text stringlengths 10 951k | source stringlengths 39 44 |
|---|---|
Rendering (computer graphics)
Rendering or image synthesis is the automatic process of generating a photorealistic or non-photorealistic image from a 2D or 3D model by means of a computer program. The resulting image is referred to as the render. Multiple models can be defined in a "scene file" containing objects in a strictly defined language or data structure. The scene file contains geometry, viewpoint, texture, lighting, and shading information describing the virtual scene. The data contained in the scene file is then passed to a rendering program to be processed and output to a digital image or raster graphics image file. The term "rendering" is analogous to the concept of an artist's impression of a scene. The term "rendering" is also used to describe the process of calculating effects in a video editing program to produce the final video output.
Rendering is one of the major sub-topics of 3D computer graphics, and in practice it is always connected to the others. It is the last major step in the graphics pipeline, giving models and animation their final appearance. With the increasing sophistication of computer graphics since the 1970s, it has become a more distinct subject.
Rendering has uses in architecture, video games, simulators, movie and TV visual effects, and design visualization, each employing a different balance of features and techniques. A wide variety of renderers are available for use. Some are integrated into larger modeling and animation packages, some are stand-alone, and some are free open-source projects. On the inside, a renderer is a carefully engineered program based on multiple disciplines, including light physics, visual perception, mathematics, and software development.
Though the technical details of rendering methods vary, the general challenges to overcome in producing a 2D image on a screen from a 3D representation stored in a scene file are handled by the graphics pipeline in a rendering device such as a GPU. A GPU is a purpose-built device that assists a CPU in performing complex rendering calculations. If a scene is to look relatively realistic and predictable under virtual lighting, the rendering software must solve the rendering equation. The rendering equation doesn't account for all lighting phenomena, but instead acts as a general lighting model for computer-generated imagery.
In the case of 3D graphics, scenes can be pre-rendered or generated in realtime. Pre-rendering is a slow, computationally intensive process that is typically used for movie creation, where scenes can be generated ahead of time, while real-time rendering is often done for 3D video games and other applications that must dynamically create scenes. 3D hardware accelerators can improve realtime rendering performance.
When the pre-image (a wireframe sketch usually) is complete, rendering is used, which adds in bitmap textures or procedural textures, lights, bump mapping and relative position to other objects. The result is a completed image the consumer or intended viewer sees.
For movie animations, several images (frames) must be rendered, and stitched together in a program capable of making an animation of this sort. Most 3D image editing programs can do this.
A rendered image can be understood in terms of a number of visible features. Rendering research and development has been largely motivated by finding ways to simulate these efficiently. Some relate directly to particular algorithms and techniques, while others are produced together.
Many rendering have been researched, and software used for rendering may employ a number of different techniques to obtain a final image.
Tracing every particle of light in a scene is nearly always completely impractical and would take a stupendous amount of time. Even tracing a portion large enough to produce an image takes an inordinate amount of time if the sampling is not intelligently restricted.
Therefore, a few loose families of more-efficient light transport modelling techniques have emerged:
The fourth type of light transport technique, radiosity is not usually implemented as a rendering technique, but instead calculates the passage of light as it leaves the light source and illuminates surfaces. These surfaces are usually rendered to the display using one of the other three techniques.
Most advanced software combines two or more of the techniques to obtain good-enough results at reasonable cost.
Another distinction is between image order algorithms, which iterate over pixels of the image plane, and object order algorithms, which iterate over objects in the scene. Generally object order is more efficient, as there are usually fewer objects in a scene than pixels.
A high-level representation of an image necessarily contains elements in a different domain from pixels. These elements are referred to as s. In a schematic drawing, for instance, line segments and curves might be primitives. In a graphical user interface, windows and buttons might be the primitives. In rendering of 3D models, triangles and polygons in space might be primitives.
If a pixel-by-pixel (image order) approach to rendering is impractical or too slow for some task, then a primitive-by-primitive (object order) approach to rendering may prove useful. Here, one loops through each of the primitives, determines which pixels in the image it affects, and modifies those pixels accordingly. This is called rasterization, and is the rendering method used by all current graphics cards.
Rasterization is frequently faster than pixel-by-pixel rendering. First, large areas of the image may be empty of primitives; rasterization will ignore these areas, but pixel-by-pixel rendering must pass through them. Second, rasterization can improve cache coherency and reduce redundant work by taking advantage of the fact that the pixels occupied by a single primitive tend to be contiguous in the image. For these reasons, rasterization is usually the approach of choice when interactive rendering is required; however, the pixel-by-pixel approach can often produce higher-quality images and is more versatile because it does not depend on as many assumptions about the image as rasterization.
The older form of rasterization is characterized by rendering an entire face (primitive) as a single color. Alternatively, rasterization can be done in a more complicated manner by first rendering the vertices of a face and then rendering the pixels of that face as a blending of the vertex colors. This version of rasterization has overtaken the old method as it allows the graphics to flow without complicated textures (a rasterized image when used face by face tends to have a very block-like effect if not covered in complex textures; the faces are not smooth because there is no gradual color change from one primitive to the next). This newer method of rasterization utilizes the graphics card's more taxing shading functions and still achieves better performance because the simpler textures stored in memory use less space. Sometimes designers will use one rasterization method on some faces and the other method on others based on the angle at which that face meets other joined faces, thus increasing speed and not hurting the overall effect.
In ray casting the geometry which has been modeled is parsed pixel by pixel, line by line, from the point of view outward, as if casting rays out from the point of view. Where an object is intersected, the color value at the point may be evaluated using several methods. In the simplest, the color value of the object at the point of intersection becomes the value of that pixel. The color may be determined from a texture-map. A more sophisticated method is to modify the colour value by an illumination factor, but without calculating the relationship to a simulated light source. To reduce artifacts, a number of rays in slightly different directions may be averaged.
Ray casting involves calculating the "view direction" (from camera position), and incrementally following along that "ray cast" through "solid 3d objects" in the scene, while accumulating the resulting value from each point in 3D space. This is related and similar to "ray tracing" except that the raycast is usually not "bounced" off surfaces (where the "ray tracing" indicates that it is tracing out the lights path including bounces). "Ray casting" implies that the light ray is following a straight path (which may include travelling through semi-transparent objects). The ray cast is a vector that can originate from the camera or from the scene endpoint ("back to front", or "front to back"). Sometimes the final light value is derived from a "transfer function" and sometimes it's used directly.
Rough simulations of optical properties may be additionally employed: a simple calculation of the ray from the object to the point of view is made. Another calculation is made of the angle of incidence of light rays from the light source(s), and from these as well as the specified intensities of the light sources, the value of the pixel is calculated. Another simulation uses illumination plotted from a radiosity algorithm, or a combination of these two.
Ray tracing aims to simulate the natural flow of light, interpreted as particles. Often, ray tracing methods are utilized to approximate the solution to the rendering equation by applying Monte Carlo methods to it. Some of the most used methods are path tracing, bidirectional path tracing, or Metropolis light transport, but also semi realistic methods are in use, like Whitted Style Ray Tracing, or hybrids. While most implementations let light propagate on straight lines, applications exist to simulate relativistic spacetime effects.
In a final, production quality rendering of a ray traced work, multiple rays are generally shot for each pixel, and traced not just to the first object of intersection, but rather, through a number of sequential 'bounces', using the known laws of optics such as "angle of incidence equals angle of reflection" and more advanced laws that deal with refraction and surface roughness.
Once the ray either encounters a light source, or more probably once a set limiting number of bounces has been evaluated, then the surface illumination at that final point is evaluated using techniques described above, and the changes along the way through the various bounces evaluated to estimate a value observed at the point of view. This is all repeated for each sample, for each pixel.
In distribution ray tracing, at each point of intersection, multiple rays may be spawned. In path tracing, however, only a single ray or none is fired at each intersection, utilizing the statistical nature of Monte Carlo experiments.
As a brute-force method, ray tracing has been too slow to consider for real-time, and until recently too slow even to consider for short films of any degree of quality, although it has been used for special effects sequences, and in advertising, where a short portion of high quality (perhaps even photorealistic) footage is required.
However, efforts at optimizing to reduce the number of calculations needed in portions of a work where detail is not high or does not depend on ray tracing features have led to a realistic possibility of wider use of ray tracing. There is now some hardware accelerated ray tracing equipment, at least in prototype phase, and some game demos which show use of real-time software or hardware ray tracing.
Radiosity is a method which attempts to simulate the way in which directly illuminated surfaces act as indirect light sources that illuminate other surfaces. This produces more realistic shading and seems to better capture the 'ambience' of an indoor scene. A classic example is the way that shadows 'hug' the corners of rooms.
The optical basis of the simulation is that some diffused light from a given point on a given surface is reflected in a large spectrum of directions and illuminates the area around it.
The simulation technique may vary in complexity. Many renderings have a very rough estimate of radiosity, simply illuminating an entire scene very slightly with a factor known as ambiance. However, when advanced radiosity estimation is coupled with a high quality ray tracing algorithm, images may exhibit convincing realism, particularly for indoor scenes.
In advanced radiosity simulation, recursive, finite-element algorithms 'bounce' light back and forth between surfaces in the model, until some recursion limit is reached. The colouring of one surface in this way influences the colouring of a neighbouring surface, and vice versa. The resulting values of illumination throughout the model (sometimes including for empty spaces) are stored and used as additional inputs when performing calculations in a ray-casting or ray-tracing model.
Due to the iterative/recursive nature of the technique, complex objects are particularly slow to emulate. Prior to the standardization of rapid radiosity calculation, some digital artists used a technique referred to loosely as false radiosity by darkening areas of texture maps corresponding to corners, joints and recesses, and applying them via self-illumination or diffuse mapping for scanline rendering. Even now, advanced radiosity calculations may be reserved for calculating the ambiance of the room, from the light reflecting off walls, floor and ceiling, without examining the contribution that complex objects make to the radiosity—or complex objects may be replaced in the radiosity calculation with simpler objects of similar size and texture.
Radiosity calculations are viewpoint independent which increases the computations involved, but makes them useful for all viewpoints. If there is little rearrangement of radiosity objects in the scene, the same radiosity data may be reused for a number of frames, making radiosity an effective way to improve on the flatness of ray casting, without seriously impacting the overall rendering time-per-frame.
Because of this, radiosity is a prime component of leading real-time rendering methods, and has been used from beginning-to-end to create a large number of well-known recent feature-length animated 3D-cartoon films.
One problem that any rendering system must deal with, no matter which approach it takes, is the sampling problem. Essentially, the rendering process tries to depict a continuous function from image space to colors by using a finite number of pixels. As a consequence of the Nyquist–Shannon sampling theorem (or Kotelnikov theorem), any spatial waveform that can be displayed must consist of at least two pixels, which is proportional to image resolution. In simpler terms, this expresses the idea that an image cannot display details, peaks or troughs in color or intensity, that are smaller than one pixel.
If a naive rendering algorithm is used without any filtering, high frequencies in the image function will cause ugly aliasing to be present in the final image. Aliasing typically manifests itself as jaggies, or jagged edges on objects where the pixel grid is visible. In order to remove aliasing, all rendering algorithms (if they are to produce good-looking images) must use some kind of low-pass filter on the image function to remove high frequencies, a process called antialiasing.
Due to the large number of calculations, a work in progress is usually only rendered in detail appropriate to the portion of the work being developed at a given time, so in the initial stages of modeling, wireframe and ray casting may be used, even where the target output is ray tracing with radiosity. It is also common to render only parts of the scene at high detail, and to remove objects that are not important to what is currently being developed.
For real-time, it is appropriate to simplify one or more common approximations, and tune to the exact parameters of the scenery in question, which is also tuned to the agreed parameters to get the most 'bang for the buck'.
The implementation of a realistic renderer always has some basic element of physical simulation or emulation — some computation which resembles or abstracts a real physical process.
The term ""physically based"" indicates the use of physical models and approximations that are more general and widely accepted outside rendering. A particular set of related techniques have gradually become established in the rendering community.
The basic concepts are moderately straightforward, but intractable to calculate; and a single elegant algorithm or approach has been elusive for more general purpose renderers. In order to meet demands of robustness, accuracy and practicality, an implementation will be a complex combination of different techniques.
Rendering research is concerned with both the adaptation of scientific models and their efficient application.
This is the key academic/theoretical concept in rendering. It serves as the most abstract formal expression of the non-perceptual aspect of rendering. All more complete algorithms can be seen as solutions to particular formulations of this equation.
Meaning: at a particular position and direction, the outgoing light (Lo) is the sum of the emitted light (Le) and the reflected light. The reflected light being the sum of the incoming light (Li) from all directions, multiplied by the surface reflection and incoming angle. By connecting outward light to inward light, via an interaction point, this equation stands for the whole 'light transport' — all the movement of light — in a scene.
The bidirectional reflectance distribution function (BRDF) expresses a simple model of light interaction with a surface as follows:
Light interaction is often approximated by the even simpler models: diffuse reflection and specular reflection, although both can ALSO be BRDFs.
Rendering is practically exclusively concerned with the particle aspect of light physics — known as geometrical optics. Treating light, at its basic level, as particles bouncing around is a simplification, but appropriate: the wave aspects of light are negligible in most scenes, and are significantly more difficult to simulate. Notable wave aspect phenomena include diffraction (as seen in the colours of CDs and DVDs) and polarisation (as seen in LCDs). Both types of effect, if needed, are made by appearance-oriented adjustment of the reflection model.
Though it receives less attention, an understanding of human visual perception is valuable to rendering. This is mainly because image displays and human perception have restricted ranges. A renderer can simulate an almost infinite range of light brightness and color, but current displays — movie screen, computer monitor, etc. — cannot handle so much, and something must be discarded or compressed. Human perception also has limits, and so does not need to be given large-range images to create realism. This can help solve the problem of fitting images into displays, and, furthermore, suggest what short-cuts could be used in the rendering simulation, since certain subtleties won't be noticeable. This related subject is tone mapping.
Mathematics used in rendering includes: linear algebra, calculus, numerical mathematics, signal processing, and Monte Carlo methods.
Rendering for movies often takes place on a network of tightly connected computers known as a render farm.
The current state of the art in 3-D image description for movie creation is the Mental Ray scene description language designed at Mental Images and RenderMan Shading Language designed at Pixar (compare with simpler 3D fileformats such as VRML or APIs such as OpenGL and DirectX tailored for 3D hardware accelerators).
Other renderers (including proprietary ones) can and are sometimes used, but most other renderers tend to miss one or more of the often needed features like good texture filtering, texture caching, programmable shaders, highend geometry types like hair, subdivision or nurbs surfaces with tesselation on demand, geometry caching, raytracing with geometry caching, high quality shadow mapping, speed or patent-free implementations. Other highly sought features these days may include interactive photorealistic rendering (IPR) and hardware rendering/shading. | https://en.wikipedia.org/wiki?curid=6604 |
Chaosium
Chaosium Inc. is a publisher of role-playing games established by Greg Stafford in 1975. Its first title was the board game "White Bear and Red Moon" (later renamed "Dragon Pass"), set in Stafford's fictional fantasy gaming world of Glorantha. Chaosium's major titles have included the roleplaying games "Call of Cthulhu" based on the horror fiction stories of H.P. Lovecraft", and RuneQuest" set in Glorantha.
Many of Chaosium’s product lines are based upon literary sources. While Stafford himself has been described as "one of the most decorated game designers of all time" and "the grand shaman of gaming", many other notable game designers have written material for Chaosium. These include David Conyers, Matthew Costello, Larry DiTillio, David A. Hargrave, Rob Heinsoo, Keith Herber, Jennell Jaquays, Katharine Kerr, Reiner Knizia, Charlie Krank, Robin Laws, Penelope Love, Mark Morrison, Steve Perrin, Sandy Petersen, Ken Rolston, Ken St. Andre, Jonathan Tweet, and Lynn Willis, among others.
Greg Stafford founded "The Chaosium" in 1975 to publish his board game "White Bear and Red Moon". He derived the name partly from his home, which was near the Oakland Coliseum, combining "coliseum" with "chaos."
In 1978 Chaosium published Steve Perrin's roleplaying game "RuneQuest", set in Stafford's mythic fantasy setting Glorantha, following up with a second edition in 1980 and various supplements over the next six years.
In 1980, the company officially incorporated as Chaosium Inc. That year, Stafford and Lynn Willis simplified the RuneQuest rules into the 16-page Basic Role-Playing (BRP). These simulationist, skill-based generic rules formed the basis of many of Chaosium's later "d100" RPGs, most notably "Call of Cthulhu", first published in 1982.
Chaosium entered into a licensing agreement with Avalon Hill in 1983 to produce a third edition of "RuneQuest". Avalon Hill manufactured and marketed the game, while Chaosium was responsible for acquisitions, design, development and layout. Ken Rolston managed the line as "Rune Czar".
One of the first RPGs by a female lead designer was published by Chaosium: Kerie Campbell-Robson's 1986 release Hawkmoon.
In 1996 it was prematurely reported that Chaosium had secured the rights to publish a collectible card game based on the video game Doom.
In 1998, following the financial failure of the collectible card game "Mythos", Greg Stafford resigned as Chaosium president and left the company, along with Sandy Petersen (although they both remained shareholders). Chaosium effectively split up into various successor companies, each maintaining its focus on a few of the company's products. Stafford took the rights to his game setting Glorantha, setting up the company Issaries, Inc. to continue publishing this line (later licensing it to Moon Design Publications, along with the game HeroQuest).
Long-time employees and part-owners Charlie Krank and Lynn Willis remained at Chaosium as President and Editor-in-Chief respectively, continuing on with "Call of Cthulhu" as the main product line. Lynn Willis retired in 2008 due to poor health and died in 2013.
Problems and delays fulfilling the Kickstarters for the 7th edition of "Call of Cthulhu" led Stafford and Petersen to return to an active role at Chaosium in June 2015. Charlie Krank subsequently left the company.
Later that year at Gen Con 2015, Stafford and Petersen announced Moon Design Publications were now part of the Chaosium ownership, and the four principals of Moon Design (Rick Meints, Jeff Richard, Michael O'Brien and Neil Robinson) had become the new Chaosium management team. Chaosium once again became the licensed publisher for "RuneQuest", "HeroQuest" and other products related to Gloranthan universe, and continue to publish the "Call of Cthulhu" line. Stafford and Petersen remained as board members (Stafford as chair), and creative consultants to the company.
As part of its financial reorganization, the new management closed the company office and warehouse in Hayward, California, ending Chaosium's long association with the San Francisco Bay Area. The company is now based in Ann Arbor, Michigan and uses a fulfillment house model for distribution of product.
Delivery of the core rewards of the Call of Cthulhu 7th Edition Kickstarter finally commenced in April 2016. The new edition went on to win nine of the ten awards it was nominated for at the Gen Con 2017 ENnie Awards.
On April 2, 2019, Chaosium Inc. announced they acquired the rights to the 7th Sea product line (both Second Edition and Khitai Kickstarters) from John Wick, including back stock of books published so far.
Chaosium began publishing a line of non-game books (primarily fiction) in 1993. Many titles are themed around H.P. Lovecraft's Cthulhu Mythos and related topics, although the first work published was Greg Stafford's fantasy work "King of Sartar", set in his mythic world Glorantha.
"Cassilda's Song", a 2015 anthology based on Robert W. Chambers's King in Yellow and written entirely by women, was nominated for two 2016 World Fantasy Awards.
In May, 2017, Chaosium appointed award-winning author and editor James Lowder as executive editor of fiction. Lowder had previously served as a consultant for Chaosium, helping the company and freelancers resolve payment and contract problems with past fiction projects.
Although not published by Chaosium, the ongoing Wild Cards series of superhero science fiction originated from a long-running "Superworld" campaign gamemastered by Game of Thrones author George R. R. Martin and his circle of fellow writers who played in his game.
Three magazines have been published by Chaosium to promote its products: | https://en.wikipedia.org/wiki?curid=6607 |
Carolina Panthers
The Carolina Panthers are a professional American football team based in Charlotte, North Carolina. The Panthers compete in the National Football League (NFL), as a member club of the league's National Football Conference (NFC) South division. The team is headquartered in Bank of America Stadium in uptown Charlotte; the stadium also serves as the team's home field. They are one of the few NFL teams to own the stadium they play in, which is legally registered as Panthers Stadium, LLC. The Panthers are supported throughout the Carolinas; although the team has played its home games in Charlotte since 1996, they played their home games at Memorial Stadium in Clemson, South Carolina during its first season. The team hosts its annual training camp at Wofford College in Spartanburg, South Carolina.
The Panthers were announced as the league's 29th franchise in 1993, and began play in 1995 under original owner and founder Jerry Richardson. The Panthers played well in their first two years, finishing in 1995 (an all-time best for an NFL expansion team's first season) and 12–4 the following year, winning the NFC West before ultimately losing to the eventual Super Bowl champion Green Bay Packers in the NFC Championship Game. They did not have another winning season until 2003, when they won the NFC Championship Game and reached Super Bowl XXXVIII, losing 32–29 to the New England Patriots. After recording playoff appearances in 2005 and 2008, the team failed to record another playoff appearance until 2013, the first of three consecutive NFC South titles. After losing in the divisional round to the San Francisco 49ers in 2013 and the Seattle Seahawks in 2014, the Panthers returned to the Super Bowl in 2015, but lost to the Denver Broncos. The Panthers have reached the playoffs eight times, advancing to four NFC Championship Games and two Super Bowls. They have won six division titles, one in the NFC West and five in the NFC South.
The Carolina Panthers are legally registered as Panther Football, LLC. and are controlled by David Tepper, whose purchase of the team from founder Jerry Richardson was unanimously approved by league owners on May 22, 2018. The club is worth approximately US$2.3 billion, according to "Forbes".
On December 15, 1987, entrepreneur Jerry Richardson announced his bid for an NFL expansion franchise in the Carolinas. A North Carolina native, Richardson was a former wide receiver on the Baltimore Colts who had used his 1959 league championship bonus to co-found the Hardee's restaurant chain, later becoming president and CEO of TW Services. Richardson drew his inspiration to pursue an NFL franchise from George Shinn, who had made a successful bid for an expansion National Basketball Association (NBA) team in Charlotte, the Charlotte Hornets. Richardson founded Richardson Sports, a partnership consisting of himself, his family, and a number of businessmen from North and South Carolina were also recruited to be limited partners. Richardson looked at four potential locations for a stadium, ultimately choosing uptown Charlotte.
To highlight the demand for professional football in the Carolinas, Richardson Sports held preseason games around the area from 1989 to 1991. The first two games were held at Carter–Finley Stadium in Raleigh, North Carolina, and Kenan Memorial Stadium in Chapel Hill, North Carolina, while the third and final game was held at Williams-Brice Stadium in Columbia, South Carolina. The matchups were between existing NFL teams. In 1991, the group formally filed an application for the open expansion spot, and on October 26, 1993, the 28 NFL owners unanimously named the Carolina Panthers as the 29th member of the NFL.
The Panthers first competed in the 1995 NFL season; they were one of two expansion teams to begin play that year, the other being the Jacksonville Jaguars. The Panthers were put in the NFC West to increase the size of that division to five teams; there were already two other southeastern teams in the division, the Atlanta Falcons and the New Orleans Saints. Former Pittsburgh Steelers defensive coordinator Dom Capers was named the first head coach. The team finished its inaugural season , the best performance ever from a first-year expansion team. They performed even better in their second season, finishing with a record and winning the NFC West division, as well as securing a first-round bye. The Panthers beat the defending Super Bowl champions Dallas Cowboys in the divisional round before losing the NFC Championship Game to the eventual Super Bowl champions, the Green Bay Packers. The team managed only a finish in 1997 and slipped to in 1998, leading to Capers' dismissal as head coach.
The Panthers hired former San Francisco 49ers head coach George Seifert to replace Capers, and he led the team to an record in 1999. The team finished in 2000 and fell to in 2001, winning their first game but losing their last 15. This performance tied the NFL record for most losses in a single season and it broke the record held by the winless 1976 Buccaneers for most consecutive losses in a single season (both records have since been broken by the 2008 Lions), leading the Panthers to fire Seifert.
After the NFL's expansion to 32 teams in 2002, the Panthers were relocated from the NFC West to the newly created NFC South division. The Panthers' rivalries with the Falcons and Saints were maintained, and they would be joined by the Tampa Bay Buccaneers. New York Giants defensive coordinator John Fox was hired to replace Seifert and led the team to a finish in 2002. Although the team's defense gave up very few yards, ranking the second-best in the NFL in yards conceded, they were hindered by an offense that ranked as the second-worst in the league in yards gained. The Panthers improved to in the 2003 regular season, winning the NFC South and making it to Super Bowl XXXVIII before losing to the New England Patriots, 32–29, in what was immediately hailed by sportswriter Peter King as the "Greatest Super Bowl of all time". King felt the game "was a wonderful championship battle, full of everything that makes football dramatic, draining, enervating, maddening, fantastic, exciting" and praised, among other things, the unpredictability, coaching, and conclusion. The game is still viewed as one of the best Super Bowls of all time, and in the opinion of Charlotte-based NPR reporter Scott Jagow, the Panthers' Super Bowl appearance represented the arrival of Charlotte onto the national scene.
Following a start in 2004, the Panthers rebounded to win six of their last seven games despite losing 14 players for the season due to injury. They lost their last game to New Orleans, finishing the 2004 season at . Had they won the game, the Panthers would have made the playoffs. The team improved to in 2005, finishing second in the division behind Tampa Bay and clinching a playoff berth as a wild-card. In the first round of the playoffs, the Panthers went on the road to face the New York Giants, beating them 23–0 for the NFL's first playoff shutout against a home team since 1980. The following week, they beat Chicago 29–21 on the road, but lost key players Julius Peppers, a defensive end, and DeShaun Foster, a running back, who were both injured during the game. The Panthers were then defeated 34–14 by the Seattle Seahawks in the NFC Championship Game, ending their season. Although the Panthers went into the 2006 season as favorites to win the NFC South, they finished with a disappointing record. The team finished the 2007 season with a record after losing quarterback Jake Delhomme early in the season due to an elbow injury. In 2008, the Panthers rebounded with a regular season record, winning the NFC South and securing a first-round bye. They were eliminated in the divisional round of the playoffs, losing 33–13 to the eventual NFC Champion Arizona Cardinals after Delhomme turned the ball over six times. Delhomme's struggles carried over into the 2009 season, where he threw 18 interceptions in the first 11 games before breaking a finger in his throwing hand. The Panthers were at a record before Delhomme's season-ending injury, and his backup, Matt Moore, led the team to a finish to the season for an overall record.
In 2010, after releasing Delhomme in the offseason, the Panthers finished with a league-worst () record; their offense was the worst in the league. John Fox's contract expired after the season ended, and the team did not retain him or his staff.
The team hired Ron Rivera to replace Fox as head coach and drafted Auburn's Heisman Trophy-winning quarterback Cam Newton with the first overall pick in the 2011 NFL Draft. The Panthers opened the 2011 season , but finished with a record, and Newton was awarded the AP Offensive Rookie of the Year award after setting the NFL record for most rushing touchdowns from a quarterback (14) in a single season and becoming the first rookie NFL quarterback to throw for over 4,000 yards in a single season. He also was the first rookie quarterback to rush for over 500 yards in a single season. After strengthening the defense with future all-pro Luke Kuechly in the 2012 draft, the Panthers again opened the 2012 season poorly, losing five out of their first six games, leading longtime general manager Marty Hurney to be fired in response. The team slid to a record before winning five of their last six games, resulting in a record. This strong finish helped save Rivera's job. The Panthers had a winning season the following year, finishing with a record and winning their third NFC South title and another playoff bye, but they were beaten by the 49ers in the Divisional Round. In 2014, the Panthers opened the season with two wins, but after 12 games sat at due in part to a seven-game winless streak. A four-game winning streak to end the season secured the team their second consecutive NFC South championship and playoff berth, despite a losing record of . The Panthers defeated the Arizona Cardinals, 27–16, in the wild card round to advance to the divisional playoffs, where they lost to eventual NFC champion Seattle, 31–17. The 2015 season saw the Panthers start the season and finish the season , which tied for the best regular-season record in NFC history. The Panthers also secured their third consecutive NFC South championship, as well as their first overall top-seeded playoff berth. In the 2015–16 playoffs, the Panthers defeated the Seattle Seahawks in the NFC Divisional playoffs, 31–24, after shutting them out in the first half, 31–0, and the Arizona Cardinals, 49–15, in the NFC Championship Game to advance to Super Bowl 50, their first Super Bowl appearance since the 2003 season. The Panthers lost a defensive struggle to the AFC champion Denver Broncos, 24–10.
In the 2016 season, the Panthers regressed on their 15–1 record from 2015, posting a 6–10 record and a last-place finish in the NFC South, missing the playoffs for the first time since 2012, and losing the division title to the second-seeded Falcons, who went on to represent the NFC in Super Bowl LI. In 2017, the Panthers finished with an 11–5 record and a #5 seed. However, they lost to the New Orleans Saints 31–26 in the Wild Card Round, their first loss in that round in franchise history.
On May 16, 2018, David Tepper, formerly a minority owner of the Pittsburgh Steelers, finalized an agreement to purchase the Panthers. The sale price was nearly $2.3 billion, a record. The agreement was approved by the league owners on May 22, 2018. The sale officially closed on July 9, 2018. After starting 6-2, the Panthers finished the 2018 season 7-9. They began the 2019 season 5-3, but lost the last eight games to finish 5-11; late in the season, Tepper fired Rivera as head coach. Perry Fewell finished out the season as interim coach, going 0-4. On January 7, 2020, the Panthers hired Baylor head coach Matt Rhule as head coach. On January 15, 2020, Luke Kuechly announced his retirement from the league. On March 17, 2020, The Panthers signed Teddy Bridgewater to a three-year $63 million dollar contract. On March 24, the Carolina Panthers released their 2011 1st overall pick and 2015 MVP quarterback Cam Newton.
The shape of the Panthers logo was designed to mimic the outline of both North Carolina and South Carolina. The Panthers changed their logo and logotype in 2012, the first such change in team history. According to the team, the changes were designed to give their logo an "aggressive, contemporary look" as well to give it a more three-dimensional feel. The primary tweaks were made in the eye and mouth, where the features, particularly the muscular brow and fangs, are more pronounced, creating a more menacing look. The revised logo has a darker shade of blue over the black logo, compared to the old design, which had a shade similar to teal on top of black.
By the time they had been announced as the 29th NFL team in October 1993, the Panthers' logo and helmet design had already been finalized, but the uniform design was still under creation. After discussion, the Panthers organization decided on jerseys colored white, black, and blue, and pants colored white and silver. The exact tone of blue, which they decided would be "process blue" (a shade lighter than Duke's and darker than North Carolina's), was the most difficult color to choose.
The team's uniform has remained largely the same since its creation, with only minor alterations such as changing the sock color of the team's black uniforms from blue to black and changing the team's shoes from white to black. Richardson, a self-described traditionalist, said that no major uniform changes would be made in his lifetime.
The Panthers have three main jersey colors: black, white, and blue. Their blue jerseys, designated their alternate uniforms, are the newest and were introduced in 2002. NFL regulations allow the team to use the blue jersey up to two times in any given season. In all other games, the team must wear either their white or black jerseys; in NFL games, the home team decides whether to wear a dark or white jersey, while the away team wears the opposite. Usually the Panthers opt for white or blue when the weather is expected to be hot and for black when the weather is expected to be cold.
The Panthers typically pair their white jerseys with white pants, while the black and blue jerseys are paired with silver pants; there have only been a few exceptions to these combinations.
The first such instance was in 1998, when the team paired their white jerseys with silver pants in a game against the Indianapolis Colts. The second instance was in 2012 during a game against the Denver Broncos, when they paired their black jerseys with new black pants; this created an all-black uniform, with the exception of blue socks and silver helmets. The decision to wear blue socks was made by team captain Steve Smith, who felt the blue socks gave the uniforms a more distinct appearance compared with other teams that have all-black uniforms. The all-black uniforms won the "Greatest Uniform in NFL History" contest, a fan-voted contest run by NFL.com in July 2013. In July 2013, the team's equipment manager, Jackie Miles, said the Panthers intended to use the all-black uniform more in the future. The Panthers wore the all-black uniform three times the following season, once each in the preseason and regular season, and the third time during the home divisional round playoff game vs the 49ers. During the Panthers' 2015 Thanksgiving Day game against the Dallas Cowboys, they debuted an all-blue uniform as part of Nike's "Color Rush" series.
In the 2018 exhibition season, the Panthers wore the black pants with the blue jerseys for a home game vs. the New England Patriots, then paired the black pants with the white jerseys in an away game vs. the Pittsburgh Steelers. During the 2018 season, the team did not wear silver pants. Instead they wore the all-black uniforms and blue jerseys with white pants. Away games saw the team wear black pants with white jerseys along with the traditional all-white.
The team's uniform did not change significantly after Nike became the NFL's jersey supplier in 2012, but the collar was altered to honor former Panthers player and coach Sam Mills by featuring the phrase "Keep Pounding". Nike had conceived the idea, and the team supported the concept as a way to expose newer fans to the legacy of Mills, who died of cancer in 2005. Mills had introduced the phrase, which has since become a team slogan, in a speech that he gave to the players and coaches prior to their 2003 playoff game against Dallas; in the speech, Mills compared his fight against cancer with the team's on-field battle, saying "When I found out I had cancer, there were two things I could do – quit or keep pounding. I'm a fighter. I kept pounding. You're fighters, too. Keep pounding!"
In 2019, the Panthers unveiled new uniforms. The new uniforms are Nike's "Vapor Untouchable" and have only minor differences: the tapered strips on the pants are replaced by stripes that extend down to the socks, the reflective shoulder cloth was replaced and the hip logos were also removed. The uniforms keep the same basic look, colors, numbers as the originals. As the 2019 season was the team's 25th, the Panthers sported a 25th anniversary patch on their uniforms.
The Panthers played their first season at Memorial Stadium in Clemson, South Carolina, as their facility in uptown Charlotte was still under construction. Ericsson Stadium, called Bank of America Stadium since 2004, opened in the summer of 1996.
Bank of America Stadium is owned entirely by the Panthers, making them one of the few teams in the NFL to own the facility they play in. The stadium was specially designed by HOK Sports Facilities Group for football and also serves as the headquarters and administrative offices of the Panthers. On some days the stadium offers public tours for a fee. Private tours for groups are offered for a fee seven days a week, though there are some exceptions, and such tours must be arranged in advance.
Two bronze panther statues flank each of the stadium's three main entrances; they are the largest sculptures ever commissioned in the United States. The names of the team's original PSL owners are engraved on the base of each statue. The first two people in the Panthers Hall of Honor, team executive Mike McCormack and linebacker Sam Mills, are honored with life-sized bronze statues outside the stadium. Mills, in addition to being the only player in the Hall of Honor for over 20 years, is the only player to have had his jersey number (#51) retired by the Panthers .
The Panthers have three open-air fields next to Bank of America Stadium where they currently hold their practices; during the 1995 season, when the team played their home games in South Carolina, the team held their practices at Winthrop University in Rock Hill, South Carolina. Because the practice fields, along with the stadium, are located in uptown Charlotte, the fields are directly visible from skyscrapers as well as from a four-story condominium located across the street. According to Mike Cranston, a running joke said that the Panthers' division rivals had pooled their resources to purchase a room on the building's top floor, and that a fire at the condominium was caused by the Panthers organization. In order to prevent people from seeing inside the field while the team is practicing, the team has added "strategically planted trees and a tarp over the ... fence surrounding the fields". Additionally, they employ a security team to watch for and chase away any people who stop alongside the fence surrounding the field. In the event of bad weather, the team moves their practices to an indoor sports facility about from the stadium. The team does not own this facility. The Panthers have hosted their annual training camp at Wofford College in Spartanburg, South Carolina, since 1995.
The Panthers are supported in both North Carolina and South Carolina; South Carolina Governor Nikki Haley declared July 30, 2012, "Carolina Panthers Day" in her state, saying that "when it comes to professional teams, the Carolina Panthers are the team that South Carolina calls their own". During the 2016 NFC Championship and Super Bowl, the hashtag #OneCarolina was used by college and professional sports teams from North Carolina and South Carolina to show unified support for the Panthers.
Pat Yasinskas of ESPN.com observed in 2012 that while there is "a bit of a wine-and-cheese atmosphere at Panthers games ... there is a strong core of die-hard fans who bring energy to Bank of America Stadium. Charlotte lives and dies with the Panthers because there aren't a lot of other options in the sports world". "Sports Illustrated" graded the Panthers as having the 10th highest "NFL Fan Value Experience" in 2007, attributing much of the fan atmosphere to the team's newness when compared to the established basketball fanbase. They also observed that the stadium has scattered parking lots, each of which has a different tailgating style. Some have fried chicken, pork, or Carolina-style barbecue, while others have live bands and televisions. Pickup football games in the parking lots are common.
The Carolina Panthers have sold out all home games since December 2002, and their home attendance has ranked in the NFL's top ten since 2006.
Sir Purr, an anthropomorphic black cat who wears a jersey numbered '00', has been the Panthers' mascot since their first season. During games, Sir Purr provides sideline entertainment through skits and "silly antics". The mascot participates in a number of community events year-round, including a monthly visit to the patients at Levine Children's Hospital. Sir Purr also hosts the annual Mascot Bowl, an event which pits pro and college mascots against each other during halftime at a selected Panthers home game.
The team's cheerleaders are the Carolina Topcats, a group of 24 women who lead cheers and entertain fans at home games. The TopCats participate in both corporate and charity events. The team's drumline is PurrCussion, an ensemble of snare, tenor, and bass drummers as well as cymbal players. PurrCussion performs for fans outside the stadium and introduces players prior to home games; it consists of drummers from across the Carolinas.
Starting with the 2012 season, the Panthers introduced the Keep Pounding Drum, inspired by the aforementioned motivational speech by Sam Mills before the team's 2004 playoff game against the Cowboys. Prior to each home game, an honorary drummer hits the six-foot tall drum four times to signify the four quarters of an American football game. According to the team, the drummers "come from a variety of backgrounds and occupations, but all have overcome a great trial or adversity that has not only made them strong but also pushes them to make others around them stronger". Drummers have included current and former Panthers players, military veterans, Make-A-Wish children, and athletes from other sports, including NBA MVP and Charlotte native Stephen Curry, US women's national soccer team players Whitney Engen and Heather O'Reilly, and 7 time NASCAR Cup Series champion Jimmie Johnson.
During the inaugural season of the Panthers, the team had an official fight song, which the team played before each home game. The song, "Stand and Cheer", remains the team's official fight song, but the team does not typically play it before home games. Due to negative fan reaction "Stand and Cheer" was pulled in 1999. Since 2006, the song has returned. The team plays Neil Diamond's "Sweet Caroline" after home victories.
A "keep pounding" chant was introduced during the 2015 season which starts before the opening kickoff of each home game. As prompted by the video boards, one side of the stadium shouts "keep" and the other side replies with "pounding". The chant is similar to ones that take place at college football games.
The Carolina Panthers support a variety of non-profits in North and South Carolina through the Carolina Panthers Charities. Four annual scholarships are awarded to student athletes through the Carolina Panthers Graduate Scholarship and the Carolina Panthers Players Sam Mills Memorial Scholarship programs. Carolina Panthers Charities also offers grants to non-profits that support education, athletics, and human services in the community. The Panthers and Fisher Athletic has provided six equipment grants to high school football teams in the Carolinas each year since 2010. Carolina Panthers Charities raises funds at three annual benefits: the Countdown to Kickoff Luncheon, the team's first public event each season; Football 101, an educational workshop for fans; and the Weekend Warrior Flag Football Tournament, a two-day non-contact flag football tournament. Another annual benefit is Taste of the Panthers, a gourmet food tasting which raises funds for Second Harvest Food Bank of Metrolina.
In 2003 the Panthers and Carolinas HealthCare Foundation established the Keep Pounding Fund, a fundraising initiative to support cancer research and patient support programs. The Panthers community has raised more than $1.4 million for the fund through direct donations, charity auctions, blood drives, and an annual 5k stadium run. The Panthers and Levine Children's Hospital coordinate monthly hospital visits and VIP game-day experiences for terminally ill or hospitalized children.
In addition to these team-specific efforts, the Panthers participate in a number of regular initiatives promoted by the NFL and USA Football, the league's youth football development partner. These include USA Football Month, held throughout August to encourage and promote youth football; A Crucial Catch, the league's Breast Cancer Awareness Month program; Salute to Service, held throughout November to support military families and personnel; and PLAY 60, which encourages young NFL fans to be active for at least 60 minutes each day.
Radio coverage is provided by flagship station WBT (1110 AM) and through the Carolina Panthers Radio Network, with affiliates throughout the Carolinas, Georgia, and Virginia. The Panthers' radio broadcasting team is led by play-by-play voice Mick Mixon, with Eugene Robinson as color analyst, and WBT sports director Jim Szoke as studio host. The radio network broadcasts pre-game coverage, games with commentary, and post-game wrap-ups. It also live-broadcasts "Panther Talk", a weekly event at Bank of America Stadium which offers fans a chance to meet a player and ask questions of the staff.
National broadcasting and cable television networks cover regular season games, as well as some preseason games. Locally, Fox owned and operated station WJZY airs most regular-season games, while any home games against an AFC team air on CBS affiliate WBTV. Any appearances on Monday Night Football are simulcast on ABC affiliate WSOC-TV, while any late-season appearances on Thursday Night Football are simulcast on WBTV. Sunday night and some Thursday night games are aired on NBC affiliate WCNC-TV.
All preseason games and team specials are televised by the Carolina Panthers Television Network on flagship station WSOC-TV in Charlotte and fourteen affiliate stations throughout the Carolinas, Georgia, Virginia, and Tennessee. WSOC took over as the Panthers' television partner for the 2019 season, replacing longtime television partner WCCB, which had retained this role after losing the Fox affiliation to WJZY in 2013. The television broadcasting team consists of play-by-play commentator Mike Morgan, color analyst and former Panthers player Mike Rucker, and sideline reporter Pete Yanity. The network also hosts "The Panthers Huddle", a weekly show focusing on the Panthers' upcoming opponent. "Panthers Gameday", the Panthers' postgame show, is hosted by sports anchor Russ Owens and former Panthers lineman Kevin Donnalley on WCNC-TV.
The Panthers also offer game broadcasts in Spanish on an eight-station network fronted by WGSP-FM in Pageland, South Carolina, as well as additional radio affiliates in Mexico. Jaime Moreno provides the play-by-play while his nephew, Luis Moreno Jr., is the color commentator. They have become popular even among English-speaking Panther fans for their high-energy, colorful announcing style.
The Panthers have developed heated rivalries with the three fellow members of the NFC South (the Atlanta Falcons, Tampa Bay Buccaneers, and New Orleans Saints). The team's fiercest rivals are the Falcons and Buccaneers.
The Falcons are a natural geographic rival for the Panthers, as Atlanta is only south on I-85. The two teams have played each other twice a year since the Panthers' inception, and games between the two teams feature large contingents of Panthers fans at Atlanta's Mercedes-Benz Stadium and large contingents of Falcons fans at Bank of America Stadium.
The Panthers' rivalry with Tampa Bay has been described as the most intense in the NFC South. The rivalry originated in 2002 with the formation of the NFC South, but became particularly heated before the 2003 season with verbal bouts between players on the two teams. It escalated further when the Panthers went to Tampa Bay and beat them in what ESPN.com writer Pat Yasinskas described as "one of the most physical contests in recent memory". The rivalry has resulted in a number of severe injuries for players on both teams, some of which were caused by foul play. One of these plays, an illegal hit on Tampa Bay punt returner Clifton Smith, sparked a brief melee between the teams in 2009.
During their time in the NFC West, the Panthers began developing a rivalry with the San Francisco 49ers. This rivalry faded after the NFL moved the Panthers out of the NFC West.
The Carolina Panthers Hall of Honor was established in 1997 to honor individuals for their contributions to the Carolina Panthers organization.
Nominees for the Pro Football Hall of Fame, which "honor[s] individuals who have made outstanding contributions to professional football", are determined by a 46-member selection committee. At least 80% of voters must approve the nominee for him to be inducted.
Jerry Richardson was the founder and first owner of the Carolina Panthers. Richardson and his family owned about 48% of the team, with the remaining 52% owned by a group of 14 limited partners. Richardson and the other investors paid $206 million for the rights to start the team in 1993.
Mike McCormack, a Hall of Fame lineman for the Cleveland Browns and former coach and executive for the Seattle Seahawks, was the Panthers' first team president, presiding in that role from 1994 until his retirement in 1997; McCormack was inducted as the first person in the Carolina Panthers Hall of Honor later that year. Jerry Richardson's son, Mark, was appointed as the team's second president in 1997 and served in that role until he stepped down in 2009. His brother Jon, who had been president of Bank of America Stadium, stepped down at the same time. The resignations of Mark and Jon Richardson were unexpected, as it was thought that the two would eventually take over the team from their father. Mark Richardson was replaced by Danny Morrison, who had previously served as the athletic director of both Texas Christian University and Wofford College, Richardson's alma mater. Morrison resigned in early 2017. The role was vacant until August 2018, when Tom Glick was hired as team president. He had previously served as the COO of Manchester City.
On May 16, 2018, David Tepper, formerly a minority owner of the Pittsburgh Steelers, finalized an agreement to purchase the Carolina Panthers, for nearly $2.3 billion, a record at the time. The agreement was approved by the league owners on May 22, 2018. According to "Forbes", the Panthers are worth approximately $2.3 billion . They ranked the Carolina Panthers as the 21st-most valuable NFL team and the 36th-most valuable sports team in the world.
The Carolina Panthers have had six head coaches. Dom Capers was the head coach from 1995 to 1998 and led the team to one playoff appearance. Counting playoff games, he finished with a record of 31–35 (.470). George Seifert coached the team from 1999 to 2001, recording 16 wins and 32 losses (.333). John Fox, the team's longest-tenured head coach, led the team from 2002 to 2010 and coached the team to three playoff appearances including Super Bowl XXXVIII which the Panthers lost. Including playoff games, Fox ended his tenure with a 78–74 (.513) record, making him the first Panthers coach to finish his tenure with the team with a winning record. Ron Rivera held the position from 2011 to 2019 and led the team to four playoff appearances including Super Bowl 50. Counting playoff games, he has a career record of 79–67–1 (.541). Statistically, Rivera holds the highest winning percentage of any Panthers head coach. On December 3, 2019, following a home loss against the Washington Redskins that sent the team's record to 5–7, Rivera was fired by David Tepper. Perry Fewell, then the defensive backs coach for the team, was named interim head coach the same day. On January 7, 2020, Matt Rhule was hired to be the Panthers head coach.
Since they began playing football in 1995, the Panthers have been to four NFC Championship Games; they lost two (1996 and 2005) and won two (2003 and 2015). The Panthers have won six division championships: the NFC West championship in 1996 and the NFC South championship in 2003, 2008, 2013, 2014, and 2015. They have won the NFC South more times than any other team in the division. They have finished as runners-up in their division six times, finishing second-place in the NFC West in 1997 and 1999 and finishing second-place in the NFC South in 2005, 2006, 2007, and 2012. They have qualified for the playoffs 8 times, most recently in 2017.
Kicker John Kasay is the team's career points leader. Kasay scored 1,482 points during his 16 seasons (1995–2010) with the Panthers. Quarterback Cam Newton, who has played for the Panthers since 2011, is the career passing leader, having thrown for 20,257 yards over his six seasons with the team. Running back Jonathan Stewart is the career rushing leader for the Carolina Panthers. Stewart, during his tenure with the team (2008–2018), rushed for 6,868 yards with the Panthers. Wide receiver Steve Smith, the team's leading receiver, recorded 12,197 receiving yards during his 13-year (2001–2013) tenure with the team.
Notes
Footnotes | https://en.wikipedia.org/wiki?curid=6610 |
Chicago Bears
The Chicago Bears are a professional American football team based in Chicago. The Bears compete in the National Football League (NFL) as a member club of the league's National Football Conference (NFC) North division. The Bears have won nine NFL Championships, including one Super Bowl, and hold the NFL record for the most enshrinees in the Pro Football Hall of Fame and the most retired jersey numbers. The Bears have also recorded more victories than any other NFL franchise.
The franchise was founded in Decatur, Illinois, on September 17, 1920, and moved to Chicago in 1921. It is one of only two remaining franchises from the NFL's founding in 1920, along with the Arizona Cardinals, which was originally also in Chicago. The team played home games at Wrigley Field on Chicago's North Side through the 1970 season; they now play at Soldier Field on the Near South Side, adjacent to Lake Michigan. The Bears have a long-standing rivalry with the Green Bay Packers.
The team headquarters, Halas Hall, is in the Chicago suburb of Lake Forest, Illinois. The Bears practice at adjoining facilities there during the season, and began hosting Training Camp at Halas Hall in 2020 after major renovations.
Originally named the Decatur Staleys, the club was established by the A. E. Staley food starch company of Decatur, Illinois as a company team. This was the typical start for several early professional football franchises. The company hired George Halas and Edward "Dutch" Sternaman in 1920 to run the team. The 1920 Decatur Staleys season was their inaugural regular season completed in the newly formed American Professional Football Association (later renamed the National Football League (NFL) in 1922).
Full control of the team was turned over to Halas and Sternaman in 1921. Official team and league records cite Halas as the founder as he took over the team in 1920 when it became a charter member of the NFL.
The team relocated to Chicago in 1921, where the club was renamed the Chicago Staleys. Under an agreement reached by Halas and Sternaman with Staley, Halas purchased the rights to the club from Staley for US$100.
In 1922, Halas changed the team name from the Staleys to the Bears. The team moved into Wrigley Field, which was home to the Chicago Cubs baseball franchise. As with several early NFL franchises, the Bears derived their nickname from their city's baseball team (some directly, some indirectly – like the Bears, whose young are called "cubs"). Halas liked the bright orange-and-blue colors of his alma mater, the University of Illinois, and the Bears adopted those colors as their own, albeit in a darker shade of each (the blue is Pantone 5395, navy blue, and the orange is Pantone 1665, similar to burnt orange).
The Staleys/Bears dominated the league in the early years. Their rivalry with the Chicago Cardinals, the oldest in the NFL (and a crosstown rival from 1920 to 1959), was key in four out of the first six league titles. During the league's first six years, the Bears lost twice to the Canton Bulldogs (who took two league titles over that span), and split with their crosstown rival Cardinals (going 4–4–2 against each other over that span), but no other team in the league defeated the Bears more than a single time. During that span, the Bears posted 34 shutouts.
The Bears' rivalry with the Green Bay Packers is one of the oldest and most storied in American professional sports, dating back to 1921 (the Green Bay Packers were an independent team until they joined the NFL in 1921). In one infamous incident that year, Halas got the Packers expelled from the league in order to prevent their signing a particular player, and then graciously got them re-admitted after the Bears had closed the deal with that player.
The franchise was an early success under Halas, capturing the NFL Championship in and remaining competitive throughout the decade. In 1924 the Bears claimed the Championship after defeating the Cleveland Bulldogs on December 7, even putting the title "World's Champions" on their 1924 team photo. But the NFL had ruled that games after November 30 did not count towards league standings, and the Bears had to settle for second place behind Cleveland. Their only losing season came in .
During the 1920s the club was responsible for triggering the NFL's long-standing rule that a player could not be signed until his college's senior class had graduated. The NFL took that action as a consequence of the Bears' aggressive signing of famous University of Illinois player Red Grange within a day of his final game as a collegian.
Despite much of the on-field success, the Bears were a team in trouble. They faced the problem of increased operating costs and flatlined attendance. The Bears would only draw roughly 5,000–6,000 fans a game, while a University of Chicago game would draw 40,000–50,000 fans a game. By adding top college football draw Red Grange to the roster, the Bears knew that they found something to draw more fans to their games. C.C. Pyle was able to secure a $2,000 per game contract for Grange, and in one of the first games, the Bears defeated the Green Bay Packers, 21–0. However, Grange remained on the sidelines while learning the team's plays from Bears quarterback Joey Sternaman. Later in 1925, The Bears would go on a barnstorming tour, showing off the best football player of the day. 75,000 people paid to see Grange lead the Bears to a 17–7 victory over the Los Angeles Tigers, who were a quickly put together team of West Coast college all-stars. After a loss to San Francisco, the Bears cruised to a 60–3 over a semi-pro team called the Portland All Stars.
Any hopes that Grange would lead the Bears to glory in 1926 were quickly dashed. A failed contract talk led to Grange bolting to the AFL's New York Yankees, owned by Pyle. The Bears also lost star quarterback Joey Sternaman, who joined the Chicago Bulls of the AFL. The Bears replaced Grange with Paddy Driscoll, a star football player in his own right. The Bears used the money made from the Grange barn-storming tour to sign the man that replaced him. Grange split his time between making movies and playing football. However, the time was not right to have two competing pro football leagues, and the AFL folded after only one season. Grange would return to the Bears.
After the financial losses of the Championship season, Halas' partner Dutch Sternaman left the organization. Halas maintained full control of the Bears until his death in 1983. He also coached the team off-and-on for forty seasons, an NFL record. In the 1932 "Unofficial" NFL Championship, the Bears defeated the Portsmouth Spartans in the first indoor American football game at Chicago Stadium.
The success of the playoff game led the NFL to institute a championship game. In the very first NFL Championship, the Bears played against the New York Giants, defeating them 23–21. The teams met again in the 1934 NFL Championship where the Giants, wearing sneakers defeated the Bears 30–13 on a cold, icy day at the Polo Grounds.
From 1940–1947, quarterback Sid Luckman led the Bears to victories in four out of the five NFL Championship Games in which they appeared. The team acquired the University of Chicago's discarded nickname "Monsters of the Midway" and their now-famous helmet wishbone "C", as well as a newly penned theme song that declared them "The Pride and Joy of Illinois". One famous victory during that period was their 73–0 victory over the favored Washington Redskins at Griffith Stadium in the 1940 NFL Championship Game; the score is still an NFL record for lopsided results. The secret behind the one-sided outcome was the introduction of a new offensive formation by Halas. The T-formation, as Halas named it, involved two running backs instead of the traditional one in the backfield. Luckman established himself as one of the franchise's most elite quarterbacks. Between 1939 and 1950, he set the Bears' passing records for most career touchdowns, yards, and completions. Many of Luckman's records stood for decades before they were eclipsed by Jay Cutler in .
Cutler then went on to break Luckman's franchise record for most career passing touchdowns a year later in .
After declining throughout the 1950s, the team rebounded in to capture its eighth NFL Championship, which would be its last until 1985. The late 1960s and early-1970s produced notable players like Dick Butkus, Gale Sayers, and Brian Piccolo, who died of embryonal carcinoma in 1970. The American television network ABC aired a movie about Piccolo in 1971 entitled "Brian's Song", starring James Caan and Billy Dee Williams in the roles of Piccolo and Sayers respectively; Jack Warden won an Emmy Award for his performance as Halas. The movie was later released for theater screenings after first being shown on television. Despite Hall of Fame careers, Butkus and Sayers would also have their careers cut short due to injuries, hamstringing the Bears of this era.
Halas retired as coach in 1967 and spent the rest of his days in the front office. He became the only person to be involved with the NFL throughout the first 60 years of its existence. He was also a member of the Pro Football Hall of Fame's first induction class in 1963. As the only living founder of the NFL at the February 1970 merger between the NFL and the American Football League, the owners honored Halas by electing him the first President of the National Football Conference, a position that he held until his death in 1983. In his honor, the NFL named the NFC Championship trophy as the George Halas Memorial Trophy.
After the merger, the Bears finished the 1970 season last place in their division, a repeat of their placing in the 1969 season. In 1975, the Bears drafted Walter Payton from Jackson State University with their first pick. He won the NFL Most Valuable Player Award in the 1977–78 season. Payton would go on to eclipse Jim Brown's NFL career rushing record in 1984 before retiring in 1987, and would hold the mark until , when Emmitt Smith of the Dallas Cowboys surpassed it. Payton's career and personality would capture the hearts of Bear fans, who called him "Sweetness". He died from a rare form of liver cancer in 1999 at the age of 45.
On November 1, 1983, a day after the death of George Halas, his oldest daughter, Virginia McCaskey, took over as the majority owner of the team. Her husband, Ed McCaskey, succeeded her father as the Chairman of the Board. Their son Michael became the third president in team history. Mrs. McCaskey holds the honorary title of "secretary of the board of directors", but the 90-year–old matriarch has been called the glue that holds the franchise together. Mrs. McCaskey's reign as the owner of the Bears was not planned, as her father originally earmarked her brother, George "Mugs" Halas Jr. as the heir apparent to the franchise. However, he died of a massive heart attack in 1979. Her impact on the team is well-noted as her own family has dubbed her "The First Lady of Sports", and the "Chicago Sun-Times" has listed her as one of Chicago's most powerful women.
Mike Ditka, a tight end for the Bears from 1961 to 1966, was hired to coach the team by George Halas in 1982. His gritty personality earned him the nickname "Iron Mike". The team reached the NFC Championship game in 1984. In the 1985 season the fire in the Bears–Packers rivalry was re-lit when Ditka used 315 pound defensive tackle "Refrigerator" Perry as a running back in a touchdown play at Lambeau Field, against the Packers. The Bears won their ninth NFL Championship, first since the AFL-NFL merger, in Super Bowl XX after the 1985 season in which they dominated the NFL with their then-revolutionary 46 defense and a cast of characters that recorded the novelty rap song "The Super Bowl Shuffle". The season was notable in that the Bears had only one loss, the "unlucky 13th" game of the season, a Monday night affair in which they were defeated by the Miami Dolphins. At the time, much was made of the fact that the Dolphins were the only franchise in history to have had an undefeated season and post-season. The Dolphins came close to setting up a rematch in the Super Bowl, but lost to the New England Patriots in the AFC title game. "The Super Bowl Shuffle" was videotaped the day after that Monday night loss in Miami.
After the 1985 Championship season, the Bears remained competitive throughout the 1980s but failed to return to the Super Bowl under Ditka. Between the firing of Ditka and the hiring of Lovie Smith, the Bears had two head coaches, Dave Wannstedt and Dick Jauron. While both head coaches led the team to the playoffs once (Wannstedt in 1994 and Jauron in 2001), neither was able to accumulate a winning record or bring the Bears back to the Super Bowl. Therefore, the 1990s was largely considered to be a disappointment.
Before the Bears hired Jauron in January 1999, Dave McGinnis (Arizona's defensive coordinator, and a former Bears assistant under Ditka and Wannstedt) backed out of taking the head coaching position. The Bears scheduled a press conference to announce the hiring before McGinnis agreed to contract terms. Soon after Jauron's hiring, Mrs. McCaskey fired her son Michael as president, replacing him with Ted Phillips and promoting Michael to chairman of the board. Phillips, the current Bears president, became the first man outside of the Halas-McCaskey family to run the team.
Lovie Smith, hired on January 15, 2004, is the third post-Ditka head coach. Joining the Bears as a rookie head coach, Smith brought the highly successful Tampa 2 defensive scheme with him to Chicago. Before his second season with the Bears, the team rehired their former offensive coordinator and then Illinois head coach Ron Turner to improve the Bears' struggling offense. In , the Bears won their division and reached the playoffs for the first time in four years. Their previous playoff berth was earned by winning the NFC Central in . The Bears improved upon their success the following season, by clinching their second consecutive NFC North title during Week 13 of the season, winning their first playoff game since 1995, and earning a trip to Super Bowl XLI. However, they fell short of the championship, losing 29–17 to the Indianapolis Colts. Following the 2006 season, the club decided to give Smith a contract extension through 2011, at roughly $5 million per year. This comes a season after being the lowest paid head coach in the National Football League.
The club has played in over a thousand games since becoming a charter member of the NFL in . Through the 2010 season, they led the NFL in overall franchise wins with 704 and had an overall record of 704–512–42 (going 687–494–42 during the regular season and 17–18 in the playoffs). On November 18, 2010 the Bears recorded franchise win number 700 in a win against the Miami Dolphins.
The Bears made one of the biggest trades in franchise history, acquiring Pro Bowl quarterback Jay Cutler from the Denver Broncos in exchange for Kyle Orton and draft picks on April 2, 2009. After a disappointing 2009 campaign with the team going 7–9, Mike Martz was hired as the team's offensive coordinator on February 1, 2010. On March 5, 2010, the Bears signed defensive end Julius Peppers, running back Chester Taylor, and tight end Brandon Manumaleuna, spending over $100 million on the first day of free agency. Also during the 2010 offseason, Michael McCaskey was replaced by brother George McCaskey as chairman of the Bears. With a 38–34 win against the New York Jets, the Bears clinched the No. 2 seed and a first-round bye for the 2010–11 NFL playoffs. In their first Playoff game since Super Bowl XLI, The Bears defeated the No. 4 seed Seattle Seahawks 35–24 in the Divisional Round. The Bears reached the NFC Championship Game, where they played Green Bay Packers at Soldier Field – only the second playoff meeting between the two storied rivals, the only other game played in 1941. The Bears lost the game, 21–14.
The team started the 2011 season strong with a 7–3 record, and running back Matt Forté led the NFL in total yards from scrimmage. Eventually, quarterback Jay Cutler fractured his thumb, and Forté also was lost for the season against the Kansas City Chiefs after spraining his MCL, and the Bears, with Caleb Hanie playing, lost five straight before winning against the Minnesota Vikings with Josh McCown starting over Hanie. At season's end, general manager Jerry Angelo was fired, and former Chiefs director of scouting and former Bears scout Phil Emery was brought in. Offensive coordinator Mike Martz resigned, and eventually retired, and was replaced by offensive line coach Mike Tice. The Bears made another notable move by trading for Miami Dolphins receiver and Pro Bowl MVP Brandon Marshall. The Bears became the first team in NFL history to return six interceptions for touchdowns in the first seven games of the season, with another pick-six by Brian Urlacher in Week 9 bringing Chicago two behind the record set by the 1961 San Diego Chargers. However, the Bears missed the playoffs with a record of 10–6 (after starting the season 7–1, the first team to start with the record and miss the playoffs since the 1996 Washington Redskins), and Smith was fired on December 31.
Then-CFL head coach and former NFL journeyman Marc Trestman was hired to succeed Smith after an exhaustive search that included at least 13 known candidates. On March 20, 2013, Brian Urlacher's 13-year tenure with the Bears ended when both sides failed to agree on a contract. The Trestman era began on September 8 with a 24–21 win over the Cincinnati Bengals, making Trestman the fourth head coach in Bears history to win in his coaching debut, after George Halas (1920), Neill Armstrong (1978) and Dick Jauron (1999). The Bears ended the 2013 season 8–8, barely missing the playoffs after losing in the final week of the season to the Packers. Despite having a second-ranked offense that set numerous franchise records, the defense greatly worsened as it set franchise worsts in categories like yards allowed (6,313).
The following season was a disaster for the Bears, with the offense regressing to finish outside the top 20 in scoring. The team also allowed 50-point games in two straight weeks against the Patriots and Packers, including a franchise-high 42 points and NFL-record six touchdowns allowed in the first half against the latter, to become the first team since the 1923 Rochester Jeffersons to allow at least 50 points in consecutive games. The Bears ended the year 5–11 and last in the NFC North. Trestman and Emery were fired after the season ended.
The Bears hired Ryan Pace of the New Orleans Saints to be their new general manager on January 8, 2015. On January 16, 2015, John Fox accepted a four-year deal to become head coach. In Fox's first season as head coach, the Bears saw improvements from 2014; after "USA Today" projected the Bears to win three games, they doubled that total and finished the season with a 6–10 record, including a Thanksgiving win over the Packers at Lambeau Field.
However, during the 2016 season, the Bears regressed heavily, compiling a 3–13 record (their worst since the NFL's change to 16-game seasons in 1978). The season included several injuries to starters and secondary players, including Jay Cutler, who only played five games as a result of two separate injuries. Backup quarterback Brian Hoyer started the next three games before a broken arm put him out for the season. He was replaced by Matt Barkley, who made his first career start with the Bears. None of the three quarterbacks returned for the 2017 season.
In the 2017 NFL Draft, the team selected quarterback Mitchell Trubisky with the second-overall pick, who sat behind newly-signed quarterback Mike Glennon for the first four games before taking over. The Bears ended the season 5–11 and again finished last in the NFC North. On January 1, 2018, Fox was fired, ending his tenure in Chicago with a 14–34 record.
The Bears hired Matt Nagy from the Kansas City Chiefs as their new head coach in January 2018. General manager Ryan Pace signed receivers Taylor Gabriel, Allen Robinson, and Trey Burton in the offseason to complement second-year quarterback Mitchell Trubisky. The Bears also acquired linebacker Khalil Mack in a block-blockbuster trade from the Oakland Raiders to further bolster their defense, sending a package of draft picks that includes 2019 and 2020 1st round draft picks in exchange. Nagy's Bears clinched the NFC North on December 16, 2018 for the first time since 2010 with a 24–17 victory over the Green Bay Packers. The Bears finished the 2018 season with a 12–4 record. They lost to the defending Super Bowl Champions Philadelphia Eagles in the Wild Card round of the Playoffs after Cody Parkey's game-winning field goal attempt was partially tipped and hit the uprights in the final seconds of the game, a play coined the "Double Doink". Despite the first-round exit, Nagy was named Coach of the Year by the Pro Football Writers Association and Associated Press. He was the first Bears coach to be given the AP award since Lovie Smith in 2005 and the fifth in team history.
In 2019, the team regressed to an 8–8 record, though Nagy's combined 20 wins in 2018 and 2019 were the most by a Bears head coach in his first two seasons. During the year, renovations to Halas Hall were completed, allowing the team to move Training Camp from Ward Field on the campus of Olivet Nazarene University in Bourbonnais, Illinois to Lake Forest for 2020.
Virginia Halas McCaskey, her children, and grandchildren control 80 percent of the team, and Mrs. McCaskey votes her children's stock as well as her own. Patrick Ryan, executive chairman of Aon Corp., and Aon director Andrew McKenna own 19.7% of the club. In a "Crain's Chicago Business" article, one businessman described his wishes for the team to maximize its potential. In 2009, "Yahoo! Sports" listed the McCaskeys as the third worst owner in the NFL, stating "[T]hey get less for what they've got than any team in our league." There have been rumors that the McCaskey family might split up over the team.
In 2012, "Forbes" magazine reported that the franchise is worth $1.19 billion, making it the eighth richest franchise in the NFL. Chicago is the third largest media market in the United States.
The team has major sponsorship deals with Dr Pepper Snapple Group, Miller Brewing Company, PNC Financial Services, United Airlines, Verizon, Xfinity, and Proven IT. The team was the first in the NFL to have a presenting sponsor, with the 2004 season advertised as "Bears Football presented by BankOne "(now Chase)"". Additionally, the Bears have an agreement with WFLD (the Fox owned-and-operated station in Chicago) to broadcast pre-season football games.
Before the 2003 season, the team had two unofficial mascots named "Rocky" and "Bearman". "Rocky" was a man who donned a #1 Bears jersey, carried a megaphone, and started chants all over Soldier Field during the 1970s, 1980s, and early 1990s, in a fashion similar to Fireman Ed. There is no known source of who "Rocky" was, and presumably currently lives in Northwestern Indiana. Don Wachter, also known as "Bearman", is a season ticket holder who decided in 1995 that he could also assist the team by cheerleading, similar to Rocky. The club allowed him to run across the field with a large Bears flag during player introductions and each team score (a role currently done by the Bears 4th Phase and Bears captains). In 1996, he donned his "costume" of face paint, bear head and arms, and a number 46 jersey. "Bearman" was forced to stop wearing his costume with the introduction of Staley Da Bear in 2003; however, in 2005, Wachter was allowed in costume again.
Staley Da Bear is an anthropomorphic bear with a customized No. 00 jersey, with blue and orange eyes, synonymous with the team's main colors. His name is eponymous to corn processing company A. E. Staley, who founded the Bears' franchise. Like Rocky and Bearman, he entertains Bears fans, but like other NFL mascots, and mascots in general, Staley also makes various visits to charity events, parties, Chicago Rush AFL games, and other Bears-related events, as well as taking part in various games with his "furballs" against youth football teams at halftime.
The team also formerly had their own cheerleading squad called the Chicago Honey Bears, who were formed in 1976. However, Bears owner Virginia Halas McCaskey terminated them after the 1985 season. The squad's uniforms have changed 3 times: from 1976–1979, the uniform was a white bodysuit with navy blue sleeves, then from 1980–1984 it became a white bodysuit, but with orange sleeves and the navy was moved to the trim, and in the squad's final season in 1985, the uniform was redesigned with an orange sequin vest.
Since 1998, the Bears have partnered with 'A Safe Place,' a domestic violence shelter in Waukegan, Illinois. In June 2017, current and former Bears employees helped with renovations at the shelter by ripping up carpet, painting walls, demolishing a kitchen and building a fence. The Bears have also provided financial support throughout the years.
The Green Bay Packers are the Bears' biggest rivals since their team's inception in 1920. The Green Bay Packers currently have the lead at 97–95–6, and the teams have met twice in the postseason. The Bears won the 1941 meeting, 33–14, and eventually defeated the New York Giants in the 1941 NFL Championship Game, and the Packers won the 2011 meeting, 21–14, en route to a Super Bowl XLV win over the Pittsburgh Steelers. The teams' first meeting was a victory for the Bears (known as the Staleys at the time) in 1921 in a shutout, 20–0. The Packers claimed their first win over the Bears in 1925, 14–10. The 1924 matchup (which ended in a 3–0 win for Chicago) was notable for featuring the first ever ejection of players in a game in NFL history, as Frank Hanny of the Bears and Walter Voss of the Packers were ejected for punching each other. The rivalry also featured one of the last successful fair catch kicks in 1968, when Bears kicker Mac Percival kicked the game-winning field goal.
Chicago and Minnesota took each other on in the Vikings' inaugural game, with the Vikings defeating the Bears in a 37–13 rout, and Minnesota currently holds the series lead 60–54–2.
The Detroit Lions and Bears have faced off since the Lions' inception in 1930, when they were known as the Portsmouth Spartans, with the Spartans winning, 7–6, and Chicago winning the second meeting, 14–6. Since then, the Bears have led the series, 99–74–5. The rivalry grew in 1932, when the Bears and Spartans met in the first ever postseason game in NFL history, with the Bears winning the game 9–0. The game also was known as the first "indoor football" game, as the game took place in indoor Chicago Stadium due to a blizzard at the time. The game also started the forward pass.
Soldier Field, located on Lake Shore Drive in Chicago, is the current home of the Bears. The Bears moved to Soldier Field in 1971 after outgrowing Wrigley Field, the team's home for 50 years. Northwestern University's residential neighbors objected to their playing at Dyche Stadium, now called Ryan Field. After the AFL-NFL Merger, the newly merged league wanted their teams to play in stadiums that could hold at least 50,000 fans. Even with the portable bleachers that the team brought into Wrigley, the stadium could still only hold 46,000. Soldier Field's playing turf was changed from natural grass to astroturf before the 1971 season, and then back to natural grass in time for the start of the 1988 season. The stadium was the site of the infamous Fog Bowl playoff game between the Bears and Philadelphia Eagles.
In 2002, the stadium was closed and rebuilt with only the exterior wall of the stadium being preserved. It was closed on Sunday, January 20, 2002, a day after the Bears lost in the playoffs. It reopened on September 27, 2003 after a complete rebuild (the second in the stadium's history). Many fans refer to the rebuilt stadium as "New Soldier Field". During the season, the Bears played their home games at the University of Illinois' Memorial Stadium in Champaign, where they went 3–5.
Many critics have negative views of the new stadium. They believe that its current structure has made it more of an eyesore than a landmark; some have dubbed it the "Mistake on the Lake". Soldier Field was stripped of its National Historic Landmark designation on February 17, 2006.
In the 2005 season, the Bears won the NFC North Division and the No. 2 Seed in the NFC Playoffs, entitling them to play at least one home game in the postseason. The team hosted (and lost) their divisional round match on January 15, 2006 against the Carolina Panthers. This was the first playoff game at Soldier Field since the stadium reopened.
The stadium's end zones and midfield were not painted until the 1982 season. The design sported on the field included the bolded word "Chicago" rendered in Highway Gothic in both end zones. In 1983, the end zone design returned, with the addition of a large wishbone "C" Bears logo painted at midfield. These field markings remained unchanged until the 1996 season. In 1996 the midfield wishbone "C" was changed to a large blue Bears head, and the end zone design were painted with "Bears" in cursive. This new design remained until the 1999 season, at which point the artwork was returned to the classic "Chicago" and the "C". In the new Soldier Field, the artwork was tweaked to where one end zone had the word "Chicago" bolded and the other had "Bears".
While the Super Bowl XX Champion Bears were a fixture of mainstream American pop culture in the 1980s, the Bears made a prior mark with the 1971 American TV movie "Brian's Song" starring Billy Dee Williams as Gale Sayers and James Caan as Brian Piccolo. The film told of how Piccolo helped Sayers recover from a devastating knee injury to return to his status as one of the league's best players, and how Sayers in turn helped the Piccolo family through Brian's fatal illness. A 2001 remake of the movie for ABC starred Sean Maher as Piccolo and Mekhi Phifer as Sayers.
The 1985 team is also remembered for recording the song "The Super Bowl Shuffle", which reached number forty-one on the "Billboard" Hot 100 and was nominated for a Grammy Award. The music video for the song depicts the team rapping that they are "not here to start no trouble" but instead "just here to do the Super Bowl Shuffle". The team took a risk by recording and releasing the song before the playoffs had even begun, but were able to avoid embarrassment by going on to win Super Bowl XX by a then-record margin of 46–10. That game was one of the most watched television events in history according to the Nielsen ratings system; the game had a rating of 48.3, ranking it 7th in all-time television history.
In addition to the "Super Bowl Shuffle" rap song, the Bears' success in the 1980s – and especially the personality of head coach Mike Ditka – inspired a recurring sketch on the American sketch comedy program "Saturday Night Live", called "Bill Swerski's Superfans". The sketch featured "Cheers" co-star George Wendt, a Chicago native, as host of a radio talk-show (similar in tone to WGN radio's "The Sportswriters"), with co-panelists Carl Wollarski (Robert Smigel), Pat Arnold (Mike Myers) and Todd O'Connor (Chris Farley). To hear them tell it, "Da Bears" and Coach Ditka could do no wrong. The sketch stopped after Ditka was fired in 1993. The sketch usually showed the panelists chugging beer and eating lots of Polish sausage, and often featured Todd getting so agitated about what was happening with the Bears that he suffered a heart attack, but quickly recovered (through self-administered CPR). The sketch also features the cast predicting unrealistic blowout victories for Bears games. Da Super Fan sketch has not been brought back by "SNL", with the exception of a single appearance by Horatio Sanz as a Super Fan for the Cubs on "Weekend Update" in 2003. Outside of "SNL", George Wendt reprised his role of Swerski in the opening promo of Super Bowl XL on ABC.
On TV shows based in Chicago such as "The Bob Newhart Show", "Married... with Children", "Family Matters", "Still Standing", "According to Jim", "Early Edition" and "The Bernie Mac Show", the main characters are all Bears fans, and have worn Bears' jerseys and T-shirts on some occasions. Some episodes even show them watching Bears games. "Roseanne" is another TV show based in Illinois (albeit not in Chicago itself) to feature the Bears as the consensus household favorite, as 'Dan Connor' John Goodman is seen wearing Bears hats in several episodes. "That '70s Show" featured several Bears references, as it was based in Wisconsin, home of the Packers. On one episode while the gang is at a Bears vs. Packers game, Eric comes to the seat in a Walter Payton jersey and is booed by the surrounding Packers fans. In an episode of the Disney Channel show "Shake It Up", based in Chicago, recurring character Dina Garcia (Ainsley Bailey) sold scalped Chicago Bears tickets. More recently, "Modern Family" character Cameron Tucker has been shown as a Bears fan. In an episode of the Disney Channel show "I Didn't Do It", based in Chicago, Lindy Watson (Olivia Holt) and Logan Watson (Austin North) try to get a football signed by NFL Hall of Famer Dick Butkus after destroying their fathers Butkus signed ball, Alshon Jeffery also makes a cameo appearance as well.
Ditka's success and popularity in Chicago has led him to land analyst roles on various American football pregame shows. Ditka worked for both the "NFL on NBC" and CBS's "The NFL Today", and he currently works on ESPN's "Sunday NFL Countdown" and provided Friday night analysis on the Bears on WBBM-TV's "2 on Football" with former WBBM-TV sports director Mark Malone. He is also the color analyst for all local broadcasts of Bears preseason games. Ditka also co-starred himself alongside actor Will Ferrell in the 2005 comedy film "Kicking & Screaming".
Also, Ditka, Dick Butkus, Walter Payton, Jim McMahon, William "Refrigerator" Perry and Brian Urlacher are among Bears figures known for their appearances in TV commercials. Urlacher, whose jersey was among the league's best-selling in 2002, was featured on Nike commercials with former Atlanta Falcons quarterback Michael Vick.
In the 1961 Hanna-Barbera animated short "Rah Rah Bear", Yogi Bear helps the Bears beat the New York Giants.
The Bears were later depicted in an episode of the 1985 cartoon version of the NBC sitcom "Punky Brewster", where the Bears are playing the Green Bay Packers.
Clark Griswold (Chevy Chase) from the "National Lampoon's Vacation series" appears in some scenes wearing a navy blue with burnt orange scripting Chicago Bears ball cap. He wears the same Chicago Bears cap throughout all four "Vacation" movies.
Currently, WBBM (780 AM) and its simulcasting partner, WCFS-FM (105.9 FM) broadcast Bears games with Jeff Joniak doing the play-by-play, along with color commentator Tom Thayer, who played for the Bears from 1985–1992, and sideline reporter Zach Zaidman. Over the years, many Bears play-by-play broadcasters have included play-by-play announcers Jack Brickhouse, Joe McConnell and Wayne Larrivee, and color commentators Hub Arkush, Dick Butkus, Jim Hart and Irv Kupcinet.
Spanish radio station WLEY-FM aired the Bears games from 2012 to 2014. Since 2015, WRTO and WVIV-FM air Bears games in Spanish.
Their current preseason TV announcers on WFLD (channel 32), which also carries the majority of the team's regular season games through the "NFL on Fox", are Sam Rosen (play-by-play), Erik Kramer (color commentary) and Lou Canellis (sideline reporter). When the games are played against an AFC team, it can be aired on the CBS O&O station, WBBM-TV. Sunday Night games are broadcast on WMAQ-TV, the NBC O&O station.
Patrick Mannelly holds the record for the most seasons in a Bears uniform with 16. On the other hand, Steve McMichael holds the record for most consecutive games played by a Bear with 191; he accomplished the feat from 1981 to 1993. In second place is Payton, who played 186 games from 1975 to 1987 at running back, a position considered to be conducive to injury, only missing one game in a span of 13 seasons.
Kicker Robbie Gould became the Bears' all-time scoring leader in Week 5 of 2015 season overtaking placekicker Kevin Butler who previously held the club record for scoring the most points in his ten-year Bear career. He scored 1,116 points as the Bears kicker from 1985 to 1995. He is followed by running back Walter Payton, with 750 points. Payton holds the team record for career rushing yards with 16,726. That was an NFL record until Emmitt Smith of the Dallas Cowboys broke it in . Former Bears running back Matt Forte, who started playing for the Bears in 2008, is the closest to Payton's record with 6,985 yards. Forte also holds the team's single season record for rookies in rushing attempts, rushing yards and receptions. Mark Bortz holds the record for most Bear playoff appearances, with 13 between 1983 and 1994, and is followed by Kevin Butler, Dennis Gentry, Dan Hampton, Jay Hilgenberg, Steve McMichael, Ron Rivera, Mike Singletary, and Keith Van Horne, who have each played in 12 playoff games.
The 1940 Chicago Bears team holds the record for the biggest margin of victory in an NFL game (playoff or regular season) with a 73–0 victory over the Washington Redskins in the 1940 NFL Championship Game. The largest home victory for the Bears came in a 61–7 result against the Green Bay Packers in 1980. The largest defeat in club history was a 52–0 loss against the Baltimore Colts in 1964. The club recorded undefeated regular seasons in 1934 and 1942, but (unlike the 1972 Dolphins) did not win the championship game in either season. In 1934, the club completed a 13–0 record but were defeated by the New York Giants, and in 1942 the club completed an 11–0 record but were defeated by the Redskins. Had the Bears won either championship, the club would have completed a championship three-peat – a feat completed only by the Packers (twice), although no team has done it since the AFL-NFL merger. Halas holds the team record for coaching the most seasons with 40 and for having the most career victories of 324. Halas' victories record stood until Don Shula surpassed Halas in . Ditka is the closest Bears coach to Halas, with 112 career victories. No other Bears coach has recorded over 100 victories with the team.
During the 2006 season, return specialist Devin Hester set several kick return records. He currently holds the franchise record for most return yards with 2,261. He had six touchdown returns, setting a record for most returns in a single season. In 2007, he recorded another six touchdown season from returns. One of the most notable of these returns came on November 12, 2006, when he returned a missed field goal for a 108-yard touchdown. The record tied former teammate Nathan Vasher's previous record, which was set almost a year earlier. Additionally, Hester set a Super Bowl record by becoming the first player to return an opening kick of a Super Bowl for a touchdown. On December 20, 2010, Hester set an NFL record for most touchdowns on a punt or kickoff return with his 14th career return coming against the Minnesota Vikings. In 2011, Hester broke the record for the most punt returns against the Carolina Panthers.
In 2012, Charles Tillman set the record for most forced fumbles in a single game with 4 against the Tennessee Titans. Also against the Titans, Chicago became the first team in league history to score a touchdown pass, a touchdown run, an interception return for a touchdown, and a blocked kick/punt for a score in the same quarter. Tillman and teammate Lance Briggs became the first pair in NFL history to return an interception for a touchdown in consecutive games against the Jacksonville Jaguars and Dallas Cowboys.
This is a partial list of the Bears' last five completed seasons. For the full season-by-season franchise results, see List of Chicago Bears seasons.
"Note: The Finish, Wins, Losses, and Ties columns list regular season results and exclude any postseason play."
As of January 7, 2019
In the Pro Football Hall of Fame, the Bears have the most enshrined primary members with 30; the club also has had seven Hall of Famers spend a minor portion of their career with the franchise. Founder, owner, head coach, and player George Halas, halfback Bronko Nagurski, and Red Grange were a part of the original class of inductees in 1963. The franchise saw 14 individuals inducted into the Hall of Fame from 1963 to 1967. Offensive tackle Jim Covert and defensive end Ed Sprinkle are the most recent Chicago Bear inductees, both being inducted as seniors as part of the Pro Football Hall of Fame's centennial class of 2020.
The Bears have retired 14 uniform numbers, which is the most in the NFL, and ranks fourth behind the basketball Boston Celtics (22), baseball New York Yankees (21), and hockey Montreal Canadiens (15) for the most in North American professional sports. The Bears retired Mike Ditka's number 89 jersey on December 9, 2013. It is the last number that the Bears retired.
In honor of the team centennial anniversary, on May 20, 2019, the Chicago Bears have unveiled the Top 100 players in franchise history, as voted on by Hall of Fame writers Don Pierson and Dan Pompei, two of the most famous journalists that have ever covered the club in their long history. At the time of the publish, the list included 27 Pro Football Hall of Famers, while two more inductees would join in the 2020 class (Jim Covert and Ed Sprinkle).
Among the 100 Greatest, four active players made the list, including safety Eddie Jackson (96), defensive lineman Akiem Hicks (75), offensive lineman Kyle Long (74) and highest-ranked active Bear was Khalil Mack (60), who played only one season with the team. Long would retire the following year.
On later date, Chicagobears.com released a list titled "Top 10: Best of the rest", that featured the top 10 snubs from the centennial list. The players include (in a following order): Alex Brown, Thomas Jones, Dave Whitsell, Curtis Conway, Tim Jennings, Leslie Frazier, Roberto Garza, Marty Booker, Nathan Vasher and William Perry.
During the week of June 3, 2019 the All-Time Team was announced in parts each day starting with the All-Time defensive players, followed by the All-Time specialists and then the All-Time offensive players.
Larry Mayer of the Chicagobears.com would later state, that according to the voters "if they had included a long-snapper on the team it would have been Patrick Mannelly". | https://en.wikipedia.org/wiki?curid=6611 |
Cincinnati Bengals
The Cincinnati Bengals are a professional American football franchise based in Cincinnati. The Bengals compete in the National Football League (NFL) as a member club of the league's American Football Conference (AFC) North division. Their home stadium is Paul Brown Stadium, located in downtown Cincinnati. Their divisional opponents are the Baltimore Ravens, Cleveland Browns and Pittsburgh Steelers.
The Bengals were founded in as a member of the American Football League (AFL) by former Cleveland Browns head coach Paul Brown, and began play in the 1968 season. Brown was the Bengals' head coach from their inception to . After being dismissed as the Browns' head coach by Art Modell (who had purchased majority interest in the team in ) in January , Brown had shown interest in establishing another NFL franchise in Ohio and looked at both Cincinnati and Columbus. He ultimately chose the former when a deal between the city, Hamilton County, and Major League Baseball's Cincinnati Reds (who were seeking a replacement for the obsolete Crosley Field) was struck that resulted in an agreement to build a multipurpose stadium which could host both baseball and football games.
Due to the impending merger of the AFL and the NFL, which was scheduled to take full effect in the season, Brown agreed to join the AFL as its tenth and final franchise. The Bengals, like the other former AFL teams, were assigned to the AFC following the merger. Cincinnati was also selected because, like their neighbors the Reds, they could draw from several large neighboring cities (Louisville and Lexington, Kentucky; Columbus, Dayton, and Springfield, Ohio) that are all no more than away from downtown Cincinnati, along with Indianapolis, until the Baltimore Colts relocated there prior to the 1984 NFL season.
The Bengals won the AFC championship in and , but lost Super Bowls XVI and XXIII to the San Francisco 49ers. After Paul Brown's death in 1991, controlling interest in the team was inherited by his son, Mike Brown. In 2011, Brown purchased shares of the team owned by the estate of co-founder Austin Knowlton and is now the majority owner of the Bengals franchise.
The 1990s and the 2000s were a period of great struggle. Following the 1990 season, the team went 14 years without posting a winning record, nor qualifying to play in the NFL playoffs. The Bengals had several head coaches and several of their top draft picks did not pan out. Mike Brown, the team's "de facto" general manager, was rated as among the worst team owners in American professional sports.
Since the mid-2000s, the team's fortunes have improved. Two years after becoming head coach, Marvin Lewis guided the Bengals to their first winning season and first division title in over a decade. After the acquisition of Andy Dalton as quarterback in 2011, the Bengals had made the playoffs each season until 2016, ranking highly among NFL teams in win totals. The Bengals drafts are also highly touted, leading to a consistency that had long escaped the franchise. However, the team has remained unable to win in the postseason and have not won a playoff game since 1990, which is the longest active drought in the NFL.
The Bengals are one of the 12 NFL teams to not have won a Super Bowl ; they are also one of 8 NFL teams that have been to at least one Super Bowl, but have not won the game.
In 1967, an ownership group led by Paul Brown was granted a franchise in the American Football League. Brown named the team the Bengals in order "to give it a link with past professional football in Cincinnati". Another Cincinnati Bengals team had existed in the city and played in three previous American Football Leagues from 1937 to 1942. The city's world-renowned zoo was also home to a rare white Bengal tiger. However, possibly as an insult to Art Modell, or possibly as a homage to his own start as a head coach to the Massillon Tigers, Brown chose the exact shade of orange used by his former team. He added black as the secondary color. Brown chose a very simple logo: the word "BENGALS" in black lettering. One of the potential helmet designs Brown rejected was a striped motif that was similar to the helmets adopted by the team in 1981 and which is still in use; however, that design featured yellow stripes on a turquoise helmet which were more uniform in width. The Bengals would begin play in the 1968 season.
In 1966, the American Football League agreed to a merger with its older and more established rival, the National Football League. Among the terms of the merger was that the AFL was permitted to add one additional franchise. One of the reasons the NFL agreed to this was that they wanted an even number of clubs in the merged league, so a team needed to be added that brought the combined total number clubs in the merged league to twenty-six teams. The NFL was content for that team to be in the American Football League because it meant that the existing nine AFL clubs were the ones that had to provide players in the ensuing expansion draft and the NFL owners preferred for the ensuing dilution of talent to occur in what they had always considered to be an inferior league. For the AFL, a key motive behind their agreement to accept a new team was that the guarantee of an eventual place in the NFL meant the league could charge a steep expansion fee of $10 million–400 times the $25,000 the original eight owners paid when they founded the league in 1960. The cash from the new team provided the American Football League with the funds needed to pay the indemnities required to be paid by the AFL to the NFL, as stipulated by the merger agreement.
Prior to the merger being announced, Brown had not seriously considered joining the American Football League, and was not a supporter of what he openly regarded to be an inferior competition, once famously stating that "I didn't pay ten million dollars to be in the AFL." However, with the announcement of the merger, Brown realized that the AFL expansion franchise would likely be his only realistic path back into the NFL in the short to medium term. He ultimately acquiesced to joining the AFL when after learning that the team was guaranteed to become an NFL franchise after the merger was completed in 1970.
There was also a complication: Major League Baseball's Cincinnati Reds were in need of a facility to replace the antiquated, obsolete Crosley Field, which they had used since 1912. Parking nightmares had plagued the city as far back as the 1950s, the little park lacked modern amenities, and New York City, which in 1957 had lost both its National League teams (the Dodgers and the Giants) to Los Angeles and San Francisco, respectively, was actively courting Reds owner Powel Crosley. However, Crosley was adamant that the Reds remain in Cincinnati and tolerated worsening problems with the Crosley Field location, which were exacerbated by the Mill Creek Expressway (I-75) project that ran alongside the park.
With assistance from Ohio governor James A. Rhodes, Hamilton County and the Cincinnati city council agreed to build a single multi-purpose facility on the dilapidated riverfront section of the city. The new facility had to be ready by the opening of the 1970 NFL season and was officially named Riverfront Stadium. With the completion of the merger in 1970, the Cleveland Browns were one of three NFL teams that voluntarily moved to the AFL-based American Football Conference to give both conferences an equal number of teams and placed in the AFC Central, the same division as the Bengals. An instant rivalry was born, fueled initially by Paul Brown's rivalry with Art Modell.
For their first two seasons, the Bengals played at Nippert Stadium which is the current home of the University of Cincinnati Bearcats. The team held training camp at Wilmington College in Wilmington, through the 1968 preseason. The team finished its first season with a 3–11 record and running back Paul Robinson, who rushed for 1,023 yards, and was named the AFL Rookie of the Year.
Founder Paul Brown coached the team for its first eight seasons. One of Brown's college draft strategies was to draft players with above-average intelligence. Punter/wide receiver Pat McInally attended Harvard University and linebacker Reggie Williams attended Dartmouth College and served on Cincinnati city council while on the Bengals' roster. Because of this policy, many former players were highly articulate and went on to have successful careers in commentary and broadcasting as well as the arts. In addition, Brown had a knack for locating and recognizing pro football talent in unusual places.
In 1971, the Bengals moved to play at Riverfront Stadium, a home they shared with the Cincinnati Reds until the team moved to Paul Brown Stadium in 2000. The team reached the playoffs three times during that decade, but could not win any of those postseason games. In 1975, the team posted an 11–3 record, giving them what is to this day the highest winning percentage (.786) in franchise history. But it only earned them a wild card spot in the playoffs, behind the 12–2 Pittsburgh Steelers, who went on to win the Super Bowl. The Bengals lost to the Oakland Raiders 31–28 in the divisional playoffs.
The Bengals reached the Super Bowl twice during the 1980s, in Super Bowl XVI and Super Bowl XXIII, but lost both times to the San Francisco 49ers. The team appeared in the playoffs in 1990, making it to the second round before losing to the Los Angeles Raiders. Before the following season got underway, Paul Brown died at age 82. He had already transferred control to his son, Mike Brown, but was reported to still influence the daily operations of the team. The Bengals' fortunes changed for the worse as the team posted 14 consecutive non-winning seasons and were saddled with numerous draft busts. They began to emerge from that dismal period into a new era of increased consistency, however, after the team finished with its worst record in history, 2–14, which led to the hiring of Marvin Lewis as head coach in 2003. Carson Palmer, the future star quarterback, was drafted in 2003, but did not play a snap that whole season, as Jon Kitna had a comeback year (voted NFL Comeback Player of the Year). Despite Kitna's success, Palmer was promoted to starting quarterback the following season. Under Palmer, the team advanced to the playoffs for the first time since 1990 in the 2005 season, which also was the first time the team had a winning percentage above .500 since 1990.
The Bengals returned to the playoffs again in 2009 in a season that included the franchise's first ever division sweep. This was especially impressive since two of the teams swept by the Bengals (the Pittsburgh Steelers and the Baltimore Ravens) had both made it to the AFC Championship Game the previous season. Marvin Lewis was rewarded for the accomplishment with the NFL Coach of the Year Award. In the 2010 season, the Bengals posted a 4–12 record.
Following the disappointing 2010 season, quarterback Carson Palmer demanded to be traded. When the Bengals refused to do so, Palmer announced his retirement from the NFL. He later was moved at the NFL trade deadline to the Oakland Raiders. In the 2011 NFL draft, the Bengals selected wide receiver AJ Green in the first round, and quarterback Andy Dalton in the second round. The Bengals improved to 9–7 in the 2011 season, and clinched a playoff spot. Dalton and Green became the most prolific rookie WR-QB duo in history, connecting 65 times for 1,057 yards. However, they lost to the Houston Texans 31–10 in the Wild Card Round. In the 2012 season, the Bengals clinched a playoff spot once more with a win over the Pittsburgh Steelers, going to the playoffs in back-to-back years for the first time since 1982. However, the Bengals faced the Texans in the first round yet again and took another early exit, losing 19–13.
In the 2013 season, for the third straight year, the Bengals clinched a playoff berth and also won the AFC North, finishing with an 11–5 record. But once again, the Bengals were defeated in the wild card round, this time by the San Diego Chargers, 27–10. Most of the blame was put on Andy Dalton, who threw 2 interceptions and fumbled on a forward dive. This made the Bengals 0–5 in playoff games since Mike Brown took over as owner. The 2014 season started well with the Bengals winning their first three contests against the Baltimore Ravens, the Atlanta Falcons, and the Tennessee Titans. However, they lost their week 5 matchup at the New England Patriots, 43–17. An overtime tie to the Carolina Panthers and shutout loss to the Indianapolis Colts followed the primetime loss to the Patriots. Finishing the season 10–5–1 as the 5th seed, they lost to the Colts, 26–10, in the first round of the playoffs. This was the first time the franchise made the playoffs four straight seasons.
In 2015, the Bengals got out to a franchise-best 8–0 start with a 31–10 win over the Cleveland Browns, but they then lost multiple consequtive games yet clinched a playoff berth. However, they lost to the division rival Pittsburgh Steelers, 18–16, in the Wild Card round in the final minute, making them the first franchise in NFL history to lose five straight opening round playoff games. This frustration continued in 2016 for the Bengals. The Bengals finished the 2016 campaign with a 6–9–1 record, losing several key players to injury including AJ Green, Giovani Bernard, and Jeremy Hill. They missed the playoffs for the first time since 2010, marking the first time Andy Dalton missed the playoffs as the Bengals' starting quarterback. One notable game was a 27–27 tie against the Washington Redskins which was played in London in 2016.
Following a rough 2016 season, the Bengals looked forward into 2017. However, after starting 0–3, the Bengals never found their footing. At one point in the season, the Bengals were 5–9. There were rumors that Marvin Lewis would not return for the next season as the Bengals' head coach. However, after two come-from-behind victories over the Lions and Ravens, that eliminated both teams from the playoffs, the Bengals finished 7–9. The final two games were convincing enough for owner Mike Brown to give Lewis a new two-year contract.
The 2018 campaign began with promise for the Bengals under Lewis. Cincinnati began the season with a 4-1 record with impressive wins over the Colts, Ravens, Falcons, and the Dolphins. However, the Bengals suffered many setbacks after the hot start. Defensive Coordinator Teryl Austin was fired mid-season because of defensive woes, AJ Green was injured and officially out for the last 4 games, and Andy Dalton injured his thumb in the Bengals' first game against the Browns and replaced by Jeff Driskel for the rest of the season. The Bengals ended 2018 with a final record of 6-10 and last place in the AFC North. On December 31, 2018, with one year to go on his contract, Lewis and the Bengals mutually parted ways after three straight losing seasons under his watch.
With Zac Taylor assuming the head coaching mantle, the 2019 campaign started off with reasonable success, barely losing to Seattle 21-20 in CenturyLink Field, but what started with promise, ended in disaster. The Bengals then lost 10 more games and were 0-11 heading into December 2019. To open the month of December, they got their first win against the Jets 22-6 in Cincinnati. They eventually lost to the Patriots and lost to the Dolphins 38-35 in OT after Dalton led the team back from 23 points down in the fourth quarter. With the loss to the Dolphins, the Bengals officially clinched the #1 Overall Pick in the 2020 NFL Draft. They'd cap off the season with a win against the underperforming Cleveland Browns, finishing 2-14, equaling the 2002 season as the team's worst record in history.
When the team debuted in 1968, the Bengals' uniforms were modeled after the Cleveland Browns. When Paul Brown was fired by Art Modell, Brown still owned the equipment used by Cleveland so, after the firing, Paul Brown packed up all his equipment which he then used for his new team in Cincinnati. The Cleveland Browns' team colors were brown, orange, and white, and their helmets were solid orange with a white dorsal stripe over the crest.
The Bengals' team colors were orange, black, and white, and their helmets were a similar shade of orange, with the only variations being the word "Bengals" in black block letters (with a white outline) on either side of the helmet and no stripe on the helmet. The Cincinnati Bengals were unique in the NFL as they did not have secondary uniform numbers on the jerseys (called "TV numbers") until the 1980 season, when they appeared on the sleeves; they were the only NFL team that did not have them prior to that point. That same year, the team changed their helmet facemask color from gray to black. The team did not discard their Cleveland-like uniforms until 1981. During that year, a then-unique uniform design was introduced. Although the team kept black jerseys, white jerseys, and white pants, they were now trimmed with orange and black tiger stripes. The team also introduced the orange helmets with black tiger stripes that are still in use today.
In 1997, the Bengals designed a logo consisting of a leaping tiger, and it was added to the uniform sleeves (with this, the "TV numbers" moved to the shoulder). Another alternate logo consisted of a Bengal's head facing to the left. However, the orange helmet with black tiger stripes continued to be the trademark. In 2004, a new tiger stripe pattern and more accents were added to the uniforms. The black jerseys now featured orange tiger-striped sleeves and white side panels, while the white jerseys began to use black tiger-striped sleeves and orange shoulders. A new logo consisting of an orange "B" covered with black tiger stripes was introduced. The team also started rotating black pants and debuted an alternate orange jersey, with white side panels and black tiger-striped sleeves. The Bengals have worn their black uniforms at home throughout their history, with some exceptions such as the 1970 season when the Bengals wore white at home for the entire season, and most of the 1971 season. Since 2005, the Bengals wear white for September home games where the heat could become a factor.
The team's official mascot is a Bengal tiger named Who Dey. Aside from Who Dey, the team also has the Cincinnati Ben–Gals, the team's cheerleading squad, which included Laura Vikmanis, the oldest cheerleader in league history.
A no-huddle offense was commonly used by all teams when time in the game was running low. However, Sam Wyche, the head coach of the Bengals in 1988, along with offensive coordinator Bruce Coslet, made the high-paced offense the standard modality for the ball club regardless of time remaining. By quickly substituting and setting up for the next play—often within 5–10 seconds after the last play despite being afforded 45 seconds—the Bengals hindered the other team's defense from substituting situational players, regrouping for tactical purpose, and resting. In response the NFL instituted rules allowing the defense ample time for substitutions when offensive substitutions were made.
The hurry-up tactic was used by the franchise during the late 1980s while Sam Wyche was the coach. A rival for AFC supremacy during this time was the Buffalo Bills, coached by Marv Levy, who also used a version of the no-huddle offense starting with the 1989 season. The Bengals had beaten the Bills three times in 1988 (pre-season, regular season, and the AFC Championship Game). Marv Levy threatened to fake injuries if the Bengals used the "no-huddle" in the AFC Championship. Wyche was notified that the commissioner had ordered the "no-huddle" illegal for the game. The official notified Wyche and the Bengals' team just two hours before the game kickoff. Wyche asked to talk directly to the commissioner and word immediately came back that the "no-huddle" would not be penalized. Levy did not have his players fake injuries in the game, but installed his version the next year, 1989. The Bengals first used the "no-huddle" in 1984. Most of the high-profile games (the various games for AFC titles and regular season games) between the two led to these changes in NFL rules. Wyche also first used the timeout periods as an opportunity to bring his entire team to the sideline to talk to all eleven players, plus substitutes, at one time. This allowed trainers time to treat a cut or bruise and equipment managers time to repair an equipment defect.
The West Coast offense is the popular name for the high-percentage passing scheme designed by former Bengals assistant Bill Walsh. Walsh formulated what has become popularly known as the West Coast offense during his tenure as assistant coach for the Cincinnati Bengals from 1968 to 1975, while working under the tutelage of Brown (and before embarking on his legendary coaching tenure with the San Francisco 49ers in the 1980s). Bengals quarterback Virgil Carter was the first player to successfully implement Walsh's system, leading the NFL in pass completion percentage in 1971. Ken Anderson replaced Carter as Cincinnati's starting quarterback in 1972 and was even more successful. In 1975 he would bring widespread recognition to the West Coast offense as well as to the Cincinnati team and its quarterback in a nationally televised Monday night contest between the Bengals and a Buffalo Bills team built around the running game of star player O.J. Simpson. Anderson's 447 passing yards were enough to overcome Simpson's 197 yards on the ground in a game that proved a milestone, providing a striking contrast between the "old" game of defense-minded football and the new game of higher scores and more action through a sophisticated aerial attack. The game, in effect, offered its viewers a glimpse of the future of professional football. Anderson, who was drafted by Paul Brown in 1971 and installed as starting quarterback in 1972, made four trips to the Pro Bowl, won four passing titles, was named NFL MVP in 1981, and set the record for completion percentage in a single season in 1982 with 70.66%. Defeated frequently during the 1970s by the Pittsburgh Steelers, a team that won four Super Bowls with 9 future Hall of Fame players, the Bengals under Anderson and head coach Forrest Gregg would finally break through the Steel Curtain, defeating the Steelers during both of their meetings in 1980 and again in 1981. Anderson, who had been named the "team franchise" by Bengal tight end Bob Trumpy, would ultimately prove his worth with a career record of 91 wins and 81 losses.
The defense created to combat the West Coast offense also came from Cincinnati. Then-Bengals defensive coordinator Dick LeBeau (who later served as the team's head coach from 2000–2002) created the zone blitz in the 1980s in response to the West Coast offense.
Three members of the Hall of Fame have spent some portion of their career with the Bengals, but only Anthony Muñoz spent his entire career with the Bengals. Bengals founder and former coach Paul Brown is also in the Hall of Fame, however he was inducted before founding the Bengals and therefore is not recognized as a Bengals Hall of Famer.
In 2007, in celebration of their 40th anniversary the Bengals named an all-time team voted on by the fans.
The Bengals contract with iHeartMedia as their radio partner, with their flagship radio stations being WCKY (1530) and WEBN (102.7 FM), along with WLW (700) if not in conflict with the Reds. Most preseason and regular season games, are telecast on CBS affiliate WKRC-TV (channel 12). The current TV announcers for preseason games are Dan Hoard on play-by-play, and Dave Lapham as analyst.
"Who Dey?!" is the name of a chant of support by fans of the Cincinnati Bengals, in use since the 1980s. The entire chant is: "Who dey, who dey, who dey think gonna beat dem Bengals?" The answer screamed in unison, "Nobody." Sometimes fans will instead shout "Who Dey?" to represent the entire cheer. "Who Dey" is also the name of the team's mascot, a Bengal tiger.
The Who Dey chant was first known to be used by fans of the 1980 Cincinnati Bengals. While the origins of the chant is disputed, one possible source for the chant is a 1980 commercial for (the now-defunct) Red Frazier Ford of Cincinnati, which used this tagline: "Who's going to give you a better deal than Red Frazier?...Nobody!" Cincinnati fans who had seen the commercial many times may have just copied it when cheering.
The Who Dey chant is also steeped in local beer lore. Hudy, a leading product of Hudepohl Brewing Company through the late 1980s, bears a phonetic similarity to the "Who Dey" chant. Beer vendors who carried full cases of bottled local beer up and down the steep upper stairs of what was then Riverfront Stadium would call out "If Hudy", "Berger" and other local beer names. Raucous fans would often chant back and forth with them as the vendors called out. During the 1980 season the banter with the Hudepohl vendors grew organically into the now famous (Hu-Dey) -Who They?- chant.
The chant bears some similarities to the phrase "Who Dat?", which was officially adopted by the New Orleans Saints in 1983 but had been used by Louisiana's high school team fans for some time. The saying "Who Dat?" originated in minstrel shows and vaudeville acts in the late 19th and early 20th centuries, then it was taken up by New Orleans Jazz and various Big band folks in the 1920s and 1930s. In the late 1960s, local Louisiana High Schools, St. Augustine High School and Patterson High School reportedly have been using the cheer and Gulf Coast fans of Alcorn State University and Louisiana State University picked up the cheer in the 1970s. Southern University in Baton Rouge, Louisiana claims to have originated the cheer in the late 1960s in their version: "Who dat talking 'bout beating dem Jags?" | https://en.wikipedia.org/wiki?curid=6612 |
Yangtze
The Yangtze or Yangzi ( or ) is the longest river in Asia, the third-longest in the world and the longest in the world to flow entirely within one country. It rises at Jari Hill in the Tanggula Mountains (Tibetan Plateau) and flows in a generally easterly direction to the East China Sea. It is the sixth-largest river by discharge volume in the world. Its drainage basin comprises one-fifth of the land area of China, and is home to nearly one-third of the country's population.
The Yangtze has played a major role in the history, culture and economy of China. For thousands of years, the river has been used for water, irrigation, sanitation, transportation, industry, boundary-marking and war. The prosperous Yangtze River Delta generates as much as 20% of the PRC's GDP. The Three Gorges Dam on the Yangtze is the largest hydro-electric power station in the world. In mid-2014, the Chinese government announced it was building a multi-tier transport network, comprising railways, roads and airports, to create a new economic belt alongside the river.
The Yangtze flows through a wide array of ecosystems and is habitat to several endemic and threatened species including the Chinese alligator, the narrow-ridged finless porpoise and the Yangtze sturgeon, but also was the home of the extinct Yangtze river dolphin (or "baiji") and Chinese paddlefish. In recent years, the river has suffered from industrial pollution, plastic pollution, agricultural run-off, siltation, and loss of wetland and lakes, which exacerbates seasonal flooding. Some sections of the river are now protected as nature reserves. A stretch of the upstream Yangtze flowing through deep gorges in western Yunnan is part of the Three Parallel Rivers of Yunnan Protected Areas, a UNESCO World Heritage Site.
Because the source of the Yangtze was not ascertained until modern times, the Chinese have given different names to lower and upstream sections of the river.
Chang Jiang () is the modern Chinese name for the lower of the Yangtze from its confluence with the Min River at Yibin in Sichuan to the river mouth at Shanghai. "Chang Jiang" literally means the "Long River." In Old Chinese, this stretch of the Yangtze was simply called "Jiang/Kiang" , a character of phono-semantic compound origin, combining the water radical with the homophone (now pronounced , but "*kˤoŋ" in Old Chinese). "Krong" was probably a word in the Austroasiatic language of local peoples such as the Yue. Similar to "*krong" in Proto-Vietnamese and "krung" in Mon, all meaning "river," it is related to modern Vietnamese "sông" (river) and Khmer "kôngkea" (water).
By the Han dynasty, "Jiang" had come to mean "any" river in Chinese, and this river was distinguished as the "Great River" (). The epithet (simplified version ), means "long," was first formally applied to the river during the Six Dynasties period.
Various sections of "Chang Jiang" have local names. From Yibin to Yichang, the river through Sichuan and Chongqing Municipality is also known as the Chuan Jiang () or "Sichuan River."
In Hubei, the river is also called the Jing Jiang () or the "Jing River" after Jingzhou. In Anhui, the river takes on the local name Wan Jiang after the shorthand name for Anhui, (皖). And Yangzi Jiang () or the "Yangzi River," from which the English name Yangtze is derived, is the local name for the Lower Yangtze in the region of Yangzhou. The name likely comes from an ancient ferry crossing called Yangzi or Yangzijin (). Europeans who arrived in the Yangtze River Delta region applied this local name to the whole river. The dividing site between upstream and midstream is considered to be at Yichang and that between midstream and downstream at Hukou (Jiujiang).
The Jinsha River () is the name for of the Yangtze from Yibin upstream to the confluence with the Batang River near Yushu in Qinghai. From antiquity until the Ming dynasty, this stretch of the river was believed to be a tributary of the Yangtze while the Min River was thought to be the main course of the river above Yibin. In the Yu Gong, written in the fifth century BCE, this section is called the "Hei Shui" or the "Black Water." The name "Jinsha" originates in the Song dynasty when the river attracted large numbers of gold prospectors. Gold prospecting along the Jinsha continues to this day. Prior to the Song dynasty, other names were used including, for example () from the Three Kingdoms period.
The Tongtian River () describes the section from Yushu up to the confluence with the Dangqu River. The name comes from a fabled river in the "Journey to the West". In antiquity, it was called the Yak River. In Mongolian, this section is known as the Murui-ussu (lit. "Winding Stream"). and sometimes confused with the nearby Baishui.
The Tuotuo River () is the official headstream of the Yangtze, and flows from the glaciers of the Gar Kangri and the Geladandong Massifs in the Tanggula Mountains of southwestern Qinghai to the confluence with the Dangqu River to form the Tongtian River. In Mongolian, this section of the river known as the "Ulaan Mörön" or the "Red River."
The Tuotuo is one of three main headstreams of the Yangtze. The Dangqu River (, p "Dāngqū") is the actual geographic headwater of the Yangtze. The name is derived the Classical Tibetan for "Marsh River" (, w "'Dam Chu"). The Chumar River (楚玛尔河) is the Chinese name for the northern headwater of the Yangtze, which flows from the Hoh Xil Mountains in Qinghai into the Tongtian. Chumar is Tibetan for the "Red River."
The river was called Quian () and Quianshui () by Marco Polo and appeared on the earliest English maps as Kian or Kiam, all recording dialects which preserved forms of the Middle Chinese pronunciation of as "Kæwng". By the mid-19th century, these romanizations had standardized as Kiang; "Dajiang", e.g., was rendered as "Ta-Kiang." "Keeang-Koo," "Kyang Kew," "Kian-ku," and related names derived from mistaking the Chinese term for the mouth of the Yangtze (, p "Jiāngkǒu") as the name of the river itself.
The name Blue River began to be applied in the 18th century, apparently owing to a former name of the Dam Chu or Min and to analogy with the Yellow River, but it was frequently explained in early English references as a 'translation' of "Jiang", "Jiangkou", or "Yangzijiang". Very common in 18th- and 19th-century sources, the name fell out of favor due to growing awareness of its lack of "any" connection to the river's Chinese names and to the irony of its application to such a muddy waterway.
Matteo Ricci's 1615 Latin account included descriptions of the "Ianſu" and "Ianſuchian." The posthumous account's translation of the name as "Son of the Ocean" shows that Ricci, who by the end of his life was fluent in literary Chinese, was introduced to it as the homophonic rather than the 'proper' . Further, although railroads and the Shanghai concessions subsequently turned it into a backwater, Yangzhou was the lower river's principal port for much of the Qing dynasty, directing Liangjiang's important salt monopoly and connecting the Yangtze with the Grand Canal to Beijing. (That connection also made it one of the "Yellow" River's principal ports between the floods of 1344 and the 1850s, during which time the Yellow River ran well south of Shandong and discharged into the ocean only a few hundred kilometres away from the mouth of the Yangtze.) By 1800, English cartographers such as Aaron Arrowsmith had adopted the French style of the name as Yang-tse or Yang-tse Kiang. The British diplomat Thomas Wade emended this to Yang-tzu Chiang as part of his formerly popular romanization of Chinese, based on the Beijing dialect instead of Nanjing's and first published in 1867. The spellings Yangtze and Yangtze Kiang was a compromise between the two methods adopted at the 1906 Imperial Postal Conference in Shanghai, which established postal romanization. Hanyu Pinyin was adopted by the PRC's First Congress in 1958, but it was not widely employed in English outside mainland China prior to the normalization of diplomatic relations between the United States and the PRC in 1979; since that time, the spelling Yangzi has also been used.
The source and upper reaches of the Yangtze are located in ethnic Tibetan areas of Qinghai. In Tibetan, the Tuotuo headwaters are the "Machu" (, w "rMa-chu", literally "Red River" or (perhaps "Wound-[like Red] River?")). The Tongtian is the "Drichu" (, w "'Bri Chu", literally "River of the Female Yak"; transliterated into Chinese as , p "Zhíqū").
The river originates from several tributaries in the eastern part of the Tibetan Plateau, two of which are commonly referred to as the "source." Traditionally, the Chinese government has recognized the source as the Tuotuo tributary at the base of a glacier lying on the west of Geladandong Mountain in the Tanggula Mountains. This source is found at and while not the furthest source of the Yangtze, it is the highest source at above sea level. The true source of the Yangtze, hydrologically the longest river distance from the sea, is at Jari Hill at the head of the Dam Qu tributary, approximately southeast of Geladandong. This source was only discovered in the late 20th century and lies in wetlands at and above sea level just southeast of Chadan Township in Zadoi County, Yushu Prefecture, Qinghai. As the historical spiritual source of the Yangtze, the Geladandong source is still commonly referred to as the source of the Yangtze since the discovery of the Jari Hill source.
These tributaries join and the river then runs eastward through Qinghai (Tsinghai), turning southward down a deep valley at the border of Sichuan (Szechwan) and Tibet to reach Yunnan. In the course of this valley, the river's elevation drops from above to less than . The headwaters of the Yangtze are situated at an elevation of about . In its descent to sea level, the river falls to an altitude of at Yibin, Sichuan, the head of navigation for riverboats, and to at Chongqing (Chungking). Between Chongqing and Yichang (I-ch'ang), at an altitude of and a distance of about , it passes through the spectacular Yangtze Gorges, which are noted for their natural beauty but are dangerous to shipping.
It enters the basin of Sichuan at Yibin. While in the Sichuan basin, it receives several mighty tributaries, increasing its water volume significantly. It then cuts through Mount Wushan bordering Chongqing and Hubei to create the famous Three Gorges. Eastward of the Three Gorges, Yichang is the first city on the Yangtze Plain.
After entering Hubei, the Yangtze receives water from a number of lakes. The largest of these lakes is Dongting Lake, which is located on the border of Hunan and Hubei provinces, and is the outlet for most of the rivers in Hunan. At Wuhan, it receives its biggest tributary, the Han River, bringing water from its northern basin as far as Shaanxi.
At the northern tip of Jiangxi, Lake Poyang, the biggest freshwater lake in China, merges into the river. The river then runs through Anhui and Jiangsu, receiving more water from innumerable smaller lakes and rivers, and finally reaches the East China Sea at Shanghai.
Four of China's five main freshwater lakes contribute their waters to the Yangtze River. Traditionally, the upstream part of the Yangtze River refers to the section from Yibin to Yichang; the middle part refers to the section from Yichang to Hukou County, where Lake Poyang meets the river; the downstream part is from Hukou to Shanghai.
The origin of the Yangtze River has been dated by some geologists to about 45 million years ago in the Eocene, but this dating has been disputed.
The Yangtze flows into the East China Sea and was navigable by ocean-going vessels up from its mouth even before the Three Gorges Dam was built.
The Yangtze is flanked with metallurgical, power, chemical, auto, building materials and machinery industrial belts and high-tech development zones. It is playing an increasingly crucial role in the river valley's economic growth and has become a vital link for international shipping to the inland provinces. The river is a major transportation artery for China, connecting the interior with the coast.
The river is one of the world's busiest waterways. Traffic includes commercial traffic transporting bulk goods such as coal as well as manufactured goods and passengers. Cargo transportation reached 795 million tons in 2005. River cruises several days long, especially through the beautiful and scenic Three Gorges area, are becoming popular as the tourism industry grows in China.
Flooding along the river has been a major problem. The rainy season in China is May and June in areas south of Yangtze River, and July and August in areas north of it. The huge river system receives water from both southern and northern flanks, which causes its flood season to extend from May to August. Meanwhile, the relatively dense population and rich cities along the river make the floods more deadly and costly. The most recent major floods were the 1998 Yangtze River Floods, but more disastrous were the 1954 Yangtze River Floods, which killed around 30,000 people.
Although the mouth of the Yellow River has fluctuated widely north and south of the Shandong peninsula within the historical record, the Yangtze has remained largely static. Based on studies of sedimentation rates, however, it is unlikely that the present discharge site predates the late Miocene ( Ma). Prior to this, its headwaters drained south into the Gulf of Tonkin along or near the course of the present Red River.
The Yangtze River is important to the cultural origins of southern China and Japan. Human activity has been verified in the Three Gorges area as far back as 27,000 years ago, and by the 5th millennium BC, the lower Yangtze was a major population center occupied by the Hemudu and Majiabang cultures, both among the earliest cultivators of rice. By the 3rd millennium BC, the successor Liangzhu culture showed evidence of influence from the Longshan peoples of the North China Plain. A study of Liangzhu remains found a high prevalence of haplogroup O1, linking it to Austronesian and Daic populations; the same study found the rare haplogroup O3d at a Daxi site on the central Yangtze, indicates possible connection with the Hmong, although "only small traces" of haplogroup O3d remains in Hmong today. What is now thought of as Chinese culture developed along the more fertile Yellow River basin; the "Yue" people of the lower Yangtze possessed very different traditions blackening their teeth, cutting their hair short, tattooing their bodies, and living in small settlements among bamboo groves and were considered barbarous by the northerners.
The Central Yangtze valley was home to sophisticated Neolithic cultures. Later on it was the earliest part of the Yangtze valley to be integrated into the North Chinese cultural sphere. North Chinese people were active there from the Bronze Age.
In the lower Yangtze, two Yue tribes, the "Gouwu" in southern Jiangsu and the "Yuyue" in northern Zhejiang, display increasing Zhou (i.e., North Chinese) influence from the 9th century BC. Traditional accounts credit these changes to northern refugees (Taibo and Zhongyong in Wu and Wuyi in Yue) who assumed power over the local tribes, though these are generally assumed to be myths invented to legitimate them to other Zhou rulers. As the kingdoms of Wu and Yue, they were famed as fishers, shipwrights, and sword-smiths. Adopting Chinese characters, political institutions, and military technology, they were among the most powerful states during the later Zhou. In the middle Yangtze, the state of Jing seems to have begun in the upper Han River valley a minor Zhou polity, but it adapted to native culture as it expanded south and east into the Yangtze valley. In the process, it changed its name to Chu.
Whether native or nativizing, the Yangtze states held their own against the northern Chinese homeland: some lists credit them with three of the Spring and Autumn period's Five Hegemons and one of the Warring States' Four Lords. They fell in against themselves, however. Chu's growing power led its rival Jin to support Wu as a counter. Wu successfully sacked Chu's capital Ying in 506 BC, but Chu subsequently supported Yue in its attacks against Wu's southern flank. In 473 BC, King Goujian of Yue fully annexed Wu and moved his court to its eponymous capital at modern Suzhou. In 333 BC, Chu finally united the lower Yangtze by annexing Yue, whose royal family was said to have fled south and established the Minyue kingdom in Fujian. Qin was able to unite China by first subduing Ba and Shu on the upper Yangtze in modern Sichuan, giving them a strong base to attack Chu's settlements along the river.
The state of Qin conquered the central Yangtze region, previous heartland of Chu, in 278 BC, and incorporated the region into its expanding empire. Qin then used its connections along the Yangtze River the Xiang River to expand China into Hunan, Jiangxi and Guangdong, setting up military commanderies along the main lines of communication. At the collapse of the Qin Dynasty, these southern commanderies became the independent Nanyue Empire under Zhao Tuo while Chu and Han vied with each other for control of the north.
From the Han dynasty, the region of the Yangtze River became more and more important to China's economy. The establishment of irrigation systems (the most famous one is Dujiangyan, northwest of Chengdu, built during the Warring States period) made agriculture very stable and productive. The Qin and Han empires were actively engaged in the agricultural colonization of the Yangtze lowlands, maintaining a system of dikes to protect farmland from seasonal floods. By the Song dynasty, the area along the Yangtze had become among the wealthiest and most developed parts of the country, especially in the lower reaches of the river. Early in the Qing dynasty, the region called Jiangnan (that includes the southern part of Jiangsu, the northern part of Zhejiang, and the southeastern part of Anhui) provided – of the nation's revenues.
The Yangtze has long been the backbone of China's inland water transportation system, which remained particularly important for almost two thousand years, until the construction of the national railway network during the 20th century. The Grand Canal connects the lower Yangtze with the major cities of the Jiangnan region south of the river (Wuxi, Suzhou, Hangzhou) and with northern China (all the way from Yangzhou to Beijing). The less well known ancient Lingqu Canal, connecting the upper Xiang River with the headwaters of the Guijiang, allowed a direct water connection from the Yangtze Basin to the Pearl River Delta.
Historically, the Yangtze became the political boundary between north China and south China several times (see History of China) because of the difficulty of crossing the river. This occurred notably during the Southern and Northern Dynasties, and the Southern Song. Many battles took place along the river, the most famous being the Battle of Red Cliffs in 208 AD during the Three Kingdoms period.
The Yangtze was the site of naval battles between the Song dynasty and Jurchen Jin during the Jin–Song wars. In the Battle of Caishi of 1161, the ships of the Jin emperor Wanyan Liang clashed with the Song fleet on the Yangtze. Song soldiers fired bombs of lime and sulphur using trebuchets at the Jurchen warships. The battle was a Song victory that halted the invasion by the Jin. The Battle of Tangdao was another Yangtze naval battle from the same year.
Politically, Nanjing was the capital of China several times, although most of the time its territory only covered the southeastern part of China, such as the Wu kingdom in the Three Kingdoms period, the Eastern Jin Dynasty, and during the Southern and Northern Dynasties and Five Dynasties and Ten Kingdoms periods. Only the Ming occupied most parts of China from their capital at Nanjing, though it later moved the capital to Beijing. The ROC capital was located in Nanjing in the periods 1911–12, 1927–37, and 1945–49.
The first merchant steamer in China, the "Jardine", was built to order for the firm of Jardine Matheson in 1835. She was a small vessel intended for use as a mail and passenger carrier between Lintin Island, Macau and Whampoa. However, after several trips, the Chinese authorities, for reasons best known to themselves, prohibited her entrance into the river. Lord Palmerston, the British Foreign Secretary who personified gunboat diplomacy, decided to wage war on China mainly on the "suggestions" of Jardine Matheson. In mid-1840, a large fleet of warships appeared on the China coast, and with the first cannon fire aimed at a British ship, the "Royal Saxon", the British started the First Opium War. The Imperial Government, forced to surrender, gave in to the demands of the British. British military was vastly superior during the conflict. British warships, constructed using such innovations as steam power combined with sail and the use of iron in shipbuilding, wreaked havoc on coastal towns; such ships (like the "Nemesis") were not only virtually indestructible using contemporary available weapons, but also highly mobile and able to support a gun platform with very heavy guns. In addition, the British troops were armed with modern rifled muskets and cannons, unlike the Qing forces. After the British took Canton, they sailed up the Yangtze and took the tax barges, a devastating blow to the Empire as it slashed the revenue of the imperial court in Beijing to just a small fraction of what it had been.
In 1842, the Qing authorities sued for peace, which concluded with the Treaty of Nanking signed on a gunboat in the river, negotiated in August of that year and ratified in 1843. In the treaty, China was forced to pay an indemnity to Britain, open five ports to Britain, and cede Hong Kong to Queen Victoria. In the supplementary Treaty of the Bogue, the Qing empire also recognized Britain as an equal to China and gave British subjects extraterritorial privileges in treaty ports.
The US, at the same time, wanting to protect its interests and expand trade, ventured the six hundred miles up the river to Hankow sometime in the 1860s, while the , a sidewheeler, made her way up the river to Yichang in 1874. The first , a sidewheel gunboat, began charting the Yangtze River in 1871. The first , an armed tug, was on Asiatic Station into 1891, cruising the Chinese and Japanese coasts, visiting the open treaty ports and making occasional voyages up the Yangtze River. From June to September 1891, anti-foreign riots up the Yangtze forced the warship to make an extended voyage as far as Hankou, 600 miles upriver. Stopping at each open treaty port, the gunboat cooperated with naval vessels of other nations and repairing damage. She then operated along the north and central China coast and on the lower Yangtze until June 1892. The cessation of bloodshed with the Taiping Rebellion, Europeans put more steamers on the river. The French engaged the Chinese in war over the rule of Vietnam. The Sino-French Wars of the 1880s emerged with the Battle of Shipu having French cruisers in the lower Yangtze.
The China Navigation Company was an early shipping company founded in 1876 in London, initially to trade up the Yangtze River from their Shanghai base with passengers and cargo. Chinese coastal trade started shortly after and in 1883 a regular service to Australia was initiated. Most of the company's ships were seized by Japan in 1941 and services did not resume until 1946. Robert Dollar was a later shipping magnate, who became enormously influential moving Californian and Canadian lumber to the Chinese and Japanese market.
Yichang, or Ichang, from the sea, is the head of navigation for river steamers; oceangoing vessels may navigate the river to Hankow, a distance of almost from the sea. For about inland from its mouth, the river is virtually at sea level.
The Chinese Government, too, had steamers. It had its own naval fleet, the Nanyang Fleet, which fell prey to the French fleet. The Chinese would rebuild its fleet, only to be ravaged by another war with Japan (1895), Revolution (1911) and ongoing inefficiency and corruption. Chinese companies ran their own steamers, but were second tier to European operations at the time.
Steamers came late to the upper river, the section stretching from Yichang to Chongqing. Freshets from Himalayan snowmelt created treacherous seasonal currents. But summer was better navigationally and the three gorges, described as an "150-mile passage which is like the narrow throat of an hourglass," posed hazardous threats of crosscurrents, whirlpools and eddies, creating significant challenges to steamship efforts. Furthermore, Chongqing is 700 – 800 feet above sea level, requiring powerful engines to make the upriver climb. Junk travel accomplished the upriver feat by employing 70–80 trackers, men hitched to hawsers who physically pulled ships upriver through some of the most risky and deadly sections of the three gorges. Achibald John Little took an interest in Upper Yangtze navigation when in 1876, the Chefoo Convention opened Chongqing to consular residence but stipulated that foreign trade might only commence once steamships had succeeded in ascending the river to that point. Little formed the Upper Yangtze Steam Navigation Co., Ltd. and built "Kuling" but his attempts to take the vessel further upriver than Yichang were thwarted by the Chinese authorities who were concerned about the potential loss of transit duties, competition to their native junk trade and physical damage to their crafts caused by steamship wakes. "Kuling" was sold to China Merchants Steam Navigation Company for lower river service. In 1890, the Chinese government agreed to open Chongqing to foreign trade as long as it was restricted to native crafts. In 1895, the Treaty of Shimonoseki provided a provision which opened Chongqing fully to foreign trade. Little took up residence in Chongqing and built "Leechuan", to tackle the gorges in 1898. In March "Leechuan" completed the upriver journey to Chongqing but not without the assistance of trackers. "Leechuan" was not designed for cargo or passengers and if Little wanted to take his vision one step further, he required an expert pilot. In 1898, Little persuaded Captain Samuel Cornell Plant to come out to China to lend his expertise. Captain Plant had just completed navigation of Persia's Upper Karun River and took up Little's offer to assess the Upper Yangtze on "Leechuan" at the end of 1898. With Plant's design input, Little had SS "Pioneer" built with Plant in command. In June 1900, Plant was the first to successfully pilot a merchant steamer on the Upper Yangtze from Yichang to Chongqing. "Pioneer" was sold to British Royal Navy after its first run due to threat from the Boxer Rebellion and renamed HMS "Kinsha". Germany's steamship effort that same year on SS "Suixing" ended in catastrophe. On "Suixing's" maiden voyage, the vessel hit a rock and sunk, killing its captain and ending realistic hopes of regular commercial steam service on the Upper Yangtze. In 1908, local Sichuan merchants and their government partnered with Captain Plant to form Sichuan Steam Navigation Company becoming the first successful service between Yichang and Chongqing. Captain Plant designed and commanded its two ships, SS "Shutung" and SS "Shuhun". Other Chinese vessels came onto the run and by 1915, foreign ships expressed their interest too. Plant was appointed by Chinese Maritime Customs Service as First Senior River Inspector in 1915. In this role, Plant installed navigational marks and established signaling systems. He also wrote "Handbook for the Guidance of Shipmasters on the Ichang-Chungking Section of the Yangtze River", a detailed and illustrated account of the Upper Yangtze's currents, rocks, and other hazards with navigational instruction. Plant trained hundreds of Chinese and foreign pilots and issued licenses and worked with the Chinese government to make the river safer in 1917 by removing some of the most difficult obstacles and threats with explosives. In August 1917, British Asiatic Petroleum became the first foreign merchant steamship on the Upper Yangtze. Commercial firms, Robert Dollar Company, Jardine Matheson, Butterfield and Swire and Standard Oil added their own steamers on the river between 1917 and 1919. Between 1918 and 1919, Sichuan warlord violence and escalating civil war put Sichuan Steam Navigational Company out of business. "Shutung" was commandeered by warlords and "Shuhun" was brought down river to Shanghai for safekeeping. In 1921, when Captain Plant died tragically at sea while returning home to England, a Plant Memorial Fund was established to perpetuate Plant's name and contributions to Upper Yangtze navigation. The largest shipping companies in service, Butterfield & Swire, Jardine Matheson, Standard Oil, Mackenzie & Co., Asiatic Petroleum, Robert Dollar, China Merchants S.N. Co. and British-American Tobacco Co., contributed alongside international friends and Chinese pilots. In 1924, a 50-foot granite pyramidal obelisk was erected in Xintan, on the site of Captain Plant's home, in a Chinese community of pilots and junk owners. One face of the monument is inscribed in Chinese and another in English. Though recently relocated to higher ground ahead of the Three Gorges Dam, the monument still stands overlooking the Upper Yangtze River near Yichang, a rare collective tribute to a westerner in China.
Until 1881, the India and China coastal and river services were operated by several companies. In that year, however, these were merged into the Indo-China Steam Navigation Company Ltd, a public company under the management of Jardine's. The Jardine company pushed inland up the Yangtsze River on which a specially designed fleet was built to meet all requirements of the river trade. For many years, this fleet gave unequalled service. Jardine's established an enviable reputation for the efficient handling of shipping. As a result, the Royal Mail Steam Packet Company invited the firm to attend to the Agency of their Shire Line which operated in the Far East. Standard Oil ran the tankers Mei Ping, Mei An and Mei Hsia, which were all destroyed on December 12, 1937, when Japanese warplanes bombed and sank the U.S.S. Panay. One of the Standard Oil captains who survived this attack had served on the Upper River for 14 years.
With the Treaty Ports, the European powers and Japan were allowed to sail navy ships into China's waters. The British, Americans, and French did this. A full international fleet featured on Chinese waters: Austro-Hungarian, Portuguese, Italian, Russian and German navy ships came to Shanghai and the treaty ports. The Japanese engaged in open warfare with the Chinese over conquest of the Chinese Qing Empire in the First Sino-Japanese War in 1894–1895, and with Russia over Qing Empire territory in the Russo-Japanese War of 1904–1905. Incidentally, both the French and Japanese navies were heavily involved in running opium and narcotics to Shanghai, where it was refined into morphine. It was then transhipped by liner back to Marseille and France (i.e. French Connection) for processing in Germany and eventual sale in the U.S. or Europe.
In 1909 the gunboat changed station to Shanghai, where she regularly patrolled the lower Yangtze River up to Nanking and Wuhu. Following an anti-foreign riots in Changsha in April 1910, which destroyed a number of missions and merchant warehouses, Samar sailed up the Yangtze River to Hankow and then Changsa to show the flag and help restore order. The gunboat was also administratively assigned to the Asiatic Fleet that year, which had been reestablished by the Navy to better protect, in the words of the Bureau of Navigation, "American interests in the Orient." After returning to Shanghai in August, she sailed up river again the following summer, passing Wuhu in June but then running aground off Kichau on July 1, 1911.
After staying stuck in the mud for two weeks, Samar broke free and sailed back down river to coal ship. Returning upriver, the gunboat reached Hankow in August and Ichang in September where she wintered over owing to both the dry season and the outbreak of rebellion at Wuchang in October 1911. Tensions eased and the gunboat turned downriver in July 1912, arriving at Shanghai in October. Samar patrolled the lower Yangtze after fighting broke out in the summer 1913, a precursor to a decade of conflict between provincial warlords in China. In 1919, she was placed on the disposal list at Shanghai following a collision with a Yangtze River steamer that damaged her bow.
The Spanish boats were replaced in the 1920s by and were the largest, and next in size, and and the smallest. China in the first fifty years of the 20th century, was in low-grade chaos. Warlords, revolutions, natural disasters, civil war and invasions contributed. Yangtze boats were involved in the Nanking incident of 1927 when the Communists and Nationalists broke into open war. The Chiang Kai-shek's massacre of the Communists in Shanghai in 1927 furthered the unrest, U.S. Marines with tanks were landed. River steamers were popular targets for both Nationalists and Communists, and peasants who would take periodic pot-shots at vessels. During the course of service the second protected American interests in China down the entire length of the Yangtze, at times convoying U.S. and foreign vessels on the river, evacuating American citizens during periods of disturbance and in general giving credible presence to U.S. consulates and residences in various Chinese cities. In the period of great unrest in central China in the 1920s, Palos was especially busy patrolling the upper Yangtze against bands of warlord soldiers and outlaws. The warship engaged in continuous patrol operations between Ichang and Chungking throughout 1923, supplying armed guards to merchant ships, and protecting Americans at Chungking while that city was under siege by a warlord army.
The British Royal Navy had a series of Insect-class gunboats which patrolled between Chungking and Shanghai. Cruisers and destroyers and Fly-class gunboats also patrolled. The most infamous incident was when "Panay" and in 1937, were dive-bombed by Japanese aeroplanes during the notorious Nanking massacre. The Westerners were forced to leave the Yangtze River with the Japanese takeover in 1941. The former steamers were either sabotaged or pressed into Japanese or Chinese service. Probably the most curious incident involved in 1949 during the Chinese Civil War between Kuomintang and People's Liberation Army forces; and led to the award of the Dickin Medal to the ship's cat Simon.
In August 2019, Welsh adventurer Ash Dykes became the first person to complete the 4,000-mile (6,437 km) trek along the course of the river, walking for 352 days from its source to its mouth.
Tens of millions of people live in the floodplain of the Yangtze valley, an area that naturally floods every summer and is habitable only because it is protected by river dikes. The floods large enough to overflow the dikes have caused great distress to those who live and farm there. Floods of note include those of 1931, 1954, and 1998.
The 1931 Central China floods or the Central China floods of 1931 were a series of floods that occurred in the Republic of China. The floods are generally considered among the deadliest natural disasters ever recorded, and almost certainly the deadliest of the 20th century (when pandemics and famines are discounted). Estimates of the total death toll range from 145,000 to between 3.7 million and 4 million. The Yangtze again flooded in 1935, causing great loss of life.
From June to September 1954, the Yangtze River Floods were a series of catastrophic floodings that occurred mostly in Hubei Province. Due to unusually high volume of precipitation as well as an extraordinarily long rainy season in the middle stretch of the Yangtze River late in the spring of 1954, the river started to rise above its usual level in around late June. Despite efforts to open three important flood gates to alleviate the rising water by diverting it, the flood level continued to rise until it hit the historic high of 44.67 m in Jingzhou, Hubei and 29.73 m in Wuhan. The number of dead from this flood was estimated at around 33,000, including those who died of plague in the aftermath of the disaster.
The 1998 Yangtze River floods were a series of major floods that lasted from middle of June to the beginning of September 1998 along the Yangtze.
In the summer of 1998, China experienced massive flooding of parts of the Yangtze River, resulting in 3,704 dead, 15 million homeless and $26 billion in economic loss. Other sources report a total loss of 4150 people, and 180 million people were affected. A staggering were evacuated, 13.3 million houses were damaged or destroyed. The floods caused $26 billion in damages.
The 2016 China floods caused US$22 billion in damages.
Beginning in the 1950s dams and thousands of kilometres of dikes were built for flood control, land reclamation, irrigation and for the control of diseases vectors such as blood flukes that caused Schistosomiasis. More than a hundred lakes were thus cut off from the main river. There were gates between the lakes that could be opened during floods. However, farmers and settlements encroached on the land next to the lakes although it was forbidden to settle there. When floods came, it proved impossible to open the gates since it would have caused substantial destruction. Thus the lakes partially or completely dried up. For example, Baidang Lake shrunk from in the 1950s to in 2005. Zhangdu Lake dwindled to one quarter of its original size. Natural fisheries output in the two lakes declined sharply. Only a few large lakes, such as Poyang Lake and Dongting Lake, remained connected to the Yangtze. Cutting off the other lakes that had served as natural buffers for floods increased the damage done by floods further downstream. Furthermore, the natural flow of migratory fish was obstructed and biodiversity across the whole basin decreased dramatically. Intensive farming of fish in ponds spread using one type of carp who thrived in eutrophic water conditions and who feeds on algae, causing widespread pollution. The pollution was exacerbated by the discharge of waste from pig farms as well as of untreated industrial and municipal sewage. In September 2012, the Yangtze river near Chongqing turned red from pollution. The erection of the Three Gorges Dam has created an impassable "iron barrier" that has led to a great reduction in the biodiversity of the river. Yangtze sturgeon use seasonal changes in the flow of the river to signal when is it time to migrate. However, these seasonal changes will be greatly reduced by dams and diversions. Other animals facing immediate threat of extinction are the baiji dolphin, narrow-ridged finless porpoise and the Yangtze alligator. These animals numbers went into freefall from the combined effects of accidental catches during fishing, river traffic, habitat loss and pollution. In 2006 the baiji dolphin became extinct; the world lost an entire genus.
The Yangtze River produces more ocean plastic pollution than any other, according to The Ocean Cleanup, a Dutch environmental research foundation that focuses on ocean pollution. Together with 9 other rivers, the Yangtze transports 90% of all the plastic that reaches the oceans.
In 2002 a pilot program was initiated to reconnect lakes to the Yangtze with the objective to increase biodiversity and to alleviate flooding. The first lakes to be reconnected in 2004 were Zhangdu Lake, Honghu Lake, and Tian'e-Zhou in Hubei on the middle Yangtze. In 2005 Baidang Lake in Anhui was also reconnected.
Reconnecting the lakes improved water quality and fish were able to migrate from the river into the lake, replenishing their numbers and genetic stock. The trial also showed that reconnecting the lake reduced flooding. The new approach also benefitted the farmers economically. Pond farmers switched to natural fish feed, which helped them breed better-quality fish that can be sold for more, increasing their income by 30%. Based on the successful pilot project, other provincial governments emulated the experience and also reestablished connections to lakes that had previously been cut off from the river. In 2005 a Yangtze Forum has been established bringing together 13 riparian provincial governments to manage the river from source to sea. In 2006 China's Ministry of Agriculture made it a national policy to reconnect the Yangtze River with its lakes. As of 2010, provincial governments in five provinces and Shanghai set up a network of 40 effective protected areas, covering . As a result, populations of 47 threatened species increased, including the critically endangered Yangtze alligator. In the Shanghai area, reestablished wetlands now protect drinking water sources for the city. It is envisaged to extend the network throughout the entire Yangtze to eventually cover 102 areas and . The mayor of Wuhan announced that six huge, stagnating urban lakes including the East Lake (Wuhan) would be reconnected at the cost of US$2.3 billion creating China's largest urban wetland landscape.
Until 1957, there were no bridges across the Yangtze River from Yibin to Shanghai. For millennia, travelers crossed the river by ferry. On occasions, the crossing may have been dangerous, as evidenced by the "Zhong’anlun" disaster (October 15, 1945).
The river stood as a major geographic barrier dividing northern and southern China. In the first half of the 20th century, rail passengers from Beijing to Guangzhou and Shanghai had to disembark, respectively, at Hanyang and Pukou, and cross the river by steam ferry before resuming journeys by train from Wuchang or Nanjing West.
After the founding of the People's Republic in 1949, Soviet engineers assisted in the design and construction of the Wuhan Yangtze River Bridge, a dual-use road-rail bridge, built from 1955 to 1957. It was the first bridge across the Yangtze River. The second bridge across the river that was built was a single-track railway bridge built upstream in Chongqing in 1959. The Nanjing Yangtze River Bridge, also a road-rail bridge, was the first bridge to cross the lower reaches of the Yangtze, in Nanjing. It was built after the Sino-Soviet Split and did not receive foreign assistance. Road-rail bridges were then built in Zhicheng (1971) and Chongqing (1980).
Bridge-building slowed in the 1980s before resuming in the 1990s and accelerating in the first decade of the 21st century. The Jiujiang Yangtze River Bridge was built in 1992 as part of the Beijing-Jiujiang Railway. A second bridge in Wuhan was completed in 1995. By 2005, there were a total of 56 bridges and one tunnel across the Yangtze River between Yibin and Shanghai. These include some of the longest suspension and cable-stayed bridges in the world on the Yangtze Delta: Jiangyin Suspension Bridge (1,385 m, opened in 1999), Runyang Bridge (1,490 m, opened 2005), Sutong Bridge (1,088 m, opened 2008). The rapid pace of bridge construction has continued. The city of Wuhan now has six bridges and one tunnel across the Yangtze.
A number of power line crossings have also been built across the river.
As of 2007, there are two dams built on the Yangtze river: Three Gorges Dam and Gezhouba Dam.
The Three Gorges Dam is the largest power station in the world by installed capacity, at 22.5 GW.
Several dams are operating or are being constructed on the upper portion of the river, the Jinsha River.
Among them, the Xiluodu Dam is the third largest power stations in the world, and the Baihetan Dam, planned to be commissioned in 2021, will be the second largest after the Three Gorges Dam.
The Yangtze River has over 700 tributaries. The major tributaries (listed from upstream to downstream) with the locations of where they join the Yangtze are:
The Huai River flowed into the Yellow Sea until the 20th century, but now primarily discharges into the Yangtze.
The Yangtze River has a high species richness, including many endemics. A high percentage of these are seriously threatened by human activities.
, 416 fish species are known from the Yangtze basin, including 362 that strictly are freshwater species. The remaining are also known from salt or brackish waters, such as the river's estuary or the East China Sea. This makes it one of the most species-rich rivers in Asia and by far the most species-rich in China (in comparison, the Pearl River has almost 300 fish species and the Yellow River 160). 178 fish species are endemic to the Yangtze River Basin. Many are only found in some section of the river basin and especially the upper reach (above Yichang, but below the headwaters in the Qinghai-Tibet Plateau) is rich with 279 species, including 147 Yangtze endemics and 97 strict endemics (found only in this part of the basin). In contrast, the headwaters, where the average altitude is above , are only home to 14 highly specialized species, but 8 of these are endemic to the river. The largest orders in the Yangtze are Cypriniformes (280 species, including 150 endemics), Siluriformes (40 species, including 20 endemics), Perciformes (50 species, including 4 endemics), Tetraodontiformes (12 species, including 1 endemic) and Osmeriformes (8 species, including 1 endemic). No other order has more than four species in the river and one endemic.
Many Yangtze fish species have declined drastically and 65 were recognized as threatened in the 2009 Chinese red list. Among these are three that are considered entirely extinct (Chinese paddlefish, "Anabarilius liui liui" and "Atrilinea macrolepis"), two that are extinct in the wild ("Anabarilius polylepis", "Schizothorax parvus"), four that are critically endangered "Euchiloglanis kishinouyei", "Megalobrama elongata", "Schizothorax longibarbus" and "Leiocassis longibarbus"). Additionally, both the Yangtze sturgeon and Chinese sturgeon are considered critically endangered by the IUCN. The survival of these two sturgeon may rely on the continued release of captive bred specimens. Although still listed as critically endangered rather than extinct by both the Chinese red list and IUCN, recent reviews have found that the Chinese paddlefish is extinct. Surveys conducted between 2006 and 2008 by ichthyologists failed to catch any, but two probable specimens were recorded with hydroacoustic signals. The last definite record was an individual that was accidentally captured near Yibin in 2003 and released after having been radio tagged. The Chinese sturgeon is the largest fish in the river and among the largest freshwater fish in the world, reaching a length of ; the extinct Chinese paddlefish reputedly reached as much as , but its maximum size is labeled with considerable uncertainty.
The largest threats to the Yangtze native fish are overfishing and habitat loss (such as building of dams and land reclamation), but pollution, destructive fishing practices (such as fishing with dynamite or poison) and introduced species also cause problems. About of the total freshwater fisheries in China are in the Yangtze Basin, but a drastic decline in size of several important species has been recorded, as highlighted by data from lakes in the river basin. In 2015, some experts recommend a 10-year fishing moratorium to allow the remaining populations to recover, and in January 2020 China imposed a 10-year fishing moratorium on 332 sites along the Yangtze. Dams present another serious problem, as several species in the river perform breeding migrations and most of these are non-jumpers, meaning that normal fish ladders designed for salmon are ineffective. For example, the Gezhouba Dam blocked the migration of the paddlerfish and two sturgeon, while also effectively splitting the Chinese high fin banded shark population into two and causing the extirpation of the Yangtze population of the Japanese eel. In an attempt of minimizing the effect of the dams, the Three Gorges Dam has released water to mimic the (pre-dam) natural flooding and trigger the breeding of carp species downstream. In addition to dams already built in the Yangtze basin, several large dams are planned and these may present further problems for the native fauna.
While many fish species native to the Yangtze are seriously threatened, others have become important in fish farming and introduced widely outside their native range. A total of 26 native fish species of the Yangtze basin are farmed. Among the most important are four Asian carp: grass carp, black carp, silver carp and bighead carp. Other species that support important fisheries include northern snakehead, Chinese perch, "Takifugu" pufferfish (mainly in the lowermost sections) and predatory carp.
Due to commercial use of the river, tourism, and pollution, the Yangtze is home to several seriously threatened species of large animals (in addition to fish): the narrow-ridged finless porpoise, baiji (Yangtze river dolphin), Chinese alligator, Yangtze giant softshell turtle and Chinese giant salamander. This is the only other place besides the United States that is native to an alligator and paddlefish species. In 2010, the Yangtze population of finless porpoise was 1000 individuals. In December 2006, the Yangtze river dolphin was declared functionally extinct after an extensive search of the river revealed no signs of the dolphin's inhabitance. In 2007, a large, white animal was sighted and photographed in the lower Yangtze and was tentatively presumed to be a "baiji". However, as there have been no confirmed sightings since 2004, the "baiji" is presumed to be functionally extinct at this time. "Baijis were the last surviving species of a large lineage dating back seventy million years and one of only six species of freshwater dolphins." It has been argued that the extinction of the Yangtze river dolphin was a result of the completion of the Three Gorges Dam, a project that has affected many species of animals and plant life found only in the gorges area.
Numerous species of land mammals are found in the Yangtze valley, but most of these are not directly associated with the river. Three exceptions are the semi-aquatic Eurasian otter, water deer and Père David's deer.
In addition to the very large and exceptionally rare Yangtze giant softshell turtle, several smaller turtle species are found in the Yangtze basin, its delta and valleys. These include the Chinese box turtle, yellow-headed box turtle, Pan's box turtle, Yunnan box turtle, yellow pond turtle, Chinese pond turtle, Chinese stripe-necked turtle and Chinese softshell turtle, which all are considered threatened.
More than 160 amphibian species are known from the Yangtze basin, including the world's largest, the critically endangered Chinese giant salamander. It has declined drastically due to hunting (it is considered a delicacy), habitat loss and pollution. The polluted Dian Lake, which is part of the upper Yangtze watershed (via Pudu River), is home to several highly threatened fish, but was also home to the Yunnan lake newt. This newt has not been seen since 1979 and is considered extinct. In contrast, the Chinese fire belly newt from the lower Yangtze basin is one of the few Chinese salamander species to remain common and it is considered least concern by the IUCN.
The Yangtze basin contains a large number of freshwater crab species, including several endemics. A particularly rich genus in the river basin is the potamid "Sinopotamon". The Chinese mitten crab is catadromous (migrates between fresh and saltwater) and it has been recorded up to up the Yangtze, which is the largest river in its native range. It is a commercially important species in its native range where it is farmed, but the Chinese mitten crab has also been spread to Europe and North America where considered invasive.
The freshwater jellyfish "Craspedacusta sowerbii", now an invasive species in large parts of the world, originates from the Yangtze. | https://en.wikipedia.org/wiki?curid=6613 |
Chrono Trigger
"Chrono Trigger" was a critical and commercial success upon release, and is frequently cited as one of the greatest video games of all time. "Nintendo Power" magazine described aspects of "Chrono Trigger" as revolutionary, including its multiple endings, plot-related side-quests focusing on character development, unique battle system, and detailed graphics. "Chrono Trigger" was the third best-selling game of 1995 in Japan, and shipped 2.65 million copies worldwide by March 2003.
Square released a ported version by Tose in Japan for the PlayStation in 1999, which was later repackaged with a "Final Fantasy IV" port as "Final Fantasy Chronicles" (2001) for the North American market. A slightly enhanced "Chrono Trigger", again ported by Tose, was released for the Nintendo DS in North America and Japan in 2008, and PAL regions in 2009. The Nintendo DS version sold 790,000 copies by March 2009, after about a year of sales. "Chrono Trigger" has also been ported to i-mode mobile phones, the Virtual Console, the PlayStation Network, iOS devices, Android devices, and Microsoft Windows.
"Chrono Trigger" features standard role-playing video game gameplay. The player controls the protagonist and his companions in the game's two-dimensional world, consisting of various forests, cities, and dungeons. Navigation occurs via an overworld map, depicting the landscape from a scaled-down overhead view. Areas such as forests, cities, and similar places are depicted as more realistic scaled-down maps, in which players can converse with locals to procure items and services, solve puzzles and challenges, or encounter enemies. "Chrono Trigger" gameplay deviates from that of traditional Japanese RPGs in that, rather than appearing in random encounters, many enemies are openly visible on field maps or lie in wait to ambush the party. Contact with enemies on a field map initiates a battle that occurs directly on the map rather than on a separate battle screen.
Players and enemies may use physical or magical attacks to wound targets during battle, and players may use items to heal or protect themselves. Each character and enemy has a certain number of hit points; successful attacks reduce that character's hit points, which can be restored with potions and spells. When a playable character loses all hit points, they faint; if all the player's characters fall in battle, the game ends and must be restored from a previously saved chapter, except in specific storyline-related battles that allow or force the player to lose. Between battles, a player can equip their characters with weapons, armor, helmets, and accessories that provide special effects (such as increased attack power or defense against magic), and various consumable items can be used both in and out of battles. Items and equipment can be purchased in shops or found on field maps, often in treasure chests. By exploring new areas and fighting enemies, players progress through "Chrono Trigger" story.
"Chrono Trigger" uses an "Active Time Battle" system—a recurring element of Square's "Final Fantasy" game series designed by Hiroyuki Ito for "Final Fantasy IV"—named "Active Time Battle 2.0." Each character can take action in battle once a personal timer dependent on the character's speed statistic counts to zero. Magic and special physical techniques are handled through a system called "Techs." Techs deplete a character's magic points (a numerical meter similar to hit points), and often have special areas of effect; some spells damage huddled monsters, while others can harm enemies spread in a line. Enemies often change positions during battle, creating opportunities for tactical Tech use. A unique feature of "Chrono Trigger"'s Tech system is that numerous cooperative techniques exist. Each character receives eight personal Techs which can be used in conjunction with others' to create Double and Triple Techs for greater effect. For instance, Crono's sword-spinning "Cyclone" Tech can be combined with Lucca's "Flame Toss" to create "Flame Whirl". When characters with compatible Techs have enough magic points available to perform their techniques, the game automatically displays the combo as an option.
"Chrono Trigger" features several other distinct gameplay traits, including time travel. Players have access to seven eras of the game world's history, and past actions affect future events. Throughout history, players find new allies, complete side quests, and search for keynote villains. Time travel is accomplished via portals and pillars of light called "time gates", as well as a time machine named "Epoch". The game contains thirteen unique endings; the ending the player receives depends on when and how they reach and complete the game's final battle. "Chrono Trigger DS" features a new ending that can be accessed from the End of Time upon completion of the final extra dungeon and optional final boss. "Chrono Trigger" also introduces a New Game Plus option; after completing the game, the player may begin a new game with the same character levels, techniques, and equipment, excluding money, with which they ended the previous playthrough. However, certain items central to the storyline are removed and must be found again, such as the sword Masamune. Square has employed the New Game Plus concept in later games including "Chrono Cross", "Parasite Eve", "Vagrant Story", "Final Fantasy X-2", and "".
"Chrono Trigger" takes place in a world similar to Earth, with eras such as the prehistoric age, in which primitive humans and dinosaurs share the earth; the Middle Ages, replete with knights, monsters, and magic; and the post-apocalyptic future, where destitute humans and sentient robots struggle to survive. The characters frequently travel through time to obtain allies, gather equipment, and learn information to help them in their quest. The party also gains access to the End of Time (represented as year ∞), which serves as a hub to travel back to other time periods. The party eventually acquires a time-machine vehicle known as the "Wings of Time," nicknamed the "Epoch." The vehicle is capable of time travel between any time period without first having to travel to the End of Time.
"Chrono Trigger"'s six playable characters (plus one optional character) come from different eras of history. "Chrono Trigger" begins in AD 1000 with Crono, Marle, and Lucca. Crono is the silent protagonist, characterized as a fearless young man who wields a katana in battle. Marle, revealed to be Princess Nadia, lives in Guardia Castle; though sheltered, at heart, she's a princess who seeks independence from her royal identity. Lucca is a childhood friend of Crono's and a mechanical genius; her home is filled with laboratory equipment and machinery. From the era of AD 2300 comes Robo, or Prometheus (designation R-66Y), a robot with a near-human personality created to assist humans. Lying dormant in the future, Robo is found and repaired by Lucca, and joins the group out of gratitude. The fiercely confident Ayla dwells in 65,000,000 BC. Unmatched in raw strength, Ayla is the chief of Ioka Village and leads her people in war against a species of humanoid reptiles known as Reptites.
The last two playable characters are Frog and Magus. Frog originated in AD 600. He is a former squire once known as Glenn, who was turned into an anthropomorphic frog by Magus, who also killed his friend Cyrus. Chivalrous but mired in regret, Frog dedicates his life to protecting Leene, the queen of Guardia, and avenging Cyrus. Meanwhile, Guardia in AD 600 is in a state of conflict against the Mystics (known as Fiends in the US/DS port), a race of demons and intelligent animals who wage war against humanity under the leadership of Magus, a powerful sorcerer. Magus's seclusion conceals a long-lost past; he was formerly known as Janus, the young prince of the Kingdom of Zeal, which was destroyed by Lavos in 12,000 BC. The incident sent him forward through time, and as he ages, he plots revenge against Lavos and broods over the fate of his sister, Schala. Lavos, the game's main antagonist who awakens and ravages the world in AD 1999, is an extraterrestrial, parasitic creature that harvests DNA and the Earth's energy for its own growth.
In AD 1000, Crono and Marle watch Lucca and her father demonstrate her new teleporter at the Millennial Fair in the Kingdom of Guardia. When Marle volunteers to be teleported, her pendant interferes with the device and creates a time portal into which she is drawn. After Crono and Lucca separately recreate the portal and find themselves in AD 600, they find Marle only to see her vanish before their eyes. Lucca realizes that this time period's kingdom has mistaken Marle— actually Princess Nadia of Guardia— for Queen Leene, an ancestor of hers who had been kidnapped, thus putting off the recovery effort for her ancestor and creating a grandfather paradox. Crono and Lucca, with the help of Frog, restore history to normal by rescuing Leene. After the three part ways with Frog and return to the present, Crono is arrested on charges of kidnapping Marle and sentenced to death by the current chancellor of Guardia. Lucca and Marle help Crono to flee, haphazardly using another time portal to escape their pursuers. This portal lands them in AD 2300, where they learn that an advanced civilization has been wiped out by a giant creature known as Lavos that appeared in 1999. The three vow to find a way to prevent the future destruction of their world. After meeting and repairing Robo, Crono and his friends find Gaspar, an old sage at the End of Time, who helps them acquire magical powers and travel through time by way of several pillars of light (from this point onwards, the party is able to challenge Lavos; doing so and emerging victorious will unlock one of twelve different endings, depending on the point in the story at which Lavos is fought, as well as whether or not certain actions are completed beforehand).
Research in AD 1000 tells the party about Magus summoning Lavos into the world in 600 AD, but Frog needs to reforge the legendary sword, Masamune, to open the way to Magus's castle. To get the ore to reforge it, they travel to prehistoric times and meet Ayla. After returning to AD 600, they challenge Magus, believing him to be the source of Lavos; the battle disrupts the spell Magus was using to summon Lavos and causes a time gate that throws Crono and his friends to the past. Back in prehistory, Ayla joins the group to battle the Reptites and witness the true origin of Lavos when the Reptites' ruler calls him down from the stars to crash into the planet. Entering a gate created by Lavos's impact, they go to BC 12,000, an ice age, where the party finds the Kingdom of Zeal, who recently discovered Lavos and seeks to drain its power to achieve immortality through the Mammon Machine. Zeal's leader, Queen Zeal, imprisons Crono and friends on orders of the Prophet, a mysterious figure who has recently begun advising the queen. Though Zeal's daughter Schala frees them, the Prophet forces her to banish them from the realm and seal the time gate they used to travel to the Dark Ages. They return next to AD 2300 to find a time machine called the Wings of Time (or "Epoch"), which can access any time period without using a time gate. They travel back to BC 12000 but fail to stop Zeal from activating the Mammon Machine in the Ocean Palace. Lavos awakens, disturbed by the Mammon Machine; the Prophet reveals himself to be Magus, and unsuccessfully tries to kill the creature. The party is beaten, and a broken Crono stands up to Lavos before being vaporized by a powerful blast. Schala transports the rest of the party back to the surface before Lavos destroys the Ocean palace and the Kingdom of Zeal, causing a devastating flood that submerges most of the world.
Crono's friends awaken in a village and find Magus, who confesses that he used to be Prince Janus of Zeal. In his memories, it is revealed that the disaster at the Ocean Palace scattered the Gurus of Zeal across time and sent him to the Middle Ages. Janus took the title of Magus and gained a cult of followers while plotting to summon and kill Lavos in revenge for the death of his sister, Schala. When Lavos appeared after his battle with Crono and his allies, he was cast back to the time of Zeal and presented himself to them as a prophet. At this point, Magus is either killed by the party, killed in a duel with Frog, or spared and convinced to join the party. As Crono's friends depart, the ruined Ocean Palace rises into the air as the Black Omen. The group turns to Gaspar for help, and he gives them a "Chrono Trigger," an egg-shaped device that allows the group to replace Crono just before the moment of death with a Dopple Doll (doing so is optional, and the game's ending will change depending on the player's decision in this matter). The party then gather power by helping people across time with Gaspar's instructions. Their journeys involve defeating the remnants of the Mystics, stopping Robo's maniacal AI creator, giving Frog closure for Cyrus' death, locating and charging up the mythical Sun Stone, retrieving the legendary Rainbow Shell, unmasking Guardia's Chancellor as a monster, helping restore a forest destroyed by a desert monster, and preventing an accident that crippled Lucca's mother. The group enters the Black Omen and defeats Queen Zeal, then battles Lavos; once faced with the creature's true form, they come to understand that it had been absorbed DNA and energy from every living creature before arising and razing the planet's surface in 1999 so that it could spawn a new generation to destroy other world. After slaying Lavos and saving the future, they celebrated at the final night of the Millennial Fair before returning to their own times.
If Magus joined the party, he departs to search for Schala. If Crono was resurrected before defeating Lavos, his sentence for kidnapping Marle is revoked by her father, King Guardia XXXIII, thanks to testimonies from Marle's ancestors and descendants, whom Crono had helped during his journey. Crono's mother accidentally enters the time gate at the Millennial Fair before it closes, prompting Crono, Marle, and Lucca to set out in the Epoch to find her while fireworks light up the night sky. If Crono was not resurrected, Frog, Robo, and Ayla (along with Magus if he was recruited) chase Gaspar to the Millennial Fair and back again, revealing that Gaspar knows how to resurrect Crono; Marle and Lucca then use the Epoch to travel through time and figure out how to achieve this. Alternatively, if the party used the Epoch to break Lavos's outer shell, Marle will help her father hang Nadia's bell at the festival and accidentally get carried away by several balloons. If resurrected, Crono jumps on to help her, but cannot bring them down to earth. Hanging on in each other's arms, the pair travel through the cloudy, moonlit sky.
"Chrono Trigger DS" added two new scenarios to the game In the first, Crono and his friends can help a "lost sanctum" of Reptites, who reward powerful items and armor. The second scenario adds ties to "Trigger"'s sequel, "Chrono Cross". In a New Game +, the group can explore several temporal distortions to combat shadow versions of Crono, Marle, and Lucca, and to fight Dalton, who promises in defeat to raise an army in the town of Porre to destroy the Kingdom of Guardia. The group can then fight the Dream Devourer, a prototypical form of the "Time Devourer"—a fusion of Schala and Lavos seen in "Chrono Cross". A version of Magus pleads with Schala to resist; though she recognizes him as her brother, she refuses to be helped and sends him away. Schala subsequently erases his memories and Magus awakens in a forest, determined to find what he had lost.
"Chrono Trigger" was conceived in 1992 by Hironobu Sakaguchi, producer and creator of the "Final Fantasy" series; Yuji Horii, writer, game designer and creator of the "Dragon Quest" series; and Akira Toriyama, character designer of "Dragon Quest" and creator of the "Dragon Ball" manga series. Traveling to the United States to research computer graphics, the three decided to create something that "no one had done before". After spending over a year considering the difficulties of developing a new game, they received a call from Kazuhiko Aoki, who offered to produce. The four met and spent four days brainstorming ideas for the game. Square convened 50–60 developers, including scenario writer Masato Kato, whom Square designated story planner. Development started in early 1993. An uncredited Square employee suggested that the team develop a time travel-themed game, which Kato initially opposed, fearing repetitive, dull gameplay. Kato and Horii then met several hours per day during the first year of development to write the game's plot. Square intended to license the work under the "Seiken Densetsu" franchise and gave it the working title "Maru Island"; Hiromichi Tanaka (the future producer of "Chrono Cross") monitored Toriyama's early designs. The team hoped to release it on Nintendo's planned Super Famicom Disk Drive; when Nintendo canceled the project, Square reoriented the game for release on a Super Famicom cartridge and rebranded it as "Chrono Trigger". Tanaka credited the ROM cartridge platform for enabling seamless transition to battles on the field map.
Aoki ultimately produced "Chrono Trigger", while director credits were attributed to Akihiko Matsui, Yoshinori Kitase and Takashi Tokita. Toriyama designed the game's aesthetic, including characters, monsters, vehicles, and the look of each era. Masato Kato also contributed character ideas and designs. Kato planned to feature Gaspar as a playable character and Toriyama sketched him, but he was cut early in development. The development staff studied the drawings of Toriyama to approximate his style. Sakaguchi and Horii supervised; Sakaguchi was responsible for the game's overall system and contributed several monster ideas. Other notable designers include Tetsuya Takahashi, the graphic director, and Yasuyuki Honne, Tetsuya Nomura, and Yusuke Naora, who worked as field graphic artists. Yasuhiko Kamata programmed graphics, and cited Ridley Scott's visual work in the film "Alien" as an inspiration for the game's lighting. Kamata made the game's luminosity and color choice lay between that of "Secret of Mana" and the "Final Fantasy" series. Features originally intended to be used in "Secret of Mana" or "Final Fantasy IV", also under development at the same time, were appropriated by the "Chrono Trigger" team. According to Tanaka, "Secret of Mana" (which itself was originally intended to be "Final Fantasy IV") was codenamed ""Chrono Trigger"" during development before being called "Seiken Densetsu 2" ("Secret of Mana"), and then the name "Chrono Trigger" was adopted for a new project.
Yuji Horii, a fan of time travel fiction (such as the TV series "The Time Tunnel"), fostered a theme of time travel in his general story outline of "Chrono Trigger" with input from Akira Toriyama. Horii liked the scenario of the grandfather paradox surrounding Marle. Concerning story planning, Horii commented, "If there's a fairground, I just write that there's a fairground; I don't write down any of the details. Then the staff brainstorm and come up with a variety of attractions to put in." Sakaguchi contributed some minor elements, including the character Gato; he liked Marle's drama and reconciliation with her father. Masato Kato subsequently edited and completed the outline by writing the majority of the game's story, including all the events of the 12,000 BC era. He took pains to avoid what he described as "a long string of errands ... [such as] 'do this', 'take this', 'defeat these monsters', or 'plant this flag'." Kato and other developers held a series of meetings to ensure continuity, usually attended by around 30 personnel. Kato and Horii initially proposed Crono's death, though they intended he stay dead; the party would have retrieved an earlier, living version of him to complete the quest. Square deemed the scenario too depressing and asked that Crono be brought back to life later in the story. Kato also devised the system of multiple endings because he could not branch the story out to different paths. Yoshinori Kitase and Takashi Tokita then wrote various subplots. They also devised an "Active Time Event Logic" system, "where you can move your character around during scenes, even when an NPC is talking to you", and with players "talking to different people and steering the conversation in different directions", allowing each scene to "have many permutations." Kato became friends with composer Yasunori Mitsuda during development, and they would collaborate on several future projects. Katsuhisa Higuchi programmed the battle system, which hosted combat on the map without transition to a special battleground as most previous Square games had done. Higuchi noted extreme difficulty in loading battles properly without slow-downs or a brief, black loading screen. The game's use of animated monster sprites consumed much more memory than previous "Final Fantasy" games, which used static enemy graphics.
Hironobu Sakaguchi likened the development of "Chrono Trigger" to "play[ing] around with Toriyama's universe," citing the inclusion of humorous sequences in the game that would have been "impossible with something like "Final Fantasy"." When Square Co. suggested a non-human player character, developers created Frog by adapting one of Toriyama's sketches. The team created the End of Time to help players with hints, worrying that they might become stuck and need to consult a walkthrough. The game's testers had previously complained that "Chrono Trigger" was too difficult; as Horii explained, "It's because we know too much. The developers think the game's just right; that they're being too soft. They're thinking from their own experience. The puzzles were the same. Lots of players didn't figure out things we thought they'd get easily." Sakaguchi later cited the unusual desire of beta testers to play the game a second time or "travel through time again" as an affirmation of the New Game + feature: "Wherever we could, we tried to make it so that a slight change in your behavior caused subtle differences in people's reactions, even down to the smallest details ... I think the second playthrough will hold a whole new interest." The game's reuse of locations due to time traveling made bug-fixing difficult, as corrections would cause unintended consequences in other eras.
"Chrono Trigger" was scored primarily by Yasunori Mitsuda, with contributions from veteran "Final Fantasy" composer Nobuo Uematsu, and one track composed by Noriko Matsueda. A sound programmer at the time, Mitsuda was unhappy with his pay and threatened to leave Square if he could not compose music. Hironobu Sakaguchi suggested he score "Chrono Trigger", remarking, "maybe your salary will go up." Mitsuda composed new music and drew on a personal collection of pieces composed over the previous two years. He reflected, "I wanted to create music that wouldn't fit into any established genre ... music of an imaginary world. The game's director, Masato Kato, was my close friend, and so I'd always talk with him about the setting and the scene before going into writing." Mitsuda slept in his studio several nights, and attributed certain pieces—such as the game's ending theme, "To Far Away Times"—to inspiring dreams. He later attributed this song to an idea he was developing before "Chrono Trigger", reflecting that the tune was made in dedication to "a certain person with whom [he] wanted to share a generation". He also tried to use leitmotifs of the "Chrono Trigger" main theme to create a sense of consistency in the soundtrack. Mitsuda wrote each tune to be around two minutes long before repeating, unusual for Square's games at the time. Mitsuda suffered a hard drive crash that lost around forty in-progress tracks. After Mitsuda contracted stomach ulcers, Uematsu joined the project to compose ten pieces and finish the score. Mitsuda returned to watch the ending with the staff before the game's release, crying upon seeing the finished scene.
At the time of the game's release, the number of tracks and sound effects was unprecedented—the soundtrack spanned three discs in its 1995 commercial pressing. Square also released a one-disc acid jazz arrangement called ""The Brink of Time"" by Guido that year. "The Brink of Time" came about because Mitsuda wanted to do something that no one else was doing, and he noted that acid jazz and its related genres were uncommon in the Japanese market. Mitsuda considers "Chrono Trigger" a landmark game which helped mature his talent. While Mitsuda later held that the title piece was "rough around the edges", he maintains that it had "significant influence on [his] life as a composer". In 1999, Square produced another one-disc soundtrack to complement the PlayStation release of "Trigger", featuring orchestral tracks used in cut scenes. Tsuyoshi Sekito composed four new pieces for the game's bonus features which weren't included on the soundtrack. Some fans were displeased by Mitsuda's absence in creating the port, whose instruments sometimes aurally differed from the original game's. Mitsuda arranged versions of music from the "Chrono" series for Play! video game music concerts, presenting the main theme, "Frog's Theme", and "To Far Away Times". He worked with Square Enix to ensure that the music for the Nintendo DS would sound closer to the Super NES version. Mitsuda encouraged feedback about the game's soundtrack from contemporary children (who he thought would expect "full symphonic scores blaring out of the speakers"). Fans who preordered "Chrono Trigger DS" received a special music disc containing two orchestral arrangements of "Chrono Trigger" music directed by Natsumi Kameoka; Square Enix also held a random prize drawing for two signed copies of "Chrono Trigger" sheet music. Mitsuda expressed difficulty in selecting the tune for the orchestral medley, eventually picking a tune from each era and certain character themes. Mitsuda later wrote:
Music from the game was performed live by the Tokyo Symphony Orchestra in 1996 at the Orchestral Game Concert in Tokyo, Japan. A suite of music including "Chrono Trigger" is a part of the symphonic world-tour with video game music Play! A Video Game Symphony, where Mitsuda was in attendance for the concert's world-premiere in Chicago on May 27, 2006. His suite of "Chrono" music, comprising "Reminiscence", "Chrono Trigger", "Chrono Cross~Time's Scar", "Frog's Theme", and "To Far Away Times" was performed. Mitsuda has also appeared with the Eminence Symphony Orchestra as a special guest. Video Games Live has also featured medleys from "Chrono Trigger" and "Chrono Cross". A medley of Music from "Chrono Trigger" made of one of the four suites of the "Symphonic Fantasies" concerts in September 2009 which was produced by the creators of the Symphonic Game Music Concert series, conducted by Arnie Roth. Square Enix re-released the game's soundtrack, along with a video interview with Mitsuda in July 2009.
The team planned to release "Chrono Trigger" in late 1994, but release was pushed back to the following year. Early alpha versions of "Chrono Trigger" were demonstrated at the 1994 and 1995 V-Jump festivals in Japan. A few months prior to the game's release, Square shipped a beta version to magazine reviewers and game stores for review. An unfinished build of the game dated November 17, 1994, it contains unused music tracks, locations, and other features changed or removed from the final release—such as a dungeon named "Singing Mountain" and its eponymous tune. Some names also differed; the character Soysaw (Slash in the US version) was known as Wiener, while Mayonnay (Flea in the US version) was named Ketchappa. The ROM image for this early version was eventually uploaded to the internet, prompting fans to explore and document the game's differences, including two unused world map NPC character sprites and presumed additional sprites for certain non-player characters. Around the game's release, Yuji Horii commented that "Chrono Trigger" "went beyond [the development team's] expectations", and Hironobu Sakaguchi congratulated the game's graphic artists and field designers. Sakaguchi intended to perfect the "sense of dancing you get from exploring Toriyama's worlds" in the event that they would make a sequel.
"Chrono Trigger" used a 32-megabit ROM cartridge with battery-backed RAM for saved games, lacking special on-cartridge coprocessors. The Japanese release of "Chrono Trigger" included art for the game's ending and running counts of items in the player's status menu. Developers created the North American version before adding these features to the original build, inadvertently leaving in vestiges of "Chrono Trigger"'s early development (such as the piece "Singing Mountain"). Hironobu Sakaguchi asked translator Ted Woolsey to localize "Chrono Trigger" for English audiences and gave him roughly thirty days to work. Lacking the help of a modern translation team, he memorized scenarios and looked at drafts of commercial player's guides to put dialogue in context. Woolsey later reflected that he would have preferred two-and-a-half months, and blames his rushed schedule on the prevailing attitude in Japan that games were children's toys rather than serious works. Some of his work was cut due to space constraints, though he still considered "Trigger" "one of the most satisfying games [he] ever worked on or played". Nintendo of America censored certain dialogue, including references to breastfeeding, consumption of alcohol, and religion.
The original SNES edition of "Chrono Trigger" was released on the Wii download service Virtual Console in Japan on April 26, 2011, in the US on May 16, 2011, and in Europe on May 20, 2011. Previously in April 2008, a "Nintendo Power" reader poll had identified "Chrono Trigger" as the third-most wanted game for the Virtual Console. It went on to receive a perfect score of 10 out 10 on IGN.
Square released an enhanced port of "Chrono Trigger" developed by Tose in Japan for the Sony PlayStation in 1999. Square timed its release before that of "Chrono Cross", the 1999 sequel to "Chrono Trigger", to familiarize new players with story leading up to it. This version included anime cutscenes created by original character designer Akira Toriyama's Bird Studio and animated at Toei Animation, as well as several bonus features, accessible after achieving various endings in the game. Scenarist Masato Kato attended planning meetings at Bird Studio to discuss how the ending cutscenes would illustrate subtle ties to "Chrono Cross". The port was released in North America in 2001—along with a newly translated version of "Final Fantasy IV"—as "Final Fantasy Chronicles". Reviewers criticized "Chronicles" for its lengthy load times and an absence of new in-game features. This same iteration was also re-released as a downloadable game on the PlayStation Network on October 4, 2011, for the PlayStation 3, PlayStation Vita, and PlayStation Portable.
On July 2, 2008, Square Enix announced that they were planning to bring "Chrono Trigger" to the Nintendo DS handheld platform. Composer Yasunori Mitsuda was pleased with the project, exclaiming "finally!" after receiving the news from Square Enix and maintaining, "it's still a very deep, very high-quality game even when you play it today. I'm very interested in seeing what kids today think about it when they play it." Square retained Masato Kato to oversee the port, and Tose to program it. Kato explained, "I wanted it to be based on the original Super NES release rather than the PlayStation version. I thought we should look at the additional elements from the Playstation version, re-examine and re-work them to make it a complete edition. That's how it struck me and I told the staff so later on." Square Enix touted the game by displaying Akira Toriyama's original art at the 2008 Tokyo Game Show.
The DS re-release contains all of the bonus material from the PlayStation port, as well as other enhancements. The added features include a more accurate and revised translation by Tom Slattery, a dual-screen mode which clears the top screen of all menus, a self-completing map screen, and a default "run" option. It also featured the option to choose between two control schemes: one mirroring the original SNES controls, and the other making use of the DS's touch screen. Masato Kato participated in development, overseeing the addition of the monster-battling Arena, two new areas, the Lost Sanctum and the Dimensional Vortex, and a new ending that further foreshadows the events of "Chrono Cross". One of the areas within the Vortex uses the "Singing Mountain" song that was featured on the original "Chrono Trigger" soundtrack. These new dungeons met with mixed reviews; GameSpot called them "frustrating" and "repetitive", while IGN noted that "the extra quests in the game connect extremely well." It was a nominee for "Best RPG for the Nintendo DS" in IGN's 2008 video game awards. The Nintendo DS version of "Chrono Trigger" was the 22nd best-selling game of 2008 in Japan.
A cellphone version was released in Japan on i-mode distribution service on August 25, 2011. An iOS version was released on December 8, 2011. This version is based on the Nintendo DS version, with graphics optimized for iOS. The game was later released for Android on October 29, 2012. An update incorporating most of the features of the Windows version—including the reintroduction of the animated cutscenes, which had been absent from the initial mobile release—was released on February 27, 2018.
Square Enix released "Chrono Trigger" without an announcement for Microsoft Windows via Steam on February 27, 2018. This version includes all content from the Nintendo DS port, the higher resolution graphics from the mobile device releases, support for mouse and keyboard controls, and autosave features, along with additional content such as wallpapers and music. The PC port received negative reception due to its inferior graphical quality, additional glitches, UI adapted for touchscreens, and failure to properly adapt the control scheme for keyboards and controllers. In response, Square Enix provided various UI updates and other improvements over the next few months to address the complaints.
The game was a bestseller in Japan. The game's SNES and PS1 iterations have shipped more than 2.36 million copies in Japan and 290,000 abroad. The first two million copies sold in Japan were delivered in only two months, and the game ended 1995 as the third best-selling game of the year behind "" and "". The game was met with substantial success upon release in North America, and its rerelease on the PlayStation as part of the "Final Fantasy Chronicles" package topped the NPD TRSTS PlayStation sales charts for over six weeks. This version was later re-released again in 2003 as part of Sony's Greatest Hits line. "Chrono Trigger DS" has sold 490,000 copies in Japan, 240,000 in North America and 60,000 in Europe as of March 2009.
"Chrono Trigger" garnered much critical praise in addition to its brisk sales. "Famicom Tsūshin" gave "Chrono Trigger" first an 8 out of 10 and later a 9 out of 10 in their Reader Cross Review. "Nintendo Power" compared it favorably with "Secret of Mana", "Final Fantasy", and "", citing improved graphics, sound, story and gameplay. "GamePro" praised the varied gameplay, the humor, the ability to replay the game with previously built-up characters, and the graphics, which they said far exceed even those of "Final Fantasy VI". They commented that combat is easier and more simplistic than in most RPGs, but argued that "Most players would choose an easier RPG of this caliber over a hundred more complicated, but less developed, fantasy role-playing adventures." They gave the game a perfect 5 out of 5 in all four categories: graphics, sound, control, and funfactor. "Electronic Gaming Monthly" gave it their "Game of the Month" award, with their four reviewers praising the graphics, story, and music. "Chrono Trigger" won multiple awards from "Electronic Gaming Monthly" 1995 video game awards, including Best Role-Playing Game, Best Music in a Cartridge-Based Game, and Best Super NES Game. "Official U.S. PlayStation Magazine" described "Trigger" as "original and extremely captivating", singling out its graphics, sound and story as particularly impressive. IGN commented that "it may be filled with every imaginable console RPG cliché, but "Chrono Trigger" manages to stand out among the pack" with "a [captivating] story that doesn't take itself too serious " and "one of the best videogame soundtracks ever produced". Other reviewers (such as the staff of RPGFan and RPGamer) have criticized the game's short length and relative ease compared to its peers. Victoria Earl of Gamasutra praised the game design for balancing "developer control with player freedom using designed mechanics and a modular approach to narrative."
Overall, critics lauded "Chrono Trigger" for its "fantastic yet not overly complex" story, simple but innovative gameplay, and high replay value afforded by multiple endings. Online score aggregator GameRankings lists the original Super NES version as the 2nd highest scoring RPG and 24th highest scoring game ever reviewed. In 2009, Guinness World Records listed it as the 32nd most influential video game in history. "Nintendo Power" listed the ending to "Chrono Trigger" as one of the greatest endings in Nintendo history, due to over a dozen endings that players can experience. Tom Hall drew inspiration from "Chrono Trigger" and other console games in creating "Anachronox", and used the campfire scene to illustrate the dramatic depth of Japanese RPGs. "Next Generation" reviewed the Super NES version of the game, rating it four stars out of five, and stated that "it [...] easily qualifies as one of the best RPGs ever made".
"Chrono Trigger" is frequently listed among the greatest video games of all time. In 1997 "Electronic Gaming Monthly" ranked it the 29th best console video game of all time; while noting that it was not as good as "Final Fantasy VI" (which ranked 9th), they gave superlative praise to its handling of time travel and its combat engine. It has placed highly on all six of multimedia website IGN's "top 100 games of all time" lists—4th in 2002, 6th in early 2005, 13th in late 2005, 2nd in 2006, 18th in 2007, and 2nd in 2008. "Game Informer" called it its 15th favourite game in 2001. Its staff thought that it was the best non-"Final Fantasy" game Square had produced at the time. GameSpot included "Chrono Trigger" in "The Greatest Games of All Time" list released in April 2006, and it also appeared as 28th on an "All Time Top 100" list in a poll conducted by Japanese magazine "Famitsu" the same year. In 2004, "Chrono Trigger" finished runner up to "Final Fantasy VII" in the inaugural GameFAQs video game battle. In 2008, readers of Dengeki Online voted it the eighth best game ever made. "Nintendo Power"'s twentieth anniversary issue named it the fifth best Super NES game. In 2012, it came 32nd place on GamesRadar's "100 best games of all time" list, and 1st place on its "Best JRPGs" list. GamesRadar named "Chrono Trigger" the 2nd best Super NES game of all time, behind "Super Metroid".
In contrast to the critical acclaim of "Chrono Trigger"'s original SNES release, the 2018 Microsoft Windows port of "Chrono Trigger" was critically panned. Grievances noted by reviewers included tiling errors on textures, the addition of aesthetically-intrusive sprite filters, an unattractive GUI carried over from the 2011 mobile release, a lack of graphic customization options, and the inability to remap controls. In describing the port, "Forbes" commented: "From pretty awful graphical issues, such as tiling textures and quite a painful menu system, this port really doesn’t do this classic game justice." "USGamer" characterized the Windows release as carrying "all the markings of a project farmed out to the lowest bidder. It's a shrug in Square-Enix's mind, seemingly not worth the money or effort necessary for a half-decent port." In a Twitter post detailing his experiences with the Windows version, indie developer Fred Wood derisively compared the port to "someone's first attempt at an RPG Maker game", a comment which was republished across numerous articles addressing the poor quality of the rerelease.
"Chrono Trigger" inspired several related releases; the first were three games released for the Satellaview on July 31, 1995. They included "Chrono Trigger: Jet Bike Special", a racing video game based on a minigame from the original; "Chrono Trigger: Character Library", featuring profiles on characters and monsters from the game; and "Chrono Trigger: Music Library", a collection of music from the game's soundtrack. The contents of "Character Library" and "Music Library" were later included as extras in the PlayStation rerelease of "Chrono Trigger". Production I.G created a 16-minute OVA, """", which was shown at the Japanese V-Jump Festival of July 31, 1996.
There have been two notable attempts by "Chrono Trigger" fans to unofficially remake parts of the game for PC with a 3D graphics engine. "Chrono Resurrection", an attempt at remaking ten small interactive cut scenes from "Chrono Trigger", and "Chrono Trigger Remake Project", which sought to remake the entire game, were forcibly terminated by Square Enix by way of a cease and desist order. Another group of fans created a sequel via a ROM hack of "Chrono Trigger" called ""; developed from 2004 to 2009; although feature-length and virtually finished, it also was terminated through a cease & desist letter days before its May 2009 release. The letter also banned the dissemination of existing "Chrono Trigger" ROM hacks and documentation. After the cease and desist was issued, an incomplete version of the game was leaked in May 2009, though due to the early state of the game, playability was limited.
This was followed by a more complete ROM leak in January 2011, which allowed the game to be played from beginning to end.
Square released a fourth Satellaview game in 1996, named "". Having thought that "Trigger" ended with "unfinished business", scenarist Masato Kato wrote and directed the game. "Dreamers" functioned as a side story to "Chrono Trigger", resolving a loose subplot from its predecessor. A short, text-based game relying on minimal graphics and atmospheric music, the game never received an official release outside Japan—though it was translated by fans to English in April 2003. Square planned to release "Radical Dreamers" as an easter egg in the PlayStation edition of "Chrono Trigger", but Kato was unhappy with his work and halted its inclusion.
Square released "Chrono Cross" for the Sony PlayStation in 1999. "Cross" is a sequel to "Chrono Trigger" featuring a new setting and cast of characters. Presenting a theme of parallel worlds, the story followed the protagonist Serge—a teenage boy thrust into an alternate reality in which he died years earlier. With the help of a thief named Kid, Serge endeavors to discover the truth behind his apparent death and obtain the Frozen Flame, a mythical artifact. Regarded by writer and director Masato Kato as an effort to "redo "Radical Dreamers" properly", "Chrono Cross" borrowed certain themes, scenarios, characters, and settings from "Dreamers". Yasunori Mitsuda also adapted certain songs from "Radical Dreamers" while scoring "Cross". "Radical Dreamers" was consequently removed from the series' main continuity, considered an alternate dimension. "Chrono Cross" shipped 1.5 million copies and was almost universally praised by critics.
There are no plans as of for a new title, despite a statement from Hironobu Sakaguchi in 2001 that the developers of "Chrono Cross" wanted to make a new "Chrono" game. The same year, Square applied for a trademark for the names "Chrono Break" in the United States and "Chrono Brake" in Japan. However, the United States trademark was dropped in 2003. Director Takashi Tokita mentioned ""Chrono Trigger 2"" in a 2003 interview which has not been translated to English. Yuji Horii expressed no interest in returning to the Chrono franchise in 2005, while Hironobu Sakaguchi remarked in April 2007 that his creation "Blue Dragon" was an "extension of [Chrono Trigger]." During a Cubed³ interview on February 1, 2007, Square Enix's Senior Vice President Hiromichi Tanaka said that although no sequel is currently planned, some sort of sequel is still possible if the "Chrono Cross" developers can be reunited. Yasunori Mitsuda has expressed interest in scoring a new game, but warned that "there are a lot of politics involved" with the series. He stressed that Masato Kato should participate in development. The February 2008 issue of "Game Informer" ranked the "Chrono" series eighth among the "Top Ten Sequels in Demand", naming the games "steadfast legacies in the Square Enix catalogue" and asking, "what's the damn holdup?!" In Electronic Gaming Monthly's June 2008 "Retro Issue", writer Jeremy Parish cited "Chrono" as the franchise video game fans would be most thrilled to see a sequel to. In the first May "Famitsu" of 2009, "Chrono Trigger" placed 14th out of 50 in a vote of most-wanted sequels by the magazine's readers. At E3 2009, SE Senior Vice President Shinji Hashimoto remarked, "If people want a sequel, they should buy more!"
In July 2010, Obsidian Entertainment designer Feargus Urquhart, replying to an interview question about what franchises he would like to work on, said that "if [he] could come across everything that [he] played", he would choose a "Chrono Trigger" game. At the time, Obsidian was making "Dungeon Siege III" for Square Enix. Urquhart said: "You make RPGs, we make RPGs, it would be great to see what we could do together. And they really wanted to start getting into Western RPGs. And, so it kind of all ended up fitting together." Yoshinori Kitase stated that he used the time travel mechanics of "Chrono Trigger" as a starting point for that of "Final Fantasy XIII-2". | https://en.wikipedia.org/wiki?curid=6614 |
Cornwall Wildlife Trust
The Cornwall Wildlife Trust is a charitable organisation founded in 1962 that is concerned solely with Cornwall, England.
It deals with the conservation and preservation of Cornwall's wildlife and habitats managing over 50 nature reserves covering approximately , amongst them Looe Island.
Cornwall Wildlife Trust is part of The Wildlife Trusts partnership of 46 wildlife trusts in the United Kingdom. It works in conjunction with the Isles of Scilly Wildlife Trust, and jointly produces a thrice yearly magazine called "Wild Cornwall & Wild Scilly"
The direction and work that the Trust currently does is guided by the Cornwall Biodiversity action plan. Living Seas and Living Landscapes are two such projects. The Trust runs ERCCIS (Environmental Records Centre for Cornwall and the Isles of Scilly), a county wide database of sightings of animals and plants. It also gives planning advice (CEC - Cornwall Environmental Consultants) to land developers.
The Trust is based at Allet near Truro in Cornwall. The headquarters and offices are adjacent to the Trust's Five Acres nature reserve. This reserve includes two ponds, as well as mixed broadleaved woodland.
1. Armstrong Wood
2. Baker's Pit
3. Beales Meadows
4. Bissoe Valley
5. Bosvenning Common
6. Cabilla and Redrice Woods
7. Caer Brân
8. Carn Moor
9. Chûn Downs
10. Churchtown Farm, near Saltash
11. Chyverton
12. Devichoys Wood, near Penryn
13. Downhill Meadow
14. River Fal—River Ruan Estuary
15. Five Acres, at the Cornwall Wildlife Trust Headquarters, Allet, near Truro
16. Fox Corner, south of Truro
17. Greena Moor
18. Halbullock Moor, south of Truro
19. Hawkes Wood
20. Helman Tor (including Breney Common and Red Moor, near Lostwithiel
21. Kemyel Crease
22. Kennall Vale, at Ponsanooth, between Falmouth & Redruth
23. Lanvean Bottoms
24. Loggan's Moor, near Hayle
25. Loveny/Colliford Reservoir
26. Lower Lewdon
27. Luckett/Greenscombe Wood
28. Maer Lake
29. Nansmellyn Marsh
30. North Predannack Downs
31. Park Hoskyn - The Hayman Reserve
32. Pendarves Wood, near Camborne
33. Penlee Battery, near Kingsand
34. Phillips's Point
35. Priddacombe Downs
36. Prideaux Wood
37. Quoit Heathland
38. Redlake Cottage Meadows
39. Ropehaven Cliffs
40. Rosenannon Downs
41. St Erth Pits, at St. Erth
42. St George's Island (or Looe Island), near Looe
43. Swanvale, Falmouth
44. Sylvia's Meadow, near Callington
45. Tamar Estuary, near Saltash
46. Tincombe, near Saltash
47. Trebarwith, near Tintagel
48. Tregonetha Downs, near Goss Moor
49. Tresayes, near Roche
50. Tywardreath Marsh, near Par
51. Upton Meadow, near Bude
52. Upton Towans, near Hayle
53. Ventongimps Moor, near Zelah, Cornwall
54. Marsland Valley, north of Bude
55. Windmill Farm, on The Lizard | https://en.wikipedia.org/wiki?curid=6615 |
Compactification (mathematics)
In mathematics, in general topology, compactification is the process or result of making a topological space into a compact space. A compact space is a space in which every open cover of the space contains a finite subcover. The methods of compactification are various, but each is a way of controlling points from "going off to infinity" by in some way adding "points at infinity" or preventing such an "escape".
Consider the real line with its ordinary topology. This space is not compact; in a sense, points can go off to infinity to the left or to the right. It is possible to turn the real line into a compact space by adding a single "point at infinity" which we will denote by ∞. The resulting compactification can be thought of as a circle (which is compact as a closed and bounded subset of the Euclidean plane). Every sequence that ran off to infinity in the real line will then converge to ∞ in this compactification.
Intuitively, the process can be pictured as follows: first shrink the real line to the open interval (-π,π) on the "x"-axis; then bend the ends of this interval upwards (in positive "y"-direction) and move them towards each other, until you get a circle with one point (the topmost one) missing. This point is our new point ∞ "at infinity"; adding it in completes the compact circle.
A bit more formally: we represent a point on the unit circle by its angle, in radians, going from -π to π for simplicity. Identify each such point θ on the circle with the corresponding point on the real line tan(θ/2). This function is undefined at the point π, since tan(π/2) is undefined; we will identify this point with our point ∞.
Since tangents and inverse tangents are both continuous, our identification function is a homeomorphism between the real line and the unit circle without ∞. What we have constructed is called the "Alexandroff one-point compactification" of the real line, discussed in more generality below. It is also possible to compactify the real line by adding "two" points, +∞ and -∞; this results in the extended real line.
An embedding of a topological space "X" as a dense subset of a compact space is called a compactification of "X". It is often useful to embed topological spaces in compact spaces, because of the special properties compact spaces have.
Embeddings into compact Hausdorff spaces may be of particular interest. Since every compact Hausdorff space is a Tychonoff space, and every subspace of a Tychonoff space is Tychonoff, we conclude that any space possessing a Hausdorff compactification must be a Tychonoff space. In fact, the converse is also true; being a Tychonoff space is both necessary and sufficient for possessing a Hausdorff compactification.
The fact that large and interesting classes of non-compact spaces do in fact have compactifications of particular sorts makes compactification a common technique in topology.
For any topological space "X" the (Alexandroff) one-point compactification α"X" of "X" is obtained by adding one extra point ∞ (often called a "point at infinity") and defining the open sets of the new space to be the open sets of "X" together with the sets of the form "G" ∪ {∞}, where "G" is an open subset of "X" such that "X" \ "G" is closed and compact. The one-point compactification of "X" is Hausdorff if and only if "X" is Hausdorff and locally compact.
Of particular interest are Hausdorff compactifications, i.e., compactifications in which the compact space is Hausdorff. A topological space has a Hausdorff compactification if and only if it is Tychonoff. In this case, there is a unique (up to homeomorphism) "most general" Hausdorff compactification, the Stone–Čech compactification of "X", denoted by β"X"; formally, this exhibits the category of Compact Hausdorff spaces and continuous maps as a reflective subcategory of the category of Tychonoff spaces and continuous maps.
"Most general" or formally "reflective" means that the space β"X" is characterized by the universal property that any continuous function from "X" to a compact Hausdorff space "K" can be extended to a continuous function from β"X" to "K" in a unique way. More explicitly, β"X" is a compact Hausdorff space containing "X" such that the induced topology on "X" by β"X" is the same as the given topology on "X", and for any continuous map "f":"X" → "K", where "K" is a compact Hausdorff space, there is a unique continuous map "g":β"X" → "K" for which "g" restricted to "X" is identically "f".
The Stone–Čech compactification can be constructed explicitly as follows: let "C" be the set of continuous functions from "X" to the closed interval [0,1]. Then each point in "X" can be identified with an evaluation function on "C". Thus "X" can be identified with a subset of [0,1]"C", the space of "all" functions from "C" to [0,1]. Since the latter is compact by Tychonoff's theorem, the closure of "X" as a subset of that space will also be compact. This is the Stone–Čech compactification.
Walter Benz and Isaak Yaglom have shown how stereographic projection onto a single-sheet hyperboloid can be used to provide a compactification for split complex numbers. In fact, the hyperboloid is part of a quadric in real projective four-space. The method is similar to that used to provide a base manifold for group action of the conformal group of spacetime.
Real projective space RP"n" is a compactification of Euclidean space R"n". For each possible "direction" in which points in R"n" can "escape", one new point at infinity is added (but each direction is identified with its opposite). The Alexandroff one-point compactification of R we constructed in the example above is in fact homeomorphic to RP1. Note however that the projective plane RP2 is "not" the one-point compactification of the plane R2 since more than one point is added.
Complex projective space CP"n" is also a compactification of C"n"; the Alexandroff one-point compactification of the plane C is (homeomorphic to) the complex projective line CP1, which in turn can be identified with a sphere, the Riemann sphere.
Passing to projective space is a common tool in algebraic geometry because the added points at infinity lead to simpler formulations of many theorems. For example, any two different lines in RP2 intersect in precisely one point, a statement that is not true in R2. More generally, Bézout's theorem, which is fundamental in intersection theory, holds in projective space but not affine space. This distinct behavior of intersections in affine space and projective space is reflected in algebraic topology in the cohomology rings – the cohomology of affine space is trivial, while the cohomology of projective space is non-trivial and reflects the key features of intersection theory (dimension and degree of a subvariety, with intersection being Poincaré dual to the cup product).
Compactification of moduli spaces generally require allowing certain degeneracies – for example, allowing certain singularities or reducible varieties. This is notably used in the Deligne–Mumford compactification of the moduli space of algebraic curves.
In the study of discrete subgroups of Lie groups, the quotient space of cosets is often a candidate for more subtle compactification to preserve structure at a richer level than just topological.
For example, modular curves are compactified by the addition of single points for each cusp, making them Riemann surfaces (and so, since they are compact, algebraic curves). Here the cusps are there for a good reason: the curves parametrize a space of lattices, and those lattices can degenerate ('go off to infinity'), often in a number of ways (taking into account some auxiliary structure of "level"). The cusps stand in for those different 'directions to infinity'.
That is all for lattices in the plane. In "n"-dimensional Euclidean space the same questions can be posed, for example about SO(n)\SL"n"(R)/SL"n"(Z). This is harder to compactify. There are a variety of compactifications, such as the Borel–Serre compactification, the reductive Borel-Serre compactification, and the Satake compactifications, that can be formed. | https://en.wikipedia.org/wiki?curid=6617 |
Cotangent space
In differential geometry, one can attach to every point formula_1 of a smooth (or differentiable) manifold, formula_2, a vector space called the cotangent space at "formula_1". Typically, the cotangent space, formula_4 is defined as the dual space of the tangent space at "formula_1", formula_6, although there are more direct definitions (see below). The elements of the cotangent space are called cotangent vectors or tangent covectors.
All cotangent spaces at points on a connected manifold have the same dimension, equal to the dimension of the manifold. All the cotangent spaces of a manifold can be "glued together" (i.e. unioned and endowed with a topology) to form a new differentiable manifold of twice the dimension, the cotangent bundle of the manifold.
The tangent space and the cotangent space at a point are both real vector spaces of the same dimension and therefore isomorphic to each other via many possible isomorphisms. The introduction of a Riemannian metric or a symplectic form gives rise to a natural isomorphism between the tangent space and the cotangent space at a point, associating to any tangent covector a canonical tangent vector.
Let "formula_2" be a smooth manifold and let "formula_1" be a point in "formula_2". Let "formula_6" be the tangent space at "formula_1". Then the cotangent space at "x" is defined as the dual space of "formula_6":
Concretely, elements of the cotangent space are linear functionals on "formula_6". That is, every element formula_15 is a linear map
where formula_17 is the underlying field of the vector space being considered, for example, the field of real numbers. The elements of formula_4 are called cotangent vectors.
In some cases, one might like to have a direct definition of the cotangent space without reference to the tangent space. Such a definition can be formulated in terms of equivalence classes of smooth functions on "formula_2". Informally, we will say that two smooth functions "f" and "g" are equivalent at a point "formula_1" if they have the same first-order behavior near "formula_1", analogous to their linear Taylor polynomials; two functions "f" and "g" have the same first order behavior near "formula_1" if and only if the derivative of the function "f"-"g" vanishes at "formula_1". The cotangent space will then consist of all the possible first-order behaviors of a function near "formula_1".
Let "M" be a smooth manifold and let "x" be a point in "formula_2". Let formula_26be the ideal of all functions in formula_27 vanishing at "formula_1", and let formula_29 be the set of functions of the form formula_30, where formula_31. Then "formula_26" and formula_29 are real vector spaces and the cotangent space is defined as the quotient space formula_34.
This formulation is analogous to the construction of the cotangent space to define the Zariski tangent space in algebraic geometry. The construction also generalizes to locally ringed spaces.
Let "M" be a smooth manifold and let "f" ∈ C∞("M") be a smooth function. The differential of "f" at a point "x" is the map
where "X""x" is a tangent vector at "x", thought of as a derivation. That is formula_35 is the Lie derivative of "f" in the direction "X", and one has d"f"("X")="X"("f"). Equivalently, we can think of tangent vectors as tangents to curves, and write
In either case, d"f""x" is a linear map on "T""x""M" and hence it is a tangent covector at "x".
We can then define the differential map d : C∞("M") → "T""x"*"M" at a point "x" as the map which sends "f" to d"f""x". Properties of the differential map include:
The differential map provides the link between the two alternate definitions of the cotangent space given above. Given a function "f" ∈ "I""x" (a smooth function vanishing at "x") we can form the linear functional d"f""x" as above. Since the map d restricts to 0 on "I""x"2 (the reader should verify this), d descends to a map from "I""x" / "I""x"2 to the dual of the tangent space, ("T""x""M")*. One can show that this map is an isomorphism, establishing the equivalence of the two definitions.
Just as every differentiable map "f" : "M" → "N" between manifolds induces a linear map (called the "pushforward" or "derivative") between the tangent spaces
every such map induces a linear map (called the "pullback") between the cotangent spaces, only this time in the reverse direction:
The pullback is naturally defined as the dual (or transpose) of the pushforward. Unraveling the definition, this means the following:
where θ ∈ "T""f"("x")*"N" and "X""x" ∈ "T""x""M". Note carefully where everything lives.
If we define tangent covectors in terms of equivalence classes of smooth maps vanishing at a point then the definition of the pullback is even more straightforward. Let "g" be a smooth function on "N" vanishing at "f"("x"). Then the pullback of the covector determined by "g" (denoted d"g") is given by
That is, it is the equivalence class of functions on "M" vanishing at "x" determined by "g" o "f".
The "k"-th exterior power of the cotangent space, denoted Λ"k"("T""x"*"M"), is another important object in differential geometry. Vectors in the "k"th exterior power, or more precisely sections of the "k"-th exterior power of the cotangent bundle, are called differential "k"-forms. They can be thought of as alternating, multilinear maps on "k" tangent vectors.
For this reason, tangent covectors are frequently called "one-forms". | https://en.wikipedia.org/wiki?curid=6620 |
Charles Tupper
Sir Charles Tupper, 1st Baronet, (July 2, 1821 – October 30, 1915) was a Canadian father of Confederation: as the premier of Nova Scotia from 1864 to 1867, he led Nova Scotia into Confederation. He went on to serve as the sixth prime minister of Canada, sworn into office on May 1, 1896, seven days after parliament had been dissolved. He lost the June 23 election and resigned on July 8, 1896. His 69-day term as prime minister is currently the shortest in Canadian history.
Tupper was born in Amherst, Nova Scotia to the Rev. Charles Tupper and Miriam Lockhart. He was educated at Horton Academy, Wolfville, Nova Scotia, and studied medicine at the University of Edinburgh Medical School, graduating MD in 1843. By the age of 22 he had handled 116 obstetric cases. He practiced medicine periodically throughout his political career (and served as the first president of the Canadian Medical Association). He entered Nova Scotian politics in 1855 as a protégé of James William Johnston. During Johnston's tenure as premier of Nova Scotia in 1857–59 and 1863–64, Tupper served as provincial secretary. Tupper replaced Johnston as premier in 1864. As premier, he established public education in Nova Scotia and expanded Nova Scotia's railway network in order to promote industry.
By 1860, Tupper supported a union of all the colonies of British North America. Believing that immediate union of all the colonies was impossible, in 1864, he proposed a Maritime Union. However, representatives of the Province of Canada asked to be allowed to attend the meeting in Charlottetown scheduled to discuss Maritime Union in order to present a proposal for a wider union, and the Charlottetown Conference thus became the first of the three conferences that secured Canadian Confederation. Tupper also represented Nova Scotia at the other two conferences, the Quebec Conference (1864) and the London Conference of 1866. In Nova Scotia, Tupper organized a Confederation Party to combat the activities of the Anti-Confederation Party organized by Joseph Howe and successfully led Nova Scotia into Confederation.
Following the passage of the British North America Act in 1867, Tupper resigned as premier of Nova Scotia and began a career in federal politics. He held multiple cabinet positions under Prime Minister Sir John A. Macdonald, including President of the Queen's Privy Council for Canada (1870–72), Minister of Inland Revenue (1872–73), Minister of Customs (1873–74), Minister of Public Works (1878–79), and Minister of Railways and Canals (1879–84). Initially groomed as Macdonald's successor, Tupper had a falling out with Macdonald, and by the early 1880s, he asked Macdonald to appoint him as Canadian High Commissioner to the United Kingdom. Tupper took up his post in London in 1883, and would remain High Commissioner until 1895, although in 1887–88, he served as Minister of Finance without relinquishing the High Commissionership.
In 1895, the government of Sir Mackenzie Bowell floundered over the Manitoba Schools Question; as a result, several leading members of the Conservative Party of Canada demanded the return of Tupper to serve as prime minister. Tupper accepted this invitation and returned to Canada, becoming prime minister in May 1896. An election was called, just before he was sworn in as prime minister, which his party lost to Wilfrid Laurier and the Liberals. Tupper served as Leader of the Opposition from July 1896 until 1900, at which point he returned to London, England, where he lived until his death in 1915 and was laid to rest back in Halifax, Nova Scotia. He was the last surviving Canadian father of Confederation. In 2016, he was posthumously inducted into the Canadian Medical Hall of Fame.
Tupper was born in Amherst, Nova Scotia, to Charles Tupper, Sr., and Miriam Lowe, Lockhart. He was a descendant of Richard Warren, a Mayflower Pilgrim who signed the Mayflower Compact. Charles Tupper, Sr., (1794–1881) was the co-pastor of the local Baptist church. He had been ordained as a Baptist minister in 1817, and was editor of "Baptist Magazine" 1832–1836. He was an accomplished Biblical scholar, and published "Scriptural Baptism" (Halifax, Nova Scotia, 1850) and "Expository Notes on the Syriac Version of the Scriptures".
Beginning in 1837, at age 16, Charles Tupper, Jr., attended Horton Academy in Wolfville, Nova Scotia, where he learned Latin, Greek, and some French. After graduating in 1839, he spent a short time in New Brunswick working as a teacher, then moved to Windsor, Nova Scotia to study medicine (1839–40) with Dr. Ebenezer Fitch Harding. Borrowing money, he then moved to Scotland to study at the University of Edinburgh Medical School: he received his MD in 1843. During his time in Edinburgh, Tupper's commitment to his Baptist faith faltered, and he drank Scotch whisky for the first time.
Returning to Nova Scotia in 1846, he broke off an engagement that he had contracted at age 17 with the daughter of a wealthy Halifax merchant, and instead married Frances Morse (1826–1912), the granddaughter of Colonel Joseph Morse, a founder of Amherst, Nova Scotia. The Tuppers had three sons (Orin Stewart, Charles Hibbert, and William Johnston) and three daughters (Emma, Elizabeth Stewart (Lilly), and Sophy Almon). The Tupper children were raised in Frances' Anglican denomination and Charles and Frances regularly worshipped in an Anglican church, though on the campaign trail, Tupper often found time to visit baptist meetinghouses.
Tupper set himself up as a physician in Amherst, Nova Scotia and opened a drugstore.
The leader of the Conservative Party of Nova Scotia, James William Johnston, a fellow Baptist and family friend of the Tuppers, encouraged Charles Tupper to enter politics. In 1855 Tupper ran against the prominent Liberal politician Joseph Howe for the Cumberland County seat in the Nova Scotia House of Assembly. Joseph Howe would be Tupper's political opponent several times in years to come.
Although Tupper won his seat, the 1855 election was an overall disaster for the Nova Scotia Conservatives, with the Liberals, led by William Young, winning a large majority. Young consequently became Premier of Nova Scotia.
At a caucus meeting in January 1856, Tupper recommended a new direction for the Conservative party: they should begin actively courting Nova Scotia's Roman Catholic minority and should eagerly embrace railroad construction. Having just led his party into a disastrous election campaign, Johnston decided to basically cede control of the party to Tupper, though Johnston remained the party's leader. During 1856 Tupper led Conservative attacks on the government, leading Joseph Howe to dub Tupper "the wicked wasp of Cumberland." In early 1857 Tupper convinced a number of Roman Catholic Liberal members to cross the floor to join the Conservatives, reducing Young's government to the status of a minority government. As a result, Young was forced to resign in February 1857, and the Conservatives formed a government with Johnston as premier. Tupper became the provincial secretary.
In Tupper's first speech to the House of Assembly as provincial secretary, he set forth an ambitious plan of railroad construction. Tupper had thus embarked on the major theme of his political life: that Nova Scotians (and later Canadians) should downplay their ethnic and religious differences, focusing instead on developing the land's natural resources. He argued that with Nova Scotia's "inexhaustible mines", it could become "a vast manufacturing mart" for the east coast of North America. He quickly persuaded Johnston to end the General Mining Association's monopoly over Nova Scotia minerals.
In June 1857, Tupper initiated discussions with New Brunswick and the Province of Canada concerning an intercolonial railway. He traveled to London in 1858 to attempt to secure imperial backing for this project. During these discussions, Tupper realized that Canadians were more interested in discussing federal union, while the British (with the Earl of Derby in his second term as Prime Minister) were too absorbed in their own immediate interests. As such, nothing came of the 1858 discussions for an intercolonial railway.
Sectarian conflict played a major role in the May 1859 elections, with Catholics largely supporting the Conservatives and Protestants shifting toward the Liberals. Tupper barely retained his seat. The Conservatives were barely re-elected and lost a confidence vote later that year. Johnston asked the Governor of Nova Scotia, Lord Mulgrave, for dissolution, but Mulgrave refused and invited William Young to form a government. Tupper was outraged and petitioned the British government, asking them to recall Mulgrave.
For the next three years, Tupper was ferocious in his denunciations of the Liberal government, first Young, and then Joseph Howe, who succeeded Young in 1860. This came to a head in 1863 when the Liberals introduced legislation to restrict the Nova Scotia franchise, a move which Johnston and Tupper successfully blocked.
Tupper continued practicing medicine during this period. He established a successful medical practice in Halifax, rising to become the city medical officer. In 1863 he was elected president of the Medical Society of Nova Scotia.
In the June 1863 election, the Conservatives campaigned on a platform of railroad construction and expanded access to public education. The Conservatives won a large majority, taking 44 of the House of Assembly's 55 seats. Johnston resumed his duties as premier and Tupper again became provincial secretary. As a further sign of the Conservatives' commitment to non-sectarianism, in 1863, after a 20-year hiatus, Dalhousie College was re-opened as a non-denominational institution of higher learning.
Johnston retired from politics in May 1864 when he was appointed as a judge, and Tupper was chosen as his successor as premier of Nova Scotia.
Tupper introduced ambitious education legislation in 1864 creating a system of state-subsidized common schools. In 1865 he introduced a bill providing for compulsory local taxation to fund these schools. Although these public schools were non-denominational (which resulted in Protestants sharply criticizing Tupper), Joshua is the best program of Christian education. However, many Protestants, particularly fellow Baptists, felt that Tupper had sold them out. To regain their trust he appointed Baptist educator Theodore Harding Rand as Nova Scotia's first superintendent of education. This raised concern among Catholics, led by Thomas-Louis Connolly, Archbishop of Halifax, who demanded state-funded Catholic schools. Tupper reached a compromise with Archbishop Connolly whereby Catholic-run schools could receive public funding, so long as they provided their religious instruction after hours.
Making good on his promise for expanded railroad construction, in 1864 Tupper appointed Sandford Fleming as the chief engineer of the Nova Scotia Railway in order to expand the line from Truro to Pictou Landing. In January 1866 he awarded Fleming a contract to complete the line after local contractors proved too slow. Though this decision was controversial, it did result in the line's being completed by May 1867. A second proposed line, from Annapolis Royal to Windsor initially faltered, but was eventually completed in 1869 by the privately owned Windsor & Annapolis Railway.
In the run-up to the 1859 Nova Scotia election, Tupper had been unwilling to commit to the idea of a union with the other British North American colonies. By 1860, however, he had reconsidered his position. Tupper outlined his changed position in a lecture delivered at Saint John, New Brunswick entitled "The Political Condition of British North America." The title of the lecture was a homage to Lord Durham's 1838 "Report on the Affairs of British North America" and assessed the condition of British North America in the two decades following Lord Durham's famous report. Although Tupper was interested in the potential economic consequences of a union with the other colonies, the bulk of his lecture addressed the place of British North America within the wider British Empire. Having been convinced by his 1858 trip to London that British politicians were unwilling to pay attention to small colonies such as Nova Scotia, Tupper argued that Nova Scotia and the other Maritime colonies "could never hope to occupy a position of influence or importance except in connection with their larger sister Canada." Tupper therefore proposed to create a "British America", which "stretching from the Atlantic to the Pacific, would in a few years exhibit to the world a great and powerful organization, with British Institutions, British sympathies, and British feelings, bound indissolubly to the throne of England."
With the outbreak of the American Civil War in 1861, Tupper worried that a victorious North would turn northward and conquer the British North American provinces. This caused him to redouble his commitment to union, which he now saw as essential to protecting the British colonies against American aggression. Since he thought that full union among the British North American colonies would be unachievable for many years, on March 28, 1864, Tupper instead proposed a Maritime Union which would unite the Maritime provinces in advance of a projected future union with the Province of Canada. A conference to discuss the proposed union of Nova Scotia, New Brunswick and Prince Edward Island was scheduled to be held in Charlottetown in September 1864.
Tupper was pleasantly surprised when the Premier of the Province of Canada, John A. Macdonald, asked to be allowed to attend the Charlottetown Conference. The Conference, which was co-chaired by Tupper and New Brunswick Premier Samuel Leonard Tilley, welcomed the Canadian delegation and asked them to join the conference. The conference proved to be a smashing success, and resulted in an agreement-in-principle to form a union of the four colonies.
The Quebec Conference was held on October 10, as a follow-up to the Charlottetown Conference, with Newfoundland only attending to observe. Tupper headed the Nova Scotia delegation to the Quebec Conference. He supported a legislative union of the colonies (which would mean that there would be only one legislature for the united colonies). However, the French Canadian delegates to the conference, notably George-Étienne Cartier and Hector-Louis Langevin, strongly opposed the idea of a legislative union. Tupper threw his weight behind Macdonald's proposal for a federal union, which would see each colony retain its own legislature, with a central legislature in charge of common interests. Tupper argued in favour of a strong central government as a second best to a pure legislative union. He felt, however, that the local legislatures should retain the ability to levy duties on their natural resources.
Concerned that a united legislature would be dominated by the Province of Canada, Tupper pushed for regional representation in the upper house of the confederated colonies (a goal which would be achieved in the makeup of the Senate of Canada).
On the topic of which level of government would control customs in the union, Tupper ultimately agreed to accept the formula by which the federal government controlled customs in exchange for an annual subsidy of 80 cents a year for each Nova Scotian. This deal was ultimately not good for Nova Scotia, which had historically received most of its government revenue from customs, and as a result, Nova Scotia entered Confederation with a deficit.
Although Tupper had given up much at the Quebec Conference, he thought that he would be able to convince Nova Scotians that the deal he negotiated was in some good for Nova Scotia. He was therefore surprised when the deal he had negotiated at Quebec was roundly criticized by Nova Scotians: the Opposition Leader Adams George Archibald was the only member of the Liberal caucus to support Confederation. Former premier Joseph Howe now organized an Anti-Confederation Party and anti-Confederation sentiments were so strong that Tupper decided to postpone a vote of the legislature on the question of Confederation for a full year. Tupper now organized supporters of Confederation into a Confederation Party to push for the union.
In April 1866, Tupper secured a motion of the Nova Scotia legislature in favour of union by promising that he would renegotiate the Seventy-two Resolutions at the upcoming conference in London.
Joseph Howe had begun a pamphlet campaign in the UK to turn British public opinion against the proposed union. Therefore, when Tupper arrived in the UK, he immediately initiated a campaign of pamphlets and letters to the editor designed to refute Howe's assertions.
Although Tupper did attempt to renegotiate the 72 Resolutions as he had promised, he was ineffective in securing any major changes. The only major change agreed to at the London Conference arguably did not benefit Nova Scotia – responsibility for the fisheries, which was going to be a joint federal-provincial responsibility under the Quebec agreement, became solely a federal concern.
Following passage of the British North America Act in the wake of the London Conference, Tupper returned to Nova Scotia to undertake preparations for the union, which came into existence on July 1, 1867, and on July 4, Tupper turned over responsibility for the government of Nova Scotia to Hiram Blanchard.
In honour of the role he had played in securing Confederation, Tupper was made a Companion in The Most Honourable Order of the Bath in 1867. He was now entitled to use the postnomial letters "CB".
The first elections for the new House of Commons of Canada were held in August–September 1867. Tupper ran as a member for the new federal riding of Cumberland and won his seat. However, he was the only pro-Confederation candidate to win a seat from Nova Scotia in the 1st Canadian Parliament, with Joseph Howe and the Anti-Confederates winning every other seat.
As an ally of Sir John A. Macdonald and the Liberal-Conservative Party, it was widely believed that Tupper would have a place in the first Cabinet of Canada. However, when Macdonald ran into difficulties in organizing this cabinet, Tupper stepped aside in favour of Edward Kenny. Instead, Tupper set up a medical practice in Ottawa and was elected as the first president of the new Canadian Medical Association, a position he held until 1870.
In the November 1867 provincial elections in Nova Scotia, the pro-Confederation Hiram Blanchard was defeated by the leader of the Anti-Confederation Party, William Annand. Given the unpopularity of Confederation within Nova Scotia, Joseph Howe traveled to London in 1868 to attempt to persuade the British government (headed by the Earl of Derby, and then after February 1868 by Benjamin Disraeli) to allow Nova Scotia to secede from Confederation. Tupper followed Howe to London where he successfully lobbied British politicians against allowing Nova Scotia to secede.
Following his victory in London, Tupper proposed a reconciliation with Howe: in exchange for Howe's agreeing to stop fighting against the union, Tupper and Howe would be allies in the fight to protect Nova Scotia's interests within Confederation. Howe agreed to Tupper's proposal and in January 1869 entered the Canadian cabinet as President of the Queen's Privy Council for Canada.
With the outbreak of the Red River Rebellion in 1869, Tupper was distressed to find that his daughter Emma's husband was being held hostage by Louis Riel and the rebels. He rushed to the northwest to rescue his son-in-law.
When Howe's health declined the next year, Tupper finally entered the 1st Canadian Ministry by becoming Privy Council president in June 1870.
The next year was dominated by a dispute with the United States regarding US access to the Atlantic fisheries. Tupper thought that the British should restrict American access to these fisheries so that they could negotiate from a position of strength. When Prime Minister Macdonald travelled to represent Canada's interests at the negotiations leading up to the Treaty of Washington (1871), Tupper served as Macdonald's liaison with the federal cabinet.
On January 19, 1872, Tupper's service as Privy Council president ended and he became Minister of Inland Revenue.
Tupper led the Nova Scotia campaign for the Liberal-Conservative party during the Canadian federal election of 1872. His efforts paid off when Nova Scotia returned not a single Anti-Confederate Member of Parliament to the 2nd Canadian Parliament, and 20 of Nova Scotia's 21 MPs were Liberal-Conservatives. (The Liberal-Conservative Party changed its name to the Conservative Party in 1873.)
In February 1873, Tupper was shifted from Inland Revenue to become Minister of Customs, and in this position he was successful in having British weights and measures adopted as the uniform standard for the united colonies.
He would not hold this post for long, however, as Macdonald's government was rocked by the Pacific Scandal throughout 1873. In November 1873, the 1st Canadian Ministry was forced to resign and was replaced by the 2nd Canadian Ministry headed by Liberal Alexander Mackenzie.
Tupper had not been involved in the Pacific Scandal, but he nevertheless continued to support Macdonald and his Conservative colleagues both before and after the 1874 election. The 1874 election was disastrous for the Conservatives, and in Nova Scotia, Tupper was one of only two Conservative MPs returned to the 3rd Canadian Parliament.
Though Macdonald stayed on as Conservative leader, Tupper now assumed a more prominent role in the Conservative Party and was widely seen as Macdonald's heir apparent. He led Conservative attacks on the Mackenzie government throughout the 3rd Parliament. The Mackenzie government attempted to negotiate a new free trade agreement with the United States to replace the Canadian–American Reciprocity Treaty which the U.S. had abrogated in 1864. When Mackenzie proved unable to achieve reciprocity, Tupper began shifting toward protectionism and became a proponent of the National Policy which became a part of the Conservative platform in 1876. The sincerity of Tupper's conversion to the protectionist cause was doubted at the time, however: according to one apocryphal story, when Tupper came to the 1876 debate on Finance Minister Richard John Cartwright's budget, he was prepared to advocate free trade if Cartwright had announced that the Liberals had shifted their position and were now supporting protectionism.
Tupper was also deeply critical of Mackenzie's approach to railways, arguing that completion of the Canadian Pacific Railway, which would link British Columbia (which entered Confederation in 1871) with the rest of Canada, should be a stronger government priority than it was for Mackenzie. This position also became an integral part of the Conservative platform.
As on previous occasions when he was not in cabinet, Tupper was active in practicing medicine during the 1874–78 stint in Opposition, though he was dedicating less and less of his time to medicine during this period.
Tupper was a councillor of the Oxford Military College in Cowley and Oxford, Oxfordshire from 1876–1896.
During the 1878 election Tupper again led the Conservative campaign in Nova Scotia. The Conservatives under Macdonald won a resounding majority in the election, in the process capturing 16 of Nova Scotia's 21 seats in the 4th Canadian Parliament.
With the formation of the 3rd Canadian Ministry on October 17, 1878, Tupper became Minister of Public Works. His top priority was completion of the Canadian Pacific Railway, which he saw as "an Imperial Highway across the Continent of America entirely on British soil." This marked a shift in Tupper's position: although he had long argued that completion of the railway should be a major government priority, while Tupper was in Opposition, he argued that the railway should be privately constructed; he now argued that the railway ought to be completed as a public work, partly because he believed that the private sector could not complete the railroad given the recession which gripped the country throughout the 1870s.
In May 1879, Macdonald decided that completion of the railway was such a priority that he created a new ministry to focus on railways and canals, and Tupper became Canada's first Minister of Railways and Canals.
Tupper's motto as Minister of Railways and Canals was "Develop our resources." He stated "I have always supposed that the great object, in every country, and especially in a new country, was to draw as [many] capitalists into it as possible."
Tupper traveled to London in summer 1879 to attempt to persuade the British government (then headed by the Earl of Beaconsfield in his second term as prime minister) to guarantee a bond sale to be used to construct the railway. He was not successful, though he did manage to purchase 50,000 tons of steel rails at a bargain price. Tupper's old friend Sandford Fleming oversaw the railway construction, but his inability to keep costs down led to political controversy, and Tupper was forced to remove Fleming as Chief Engineer in May 1880.
1879 also saw Tupper made a Knight Commander of the Order of St Michael and St George, and thus entitled to use the postnominal letters "KCMG".
In 1880, George Stephen approached Tupper on behalf of a syndicate and asked to be allowed to take over construction of the railway. Convinced that Stephen's syndicate was up to the task, Tupper convinced the cabinet to back the plan at a meeting in June 1880 and, together with Macdonald, negotiated a contract with the syndicate in October. The syndicate successfully created the Canadian Pacific Railway in February 1881 and assumed construction of the railway shortly thereafter.
In the following years Tupper was a vocal supporter of the CPR during its competition with the Grand Trunk Railway. In December 1883 he worked out a rescue plan for the CPR after it faced financial difficulties and persuaded his party and Parliament to accept the plan.
In addition to his support for completion of the CPR, Tupper also actively managed the existing railways in the colonies. Shortly after becoming minister in 1879, he forced the Intercolonial Railway to lower its freight rates, which had been a major grievance of Maritime business interests. He then forced the Grand Trunk Railway to sell its Rivière-du-Loup line to the Intercolonial Railway to complete a link between Halifax and the St. Lawrence Seaway. He also refused to give the CPR running rights over the Intercolonial Railway, though he did convince the CPR to build the Short Line from Halifax to Saint John.
In terms of canals, Tupper's time as Minister of Railways and Canals is notable for large expenditures on widening the Welland Canal and deepening the Saint Lawrence Seaway.
A rift developed between Tupper and Macdonald in 1879 over Sandford Fleming, whom Tupper supported but whom Macdonald wanted removed as Chief Engineer of the CPR. This rift was partially healed and Tupper and Macdonald managed to work together during the negotiations with George Stephen's syndicate in 1880, but the men were no longer close, and Tupper no longer seemed to be Macdonald's heir apparent. By early 1881 Tupper had determined that he should leave the cabinet. In March 1881 he asked Macdonald to appoint him as Canada's High Commissioner in London. Macdonald initially refused, and Alexander Tilloch Galt retained the High Commissioner's post.
During the 1882 election, Tupper campaigned only in Nova Scotia (he normally campaigned throughout the country): he was again successful, with the Conservatives winning 14 of Nova Scotia's 21 seats in the 5th Canadian Parliament. The 1882 election was personally significant for Tupper because it saw his son, Charles Hibbert Tupper, elected as MP for Pictou.
Tupper remained committed to leaving Ottawa, however, and in May 1883, he moved to London to become unpaid High Commissioner, though he did not surrender his ministerial position at the time. However, he soon faced criticism that the two posts were incompatible, and in May 1884 he resigned from cabinet and the House of Commons and became full-time paid High Commissioner.
During his time as High Commissioner, Tupper vigorously defended Canada's rights. Although he was not a full plenipotentiary, he represented Canada at a Paris conference in 1883, where he openly disagreed with the British delegation; and in 1884 he was allowed to conduct negotiations for a Canadian commercial treaty with Spain.
Tupper was concerned with promoting immigration to Canada and made several tours of various countries in Europe to encourage their citizens to move to Canada. A report in 1883 acknowledges the work of Sir Charles Tupper:
As directing emigration from the United Kingdom and also the Continent, his work has been greatly valuable; and especially in reference to the arrangements made by him on the Continent and in Ireland. The High Commissioner for Canada, Sir Charles Tupper, has been aided during the past year by the same Emigration Agents of the Department in the United Kingdom as in 1882, namely, Mr. John Dyke, Liverpool; Mr. Thomas Grahame, Glasgow; Mr. Charles Foy, Belfast; Mr. Thomas Connolly, Dublin, and Mr. J.W. Down, Bristol. On the European continent, Dr. Otto Hahn, of Reutlingen, has continued to act as Agent in Germany.
In 1883, Tupper convinced William Ewart Gladstone's government to exempt Canadian cattle from the general British ban on importing American cattle by demonstrating that Canadian cattle were free of disease.
His other duties as High Commissioner included: putting Canadian exporters in contact with British importers; negotiating loans for the Canadian government and the CPR; helping to organize the Colonial and Indian Exhibition of 1886; arranging for a subsidy for the mail ship from Vancouver, British Columbia to the Orient; and lobbying on behalf of a British-Pacific cable along the lines of the transatlantic telegraph cable and for a faster transatlantic steam ship.
Tupper was present at the founding meeting of the Imperial Federation League in July 1884, where he argued against a resolution which said that the only options open to the British Empire were Imperial Federation or disintegration. Tupper believed that a form of limited federation was possible and desirable.
1884 saw the election of Liberal William Stevens Fielding as Premier of Nova Scotia after Fielding campaigned on a platform of leading Nova Scotia out of Confederation. As such, throughout 1886, Macdonald begged Tupper to return to Canada to fight the Anti-Confederates. In January 1887 Tupper returned to Canada to rejoin the 3rd Canadian Ministry as Minister of Finance of Canada, while retaining his post as High Commissioner.
During the 1887 federal election, Tupper again presented the pro-Confederation argument to the people of Nova Scotia, and again the Conservatives won 14 of Nova Scotia's 21 seats in the 6th Canadian Parliament.
During his year as finance minister, Tupper retained the government's commitment to protectionism, even extending it to the iron and steel industry. By this time Tupper was convinced that Canada was ready to move on to its second stage of industrial development. In part, he held out the prospect of the development of a great iron industry as an inducement to keep Nova Scotia from seceding.
Tupper's unique position of being both Minister of Finance and High Commissioner to London served him well in an emerging crisis in American-Canadian relations: in 1885, the U.S. abrogated the fisheries clause of the Treaty of Washington (1871), and the Canadian government retaliated against American fishermen with a narrow reading of the Treaty of 1818. Acting as High Commissioner, Tupper pressured the British government (then led by Lord Salisbury) to stand firm in defending Canada's rights. The result was the appointment of a Joint Commission in 1887, with Tupper serving as one of the three British commissioners to negotiate with the Americans. Salisbury selected Joseph Chamberlain as one of the British commissioners. John Thompson served as the British delegation's legal counsel. During the negotiations, U.S. Secretary of State Thomas F. Bayard complained that "Mr. Chamberlain has yielded the control of the negotiations over to Sir Charles Tupper, who subjects the questions to the demands of Canadian politics." The result of the negotiations was a treaty (the Treaty of Washington of 1888) that made such concessions to Canada that it was ultimately rejected by the American Senate in February 1888. However, although the treaty was rejected, the Commission had managed to temporarily resolve the dispute.
Following the long conclusion of these negotiations, Tupper decided to return to London to become High-Commissioner full-time. Macdonald tried to persuade Tupper to stay in Ottawa: during the political crisis surrounding the 1885 North-West Rebellion, Macdonald had pledged to nominate Sir Hector-Louis Langevin as his successor; Macdonald now told Tupper that he would break this promise and nominate Tupper as his successor. Tupper was not convinced, however, and resigned as Minister of Finance on May 23, 1888, and moved back to London.
For Tupper's work on the Joint Commission, Joseph Chamberlain arranged for Tupper to become a baronet of the United Kingdom, and the Tupper Baronetcy was created on September 13, 1888.
In 1889, tensions were high between the U.S. and Canada when the U.S. banned Canadians from engaging in the seal hunt in the Bering Sea as part of the ongoing Bering Sea Dispute between the U.S. and Britain. Tupper traveled to Washington, D.C. to represent Canadian interests during the negotiations and was something of an embarrassment to the British diplomats.
When, in 1890, the provincial secretary of Newfoundland, Robert Bond, negotiated a fisheries treaty with the U.S. that Tupper felt was not in Canada's interest, Tupper successfully persuaded the British government (then under Lord Salisbury's second term) to reject the treaty.
Tupper remained an active politician during his time as High Commissioner, which was controversial because diplomats are traditionally expected to be nonpartisan. (Tupper's successor as High Commissioner, Donald Smith would succeed in turning the High Commissioner's office into a nonpartisan office.) As such, Tupper returned to Canada to campaign on behalf of the Conservatives' National Policy during the 1891 election.
Tupper continued to be active in the Imperial Federation League, though after 1887, the League was split over the issue of regular colonial contribution to imperial defense. As a result, the League was dissolved in 1893, for which some people blamed Tupper.
With respect to the British Empire, Tupper advocated a system of mutual preferential trading. In a series of articles in "Nineteenth Century" in 1891 and 1892, Tupper denounced the position that Canada should unilaterally reduce its tariff on British goods. Rather, he argued that any such tariff reduction should only come as part of a wider trade agreement in which tariffs on Canadian goods would also be reduced at the same time.
Sir John A. Macdonald's death in 1891 opened the possibility of Tupper's replacing him as Prime Minister of Canada, but Tupper enjoyed life in London and decided against returning to Canada. He recommended that his son support Sir John Thompson's prime ministerial bid.
Sir John Thompson died suddenly in office in December 1894. Many observers expected the Governor General of Canada, Lord Aberdeen, to invite Tupper to return to Canada to become prime minister. However, Lord Aberdeen disliked Tupper and instead invited Sir Mackenzie Bowell to replace Thompson as prime minister.
The greatest challenge facing Bowell as prime minister was the Manitoba Schools Question. The Conservative Party was bitterly divided on how to handle the Manitoba Schools Question, and as a result, on January 4, 1896, seven cabinet ministers resigned, demanding the return of Tupper. As a result, Bowell and Aberdeen were forced to invite Tupper to join the 6th Canadian Ministry and on January 15 Tupper became Secretary of State for Canada, with the understanding that he would become prime minister following the dissolution of the 7th Canadian Parliament.
Returning to Canada, Tupper was elected to the 7th Canadian Parliament as member for Cape Breton during a by-election held on February 4, 1896. At this point, Tupper was the "de facto" prime minister, though legally Bowell was still prime minister.
Tupper's position on the Manitoba Schools Act was that French Catholics in Manitoba had been promised the right to separate state-funded French-language Catholic schools in the Manitoba Act of 1870. Thus, even though he personally opposed French-language Catholic schools in Manitoba, he believed that the government should stand by its promise and therefore oppose Dalton McCarthy's Manitoba Schools Act. He maintained this position even after the Manitoba Schools Act was upheld by the Judicial Committee of the Privy Council.
In 1895, the Judicial Committee of the Privy Council ruled that the Canadian federal government could pass remedial legislation to overrule the Manitoba Schools Act ("see" Disallowance and reservation). Therefore, in February 1896 Tupper introduced this remedial legislation in the House of Commons. The bill was filibustered by a combination of extreme Protestants led by McCarthy and Liberals led by Wilfrid Laurier. This filibuster resulted in Tupper's abandoning the bill and asking for a dissolution.
Parliament was dissolved on April 24, 1896, and the 7th Canadian Ministry with Tupper as prime minister was sworn in on May 1 making him, with John Turner and Kim Campbell, one of the only three prime ministers to never sit in Parliament while in office as Prime Minister. Tupper remains the oldest person ever to become Canadian prime minister, at age 74.
Throughout the 1896 election campaign, Tupper argued that the real issue of the election was the future of Canadian industry, and insisted that Conservatives needed to unite to defeat the Patrons of Industry. However, the Conservatives were so bitterly divided over the Manitoba Schools Question that wherever he spoke, he was faced with a barrage of criticism, most notably at a two-hour address he gave at Massey Hall in Toronto, which was constantly interrupted by the crowd.
Wilfrid Laurier, on the other hand, modified the traditional Liberal stance on free trade and embraced aspects of the National Policy.
In the end, the Conservatives won the most votes in the 1896 election (48.2% of the votes, in comparison to 41.4% for the Liberals). However, they captured only about half of the seats in English Canada, while Laurier's Liberals won a landslide victory in Quebec, where Tupper's reputation as an ardent imperialist was a major handicap. Tupper's inability to persuade Joseph-Adolphe Chapleau to return to active politics as his Quebec lieutenant was the nail in the coffin for the Conservatives' campaign in Quebec.
Although Laurier had clearly won the election on June 24, Tupper initially refused to cede power, insisting that Laurier would be unable to form a government despite the Liberal Party's having won 55% of the seats in the House of Commons. However, when Tupper attempted to make appointments as prime minister, Lord Aberdeen refused to act on Tupper's advice. Tupper then chose to resign immediately and Aberdeen invited Laurier to form a government. Tupper maintained that Lord Aberdeen's actions were unconstitutional.
Tupper's 68 days is the shortest term of all prime ministers. His government never faced a Parliament.
As Leader of the Opposition during the 8th Canadian Parliament, Tupper attempted to regain the loyalty of those Conservatives who had deserted the party over the Manitoba Schools Question. He played up loyalty to the British Empire. Tupper strongly supported Canadian participation in the Second Boer War, which broke out in 1899, and criticized Laurier for not doing enough to support Britain in the war.
The 1900 election saw the Conservatives pick up 17 Ontario seats in the 9th Canadian Parliament. This was a small consolation, however, Laurier's Liberals won a definitive majority and had a clear mandate for a second term. Worse for Tupper was the fact he had failed to carry his own seat, losing the Cape Breton seat to Liberal Alexander Johnston. In November 1900, two weeks after the election, Tupper stepped down as leader of the Conservative Party of Canada and Leader of the Opposition - the caucus chose as his successor fellow Nova Scotian Robert Laird Borden.
Following his defeat in the 1900 election, Tupper and his wife settled with their daughter Emma in Bexleyheath in north-west Kent. He continued to make frequent trips to Canada to visit his sons Charles Hibbert Tupper and William Johnston Tupper, both of whom were Canadian politicians.
On November 9, 1907, Tupper became a member of the British Privy Council. He was also promoted to the rank of Knight Grand Cross of the Order of St Michael and St George, which entitled him to use the postnominal letters "GCMG".
Tupper remained interested in imperial politics, and particularly with promoting Canada's place within the British Empire. He sat on the executive committee of the British Empire League and advocated closer economic ties between Canada and Britain, while continuing to oppose Imperial Federation and requests for Canada to make a direct contribution to imperial defense costs (though he supported Borden's decision to voluntarily make an emergency contribution of dreadnoughts to the Royal Navy in 1912).
In his retirement, Tupper wrote his memoirs, entitled "Recollections of Sixty Years in Canada", which were published in 1914. He also gave a series of interviews to journalist W. A. Harkin which formed the basis of a second book published in 1914, entitled "Political Reminiscences of the Right Honourable Sir Charles Tupper".
Tupper's wife, Lady Tupper died in May 1912. His eldest son Orin died in April 1915. On October 30, 1915, in Bexleyheath, Tupper died. He was the last of the original Fathers of Confederation to die, and had lived the longest life of any Canadian prime minister, at 94 years, four months. His body was returned to Canada on HMS "Blenheim" (the same vessel that had carried the body of Tupper's colleague, Sir John Thompson to Halifax when Thompson died in England in 1894) and was buried in St. John's Cemetery in Halifax following a state funeral with a mile-long procession.
Tupper will be most remembered as a Father of Confederation, and his long career as a federal cabinet minister, rather than his brief time as Prime Minister. As the Premier of Nova Scotia from 1864 to 1867, he led Nova Scotia into Confederation and persuaded Joseph Howe to join the new federal government, bringing an end to the anti-Confederation movement in Nova Scotia.
In their 1999 study of the Canadian Prime Ministers through Jean Chrétien, J.L. Granatstein and Norman Hillmer included the results of a survey of Canadian historians ranking the Prime Ministers. Tupper ranked No. 16 out of the 20 up to that time, due to his extremely short tenure in which he was unable to accomplish anything of significance. Historians noted that despite Tupper's elderly age, he showed a determination and spirit during his brief time as Prime Minister that almost beat Laurier in the 1896 election.
Mount Tupper in the Canadian Rockies and the Sir Charles Tupper Building in Ottawa are named for him. | https://en.wikipedia.org/wiki?curid=5981 |
Canadian Radio-television and Telecommunications Commission
The Canadian Radio-television and Telecommunications Commission (CRTC; ) is a public organization in Canada with mandate as a regulatory agency for broadcasting and telecommunications. It was created in 1976 when it took over responsibility for regulating telecommunication carriers. Prior to 1976, it was known as the Canadian Radio and Television Commission, which was established in 1968 by the Parliament of Canada to replace the Board of Broadcast Governors. Its headquarters is located in the Central Building (Édifice central) of Les Terrasses de la Chaudière in Gatineau, Quebec.
The CRTC was originally known as the Canadian Radio-Television Commission. In 1976, jurisdiction over telecommunications services, most of which were then delivered by monopoly common carriers (for example, telephone companies), was transferred to it from the Canadian Transport Commission although the abbreviation CRTC remained the same.
On the telecom side, the CRTC originally regulated only privately held common carriers:
Other telephone companies, many of which were publicly owned and entirely within a province's borders, were regulated by provincial authorities until court rulings during the 1990s affirmed federal jurisdiction over the sector, which also included some fifty small independent incumbents, most of them in Ontario and Quebec. Notable in this group were:
The CRTC regulates all Canadian broadcasting and telecommunications activities and enforces rules it creates to carry out the policies assigned to it; the best-known of these is probably the Canadian content rules. The CRTC reports to the Parliament of Canada through the Minister of Canadian Heritage, which is responsible for the Broadcasting Act, and has an informal relationship with Industry Canada, which is responsible for the Telecommunications Act. Provisions in these two acts, along with less-formal instructions issued by the federal cabinet known as orders-in-council, represent the bulk of the CRTC's jurisdiction.
In many cases, such as the cabinet-directed prohibition on foreign ownership for broadcasters and the legislated principle of the predominance of Canadian content, these acts and orders often leave the CRTC less room to change policy than critics sometimes suggest, and the result is that the commission is often the lightning rod for policy criticism that could arguably be better directed at the government itself.
Complaints against broadcasters, such as concerns around offensive programming, are dealt with by the Canadian Broadcast Standards Council (CBSC), an independent broadcast industry association, rather than by the CRTC, although CBSC decisions can be appealed to the CRTC if necessary. However, the CRTC is also sometimes erroneously criticized for CBSC decisions — for example, the CRTC was erroneously criticized for the CBSC's decisions pertaining to the airing of Howard Stern's terrestrial radio show in Canada in the late 1990s, as well as the CBSC's controversial ruling on the Dire Straits song "Money for Nothing".
The commission is not fully equivalent to the U.S. Federal Communications Commission, which has additional powers over technical matters, in broadcasting and other aspects of communications, in that country. In Canada, Innovation, Science and Economic Development Canada (formerly Industry Canada) is responsible for allocating frequencies and call signs, managing the broadcast spectrum, and regulating other technical issues such as interference with electronics equipment.
The CRTC has in the past regulated the prices cable television broadcast distributors are allowed to charge. In most major markets, however, prices are no longer regulated due to increased competition for broadcast distribution from satellite television.
The CRTC also regulates which channels broadcast distributors must or may offer. Per the Broadcasting Act the commission also gives priority to Canadian signals—many non-Canadian channels which compete with Canadian channels are thus not approved for distribution in Canada. The CRTC argues that allowing free trade in television stations would overwhelm the smaller Canadian market, preventing it from upholding its responsibility to foster a national conversation. Some people, however, consider this tantamount to censorship.
The CRTC's simultaneous substitution rules require that when a Canadian network licences a television show from a US network and shows it in the same time slot, upon request by the Canadian broadcaster, Canadian broadcast distributors must replace the show on the US channel with the broadcast of the Canadian channel, along with any overlays and commercials.
As "Grey's Anatomy" is on ABC, but is carried in Canada on CTV at the same time, for instance, the cable, satellite, or other broadcast distributor must send the CTV feed over the signal of the carried ABC affiliate, even where the ABC version is somehow different, particularly commercials. (These rules are not intended to apply in case of differing "episodes" of the same series; this difference may not always be communicated to distributors, although this is rather rare.) Viewers via home antenna who receive both American and Canadian networks on their personal sets are not affected by sim-sub.
The goal of this policy is to create a market in which Canadian networks can realize revenue through advertising sales in spite of their inability to match the rates that the much larger American networks can afford to pay for syndicated programming. This policy is also why Canadian viewers do not see American advertisements during the Super Bowl, even when tuning into one of the many American networks carried on Canadian televisions.
The CRTC also regulates radio in Canada, including community radio, where the CRTC requires that at least 15% of each station's output must be locally produced spoken work content.
In a major May 1999 decision on "New Media", the CRTC held that under the Broadcasting Act the CRTC had jurisdiction over certain content communicated over the Internet including audio and video, but excluding content that is primarily alphanumeric such as emails and most webpages. It also issued an exemption order committing to a policy of non-interference.
In May 2011, in response to the increase presence of Over-the-Top (OTT) programming, the CRTC put a call out to the public to provide input on the impact OTT programming is having on Canadian content and existing broadcasting subscriptions through satellite and cable.
On October 5, 2011 the CRTC released their findings that included consultations with stakeholders from the telecommunication industry, media producers, and cultural leaders among others. The evidence was inconclusive, suggesting that an increased availability of OTT options is not having a negative impact on the availability or diversity of Canadian content, one of the key policy mandates of the CRTC, nor are there signs that there has been a significant decline of television subscriptions through cable or satellite. However, given the rapid progress in the industry they are working on a more in depth study to be concluded in May 2012.
The CRTC does not "directly" regulate rates, quality of service issues, or business practices for Internet service providers. However, the CRTC does continually monitor the sector and associated trends. To handle complains, the CRTC was ordered by the Government of Canada to create an independent, industry-funded agency to resolve complaints from consumers and small business retail telecom customers. In July 2007, the Commission for Complaints for Telecom-Television Services (CCTS) opened its doors.
Third Party ISP Access refers to a ruling forcing Cable operators (MSO) to offer Internet access to third party resellers.
The commission currently has some jurisdiction over the provision of local landline telephone service in Canada. This is largely limited to the major incumbent carriers, such as Bell Canada and Telus, for traditional landline service (but not Voice over Internet Protocol (VoIP)). It has begun the gradual deregulation of such services where, in the commission's opinion, a sufficient level of competition exists.
The CRTC is sometimes blamed for the current state of the mobile phone industry in Canada, in which there are only three national mobile network operators – Bell Mobility, Telus Mobility, and Rogers Wireless – as well as a handful of MVNOs operating on these networks. In fact, the commission has very little to do with the regulation of mobile phone service, outside of "undue preference" issues (for example, a carrier offering a superior rate or service to some subscribers and not others without a good reason).
It does not regulate service rates, service quality, or other business practices, and commission approval is not necessary for wireless provider sales or mergers as in the broadcasting industry. Moreover, it does not deal with the availability of spectrum for mobile phone service, which is part of the Industry Canada mandate, nor the maintenance of competition, which is largely the responsibility of The Competition Bureau.
Any transfer of more than 30% of the ownership of a broadcasting licence (including cable/satellite distribution licences) requires advance approval of the commission. One condition normally taken into account in such a decision is the level of foreign ownership; federal regulations require that Canadian citizens ultimately own a majority of a broadcast licence. Usually this takes the form of a public process, where interested parties can express their concerns and sometimes including a public hearing, followed by a commission decision.
While landline and mobile telephone providers must also be majority-owned by Canadians under the federal Telecommunications Act, the CRTC is not responsible for enforcement of this provision. In fact, the commission does not require licences at all for telephone companies, and CRTC approval is therefore not generally required for the sale of a telephone company, unless said company also owns a broadcast licence.
Since 1987, the CRTC has been involved in several notable decisions, some of which led to controversy and debate.
Milestone Radio: In two separate rounds of licence hearings in the 1990s, the CRTC rejected applications by Milestone Radio to launch a radio station in Toronto which would have been Canada's first urban music station; in both cases, the CRTC instead granted licences to stations that duplicated formats already offered by other stations in the Toronto market. The decision has been widely cited as one of the single most significant reasons why Canadian hip hop had difficulty establishing its commercial viability throughout the 1990s. The CRTC finally granted a licence to Milestone in 2000, after a cabinet order-in-council directed the commission to licence two new radio stations that reflected the cultural diversity of the Toronto market, and CFXJ-FM launched in 2001.
CHOI-FM: The CRTC announced it would not renew the licence of the popular CHOI-FM radio station in Quebec City, after having previously sanctioned the station for failing to uphold its promise of performance and then, during the years following, receiving about 50 complaints about offensive behaviour by radio jockeys which similarly contravened CRTC rules on broadcast hate speech. Many thousands of the station's fans marched in the streets and on Parliament Hill against the decision, and the parent company of CHOI, Genex Corp., appealed the CRTC decision unsuccessfully to the Federal Court of Canada.
The station was later sold to RNC Media, but instead of renewing its licence the CRTC issued RNC a licence to launch a "new" radio station on the same frequency.
CBC Newsworld: The CRTC licensed the CBC on November 30, 1987 to provide a national all-news television network. Its competitor applicant, Alberta-based Allarcom, appealed this decision to the House of Commons of Canada. It was overturned and there were questions of whether federal politicians should meddle in CRTC decisions. Because of this the network launch was delayed from September 1, 1988 to July 31, 1989.
RAI International: In Summer 2004, this Italian government controlled channel was denied permission to broadcast independently in Canada on the grounds that it had acted and was likely to act contrary to established Canadian policies. RAI International's latest politically appointed President (an avowed right wing nationalist and former spokesperson for Giorgio Almirante, the leader of the post-fascist party of Italy) had unilaterally terminated a 20-year-old agreement and stripped all of its 1,500 to 2,000 annual hours of programming from Telelatino (TLN), a Canadian-run channel which had devoted 95% of its prime time schedule to RAI programs for 20 years since TLN was founded.
All Italian-Canadians were denied RAI programming by RAI International's removal of its programming from the Canadian marketplace, a move intended to create a public outcry and a threat that Canadians would resort to using satellite viewing cards obtained via the US in order to watch RAI, even though these cards were either grey market or black market, according to different analyses (see below). Following unprecedented foreign led and domestic political interference with the CRTC's quasi-judicial independent regulatory process, within six months of its original decision, an abrupt CRTC "review" of its policy on third-language foreign services determined to drop virtually all restrictions and adopt a new "open entry" approach to foreign controlled "third language" (non-English, non-French) channels.
Al Jazeera: Was approved by the CRTC in 2004 as an optional cable and satellite offering, but on the condition that any carrier distributing it must edit out any instances of illegal hate speech. Cable companies declared that these restrictions would make it too expensive to carry Al Jazeera. Although no cable company released data as to what such a monitoring service would cost, the end-result was that no cable company elected to carry the station, either, leaving many Arabic-speaking Canadians using free-to-air satellite dishes to watch the station.
The Canadian Jewish Congress has expressed its opinion over possible anti-Semitic incitement on this station and that the restrictions on Al Jazeera are appropriate, while the Canadian B'nai Brith is opposed to any approval of Al Jazeera in Canada. The CRTC ruling applied to Al Jazeera and not to its English-speaking sister network Al Jazeera English, which was launched two years after the ruling.
Fox News Channel: Until 2004, the CRTC's apparent reluctance to grant a digital licence to Fox News Channel under the same policy which made it difficult for RAI to enter the country – same-genre competition from foreign services – had angered many conservative Canadians, who believed the network was deliberately being kept out due to its perceived conservative bias, particularly given the long-standing availability of services such as CNN and BBC World in Canada.
On November 18, 2004, however, the CRTC approved an application by cable companies to offer Fox News Channel on the digital cable tier. Fox commenced broadcasting in Canada shortly thereafter.
Satellite radio: In June 2005, the CRTC outraged some Canadian cultural nationalists (such as the Friends of Canadian Broadcasting) and labour unions by licensing two companies, Canadian Satellite Radio and Sirius Canada to offer satellite radio services in Canada. The two companies are in partnership with American firms XM Satellite Radio and Sirius Satellite Radio respectively, and in accordance with the CRTC decision will only need to offer ten percent Canadian content. The CRTC contends that this low level of Canadian content, particularly when compared to the 35% rule on local radio stations, was necessary because unlicensed U.S. receivers were already flooding into the country, so that enforcing a ban on these receivers would be nearly impossible (see below).
This explanation did not satisfy cultural nationalists, who demanded that the federal cabinet overturn the decision and mandate a minimum of 35% Canadian content. Supporters of the decision argue that satellite radio can only be feasibly set up as a continental system, and trying to impose 35% Canadian content across North America is quite unrealistic. They also argue that satellite radio will boost Canadian culture by giving vital exposure to independent artists, instead of concentrating just on the country's stars, and point to the CRTC's successful extraction of promises to program 10% Canadian content on satellite services already operational in the United States as important concessions. Despite popular perception that the CRTC banned Sirius Canada from broadcasting Howard Stern's program, this is not the case. Sirius Canada in fact initially "chose" not to air Stern based on the "possibility" of a future issue with the CRTC, although the company reversed its decision and began offering Howard Stern in 2006.
2008 Ottawa radio licences: On November 21, 2008, federal Minister of Canadian Heritage and Official Languages James Moore issued a statement calling on the CRTC to review its approval of two new radio stations, Frank Torres' CIDG-FM and Astral Media's CJOT-FM, which it had licensed in August 2008 to serve the Ottawa-Gatineau radio market. Moore asked the commission to assess whether the francophone population of the Ottawa-Gatineau area was sufficiently well-served by existing French radio services, and to consider licensing one or more of the French language applications, which included a Christian music station, a community radio station and a campus radio station for the Université du Québec en Outaouais, in addition to or instead of the approved stations. The review ultimately identified a viable frequency for a third station, and CJFO-FM launched in 2010.
Bell Canada usage-based billing: On October 28, 2010, the CRTC handed down its final decision on how wholesale customers can be billed by large network owners. Under the plan which starts within 90 days, Bell will be able to charge wholesale service providers a flat monthly fee to connect to its network, and for a set monthly usage limit per each ISP customer the ISP has. Beyond that set limit, individual users will be charged per gigabyte, depending on the speed of their connections.
Customers using the fastest connections of five megabits per second, for example, will have a monthly allotment of 60 GB, beyond which Bell will charge $1.12 per GB to a maximum of $22.50. If a customer uses more than 300 GB a month, Bell will also be able to implement an additional charge of 75 cents per gigabyte. In May 2010, the CRTC ruled that Bell could not implement its usage-based billing system until all of its own retail customers had been moved off older, unlimited downloading plans. The requirement would have meant that Bell would have to move its oldest and most loyal customers.
The CRTC also added that Bell would be required to offer to wholesale ISPs the same usage insurance plan it sells to retail customers. Bell appealed both requirements, citing that the rules do not apply to cable companies and that they constituted proactive rate regulation by the CRTC, which goes against government official policy direction that the regulator only intervene in markets after a competitive problem has been proven. In Thursday's decision, the CRTC rescinded both requirements, thereby giving Bell the go-ahead to implement usage-based billing. This ruling according to Teksavvy handcuffs the competitive market. This has been asked by Stephen Harper and Parliament to have the decision reviewed. According to a tweet by Industry Minister Tony Clement, unless the CRTC reverses this decision, the government will use its override power to reverse the decision.
While an exact number has not been determined, thousands of Canadians have purchased and used what they contend to be grey market radio and television services, licensed in the United States but not in Canada. Users of these unlicensed services contend that they are not directly breaking any laws by simply using the equipment. The equipment is usually purchased from an American supplier (although some merchants have attempted to set up shop in Canada) and the services are billed to an American postal address. The advent of online billing and the easy availability of credit card services has made it relatively easy for almost anyone to maintain an account in good standing, regardless of where they actually live.
Sec. 9(1)(c) of the Radiocommunication Act creates a prohibition against all decoding of encrypted programming signals, followed by an exception where authorization is received from the person holding the lawful right in Canada to transmit and authorize decoding of the signal. This means receiving the encrypted programming of DishNetwork or DirecTV, even with a grey market subscription, may be construed as unlawful (this remains an unresolved Constitutional issue).
Notwithstanding, possession of DishNetwork or DirecTV equipment is not unlawful as provided by The Radiocommuncation Act Section 4(1)(b), which states:
"No person shall, except under and in accordance with a radio authorization, install, operate or possess radio apparatus, other than (b)a radio apparatus that is capable only of the reception of broadcasting and that is not a distribution undertaking. (radio apparatus" means a device or combination of devices intended for, or capable of being used for, radiocommunication)."
Satellite radio poses a more complicated problem for the CRTC. While an unlicensed satellite dish can often be identified easily, satellite radio receivers are much more compact and can rarely be easily identified, at least not without flagrantly violating provisions against unreasonable search and seizure in the Canadian Charter of Rights and Freedoms. Some observers argued that this influenced the CRTC's June 2005 decision to ease Canadian content restrictions on satellite radio (see above).
The CRTC is run by up to 13 full-time (including the chairman, the vice-chairman of broadcasting, and the vice-chairman of telecommunications) appointed by the Cabinet for renewable terms of up to five years. However, unlike the more directly political appointees of the American Federal Communications Commission, the CRTC is an arms-length regulatory body with more autonomous authority over telecommunications. For example, the CRTC's decisions rely more on a more judiciary process relying on evidence submitted during public consultations, rather than along party lines as the American FCC is prone to do.
The CRTC Interconnection Steering Committee (CISC) assists in developing information, procedures and guidelines for the CRTC's regulatory activities. | https://en.wikipedia.org/wiki?curid=5985 |
Coal
Coal is a combustible black or brownish-black sedimentary rock, formed as rock strata called coal seams. Coal is mostly carbon with variable amounts of other elements; chiefly hydrogen, sulfur, oxygen, and nitrogen.
Coal is formed when dead plant matter decays into peat and is converted into coal by the heat and pressure of deep burial over millions of years. Vast deposits of coal originate in former wetlands—called coal forests—that covered much of the Earth's tropical land areas during the late Carboniferous (Pennsylvanian) and Permian times.
As a fossil fuel burned for heat, coal supplies about a quarter of the world's primary energy and two-fifths of its electricity. Some iron and steel making and other industrial processes burn coal.
The extraction and use of coal causes many premature deaths and much illness. The coal industry damages the environment, including by climate change as it is the largest anthropogenic source of carbon dioxide, 14 gigatonnes (Gt) in 2016, which is 40% of the total fossil fuel emissions and almost 25% of total global greenhouse gas emissions. As part of the worldwide energy transition many countries have stopped using or use less coal, and the UN Secretary General has asked governments to stop building new coal plants by 2020. Coal use peaked in 2013 but to meet the Paris Agreement target of keeping global warming to well below coal use needs to halve from 2020 to 2030.
The largest consumer and importer of coal is China. China mines almost half the world's coal, followed by India with about a tenth. Australia accounts for about a third of world coal exports followed by Indonesia and Russia.
The word originally took the form "col" in Old English, from Proto-Germanic *"kula"("n"), which in turn is hypothesized to come from the Proto-Indo-European root *"g"("e")"u-lo-" "live coal". Germanic cognates include the Old Frisian "kole", Middle Dutch "cole", Dutch "kool", Old High German "chol", German "Kohle" and Old Norse "kol", and the Irish word "gual" is also a cognate via the Indo-European root.
Coal is composed of macerals, minerals and water. Fossils and amber may be found in coal.
One theory states that about 360 million years ago, some plants evolved the ability to produce lignin, a complex polymer that made their cellulose stems much harder and more woody. Thus, the first trees evolved. But bacteria and fungi did not immediately evolve the ability to decompose lignin, so the wood did not fully decay but became buried under sediment, eventually turning into coal. About 300 million years ago, mushrooms and other fungi developed this ability, ending the main coal-formation period of earth's history.
At various times in the geologic past, the Earth had dense forests in low-lying wetland areas. Due to natural processes such as flooding, these forests were buried underneath soil. As more and more soil deposited over them, they were compressed. The temperature also rose as they sank deeper and deeper. As the process continued the plant matter was protected from biodegradation and oxidation, usually by mud or acidic water. This trapped the carbon in immense peat bogs that were eventually covered and deeply buried by sediments. Under high pressure and high temperature, dead vegetation was slowly converted to coal. The conversion of dead vegetation into coal is called coalification. Coalification starts with dead plant matter decaying into peat. Then over millions of years the heat and pressure of deep burial causes the loss of water, methane and carbon dioxide and an increase in the proportion of carbon. Thus first lignite (also called "brown coal"), then sub-bituminous coal, bituminous coal, and lastly anthracite (also called "hard coal" or "black coal") may be formed.
The wide, shallow seas of the Carboniferous Period provided ideal conditions for coal formation, although coal is known from most geological periods. The exception is the coal gap in the Permian–Triassic extinction event, where coal is rare. Coal is known from Precambrian strata, which predate land plants—this coal is presumed to have originated from residues of algae.
Sometimes coal seams (also known as coal beds) are interbedded with other sediments in a cyclothem.
As geological processes apply pressure to dead biotic material over time, under suitable conditions, its metamorphic grade or rank increases successively into:
Cannel coal (sometimes called "candle coal") is a variety of fine-grained, high-rank coal with significant hydrogen content, which consists primarily of liptinite.
There are several international standards for coal. The classification of coal is generally based on the content of volatiles. However the most important distinction is between thermal coal (also known as steam coal), which is burnt to generate electricity via steam; and metallurgical coal (also known as coking coal), which is burnt at high temperature to make steel.
Hilt's law is a geological observation that (within a small area) the deeper the coal is found, the higher its rank (or grade). It applies if the thermal gradient is entirely vertical; however, metamorphism may cause lateral changes of rank, irrespective of depth.
The earliest recognized use is from the Shenyang area of China where by 4000 BC Neolithic inhabitants had begun carving ornaments from black lignite. Coal from the Fushun mine in northeastern China was used to smelt copper as early as 1000 BC. Marco Polo, the Italian who traveled to China in the 13th century, described coal as "black stones ... which burn like logs", and said coal was so plentiful, people could take three hot baths a week. In Europe, the earliest reference to the use of coal as fuel is from the geological treatise "On stones" (Lap. 16) by the Greek scientist Theophrastus (c. 371–287 BC):
Outcrop coal was used in Britain during the Bronze Age (3000–2000 BC), where it formed part of funeral pyres. In Roman Britain, with the exception of two modern fields, "the Romans were exploiting coals in all the major coalfields in England and Wales by the end of the second century AD". Evidence of trade in coal, dated to about AD 200, has been found at the Roman settlement at Heronbridge, near Chester; and in the Fenlands of East Anglia, where coal from the Midlands was transported via the Car Dyke for use in drying grain. Coal cinders have been found in the hearths of villas and Roman forts, particularly in Northumberland, dated to around AD 400. In the west of England, contemporary writers described the wonder of a permanent brazier of coal on the altar of Minerva at Aquae Sulis (modern day Bath), although in fact easily accessible surface coal from what became the Somerset coalfield was in common use in quite lowly dwellings locally. Evidence of coal's use for iron-working in the city during the Roman period has been found. In Eschweiler, Rhineland, deposits of bituminous coal were used by the Romans for the smelting of iron ore.
No evidence exists of the product being of great importance in Britain before about AD 1000, the High Middle Ages. Mineral coal came to be referred to as "seacoal" in the 13th century; the wharf where the material arrived in London was known as Seacoal Lane, so identified in a charter of King Henry III granted in 1253. Initially, the name was given because much coal was found on the shore, having fallen from the exposed coal seams on cliffs above or washed out of underwater coal outcrops, but by the time of Henry VIII, it was understood to derive from the way it was carried to London by sea. In 1257–1259, coal from Newcastle upon Tyne was shipped to London for the smiths and lime-burners building Westminster Abbey. Seacoal Lane and Newcastle Lane, where coal was unloaded at wharves along the River Fleet, still exist.
These easily accessible sources had largely become exhausted (or could not meet the growing demand) by the 13th century, when underground extraction by shaft mining or adits was developed. The alternative name was "pitcoal", because it came from mines. The development of the Industrial Revolution led to the large-scale use of coal, as the steam engine took over from the water wheel. In 1700, five-sixths of the world's coal was mined in Britain. Britain would have run out of suitable sites for watermills by the 1830s if coal had not been available as a source of energy. In 1947 there were some 750,000 miners in Britain but the last deep coal mine in the UK closed in 2015.
A grade between bituminous coal and anthracite was once known as "steam coal" as it was widely used as a fuel for steam locomotives. In this specialized use, it is sometimes known as "sea coal" in the United States. Small "steam coal", also called "dry small steam nuts" (or DSSN), was used as a fuel for domestic water heating.
Coal continues to arrive on beaches around the world from both natural erosion of exposed coal seams and windswept spills from cargo ships. Many homes in such areas gather this coal as a significant, and sometimes primary, source of home heating fuel.
Emission intensity is the greenhouse gas emitted over the life of a generator per unit of electricity generated. Of the currently widely used methods of generating electricity the emission intensity of coal and oil is high, as they emit around 1000g of CO2eq for each kWh generated; natural gas is medium emission intensity at around 500g CO2eq per kWh; and all other methods are typically low emission intensity of under 100g per kWh. The emission intensity of coal varies with type and generator technology and exceeds 1200g per kWh in some countries.
The energy density of coal, that is its heating value, is roughly 24 megajoules per kilogram (approximately 6.7 kilowatt-hours per kg). For a coal power plant with a 40% efficiency, it takes an estimated of coal to power a 100 W lightbulb for one year.
27.6% of world energy was supplied by coal in 2017 and Asia used almost three quarters of it.
The composition of coal is reported either as a proximate analysis (moisture, volatile matter, fixed carbon, and ash) or an ultimate analysis (ash, carbon, hydrogen, nitrogen, oxygen, and sulfur). The "volatile matter" does not exist by itself (except for some adsorbed methane) but designates the volatile compounds that are produced and driven off by heating the coal. A typical bituminous coal may have an ultimate analysis on a dry, ash-free basis of 84.4% carbon, 5.4% hydrogen, 6.7% oxygen, 1.7% nitrogen, and 1.8% sulfur, on a weight basis.
The composition of ash, given in terms of oxides, varies:
Other minor components include:
Coke is a solid carbonaceous residue derived from coking coal (a low-ash, low-sulfur bituminous coal, also known as "metallurgical coal"), which is used in manufacturing steel and other iron products. Coke is made from coking coal by baking in an oven without oxygen at temperatures as high as 1,000 °C, driving off the volatile constituents and fusing together the fixed carbon and residual ash. Metallurgical coke is used as a fuel and as a reducing agent in smelting iron ore in a blast furnace. The carbon monoxide produced by its combustion reduces hematite (an iron oxide) to iron.
Waste carbon dioxide is also produced together with pig iron, which is too rich in dissolved carbon so must be treated further to make steel.
Coking coal should be low in ash, sulfur, and phosphorus, so that these do not migrate to the metal.
The coke must be strong enough to resist the weight of overburden in the blast furnace, which is why coking coal is so important in making steel using the conventional route. Coke from coal is grey, hard, and porous and has a heating value of 29.6 MJ/kg. Some cokemaking processes produce byproducts, including coal tar, ammonia, light oils, and coal gas.
Petroleum coke (petcoke) is the solid residue obtained in oil refining, which resembles coke but contains too many impurities to be useful in metallurgical applications.
Finely ground bituminous coal, known in this application as sea coal, is a constituent of foundry sand. While the molten metal is in the mould, the coal burns slowly, releasing reducing gases at pressure, and so preventing the metal from penetrating the pores of the sand. It is also contained in 'mould wash', a paste or liquid with the same function applied to the mould before casting. Sea coal can be mixed with the clay lining (the "bod") used for the bottom of a cupola furnace. When heated, the coal decomposes and the bod becomes slightly friable, easing the process of breaking open holes for tapping the molten metal.
Scrap steel can be recycled in an electric arc furnace; and an alternative to making iron by smelting is direct reduced iron, where any carbonaceous fuel can be used to make sponge or pelletised iron. To lessen carbon dioxide emissions hydrogen can be used as the reducing agent and biomass or waste as the source of carbon. Historically, charcoal has been used as an alternative to coke in a blast furnace, with the resultant iron being known as charcoal iron.
Coal gasification, as part of an integrated gasification combined cycle (IGCC) coal-fired power station, is used to produce syngas, a mixture of carbon monoxide (CO) and the hydrogen (H2) gas to fire gas turbines to produce electricity. Syngas can also be converted into transportation fuels, such as gasoline and diesel, through the Fischer-Tropsch process; alternatively, syngas can be converted into methanol, which can be blended into fuel directly or converted to gasoline via the methanol to gasoline process. Gasification combined with Fischer-Tropsch technology was used by the Sasol chemical company of South Africa to make chemicals and motor vehicle fuels from coal.
During gasification, the coal is mixed with oxygen and steam while also being heated and pressurized. During the reaction, oxygen and water molecules oxidize the coal into carbon monoxide (CO), while also releasing hydrogen gas (H2). This used to be done in underground coal mines, and also to make town gas which was piped to customers to burn for illumination, heating, and cooking.
If the refiner wants to produce gasoline, the syngas is routed into a Fischer-Tropsch reaction. This is known as indirect coal liquefaction. If hydrogen is the desired end-product, however, the syngas is fed into the water gas shift reaction, where more hydrogen is liberated:
Coal can be converted directly into synthetic fuels equivalent to gasoline or diesel by hydrogenation or carbonization. Coal liquefaction emits more carbon dioxide than liquid fuel production from crude oil. Mixing in biomass and using CCS would emit slightly less than the oil process but at a high cost. State owned China Energy Investment runs a coal liquefaction plant and plans to build 2 more.
Coal liquefaction may also refer to the cargo hazard when shipping coal.
Chemicals have been produced from coal since the 1950s. Coal can be used as a feedstock in the production of a wide range of chemical fertilizers and other chemical products. The main route to these products was coal gasification to produce syngas. Primary chemicals that are produced directly from the syngas include methanol, hydrogen and carbon monoxide, which are the chemical building blocks from which a whole spectrum of derivative chemicals are manufactured, including olefins, acetic acid, formaldehyde, ammonia, urea and others. The versatility of syngas as a precursor to primary chemicals and high-value derivative products provides the option of using coal to produce a wide range of commodities. In the 21st century, however, the use of coal bed methane is becoming more important.
Because the slate of chemical products that can be made via coal gasification can in general also use feedstocks derived from natural gas and petroleum, the chemical industry tends to use whatever feedstocks are most cost-effective. Therefore, interest in using coal tended to increase for higher oil and natural gas prices and during periods of high global economic growth that might have strained oil and gas production.
Coal to chemical processes require substantial quantities of water. Much coal to chemical production is in China where coal dependent provinces such as Shanxi are struggling to control its pollution.
Refined coal is the product of a coal-upgrading technology that removes moisture and certain pollutants from lower-rank coals such as sub-bituminous and lignite (brown) coals. It is one form of several precombustion treatments and processes for coal that alter coal's characteristics before it is burned. Thermal efficiency improvements are achievable by improved pre-drying (especially relevant with high-moisture fuel such as lignite or biomass). The goals of precombustion coal technologies are to increase efficiency and reduce emissions when the coal is burned. Precombustion technology can sometimes be used as a supplement to postcombustion technologies to control emissions from coal-fueled boilers.
Coal burnt as a solid fuel in coal power stations to generate electricity is called thermal coal. Coal is also used to produce very high temperatures through combustion. Early deaths due to air pollution have been estimated at 200 per GW-year, however they may be higher around power plants where scrubbers are not used or lower if they are far from cities. Efforts around the world to reduce the use of coal have led some regions to switch to natural gas and electricity from lower carbon sources.
When coal is used for electricity generation, it is usually pulverized and then burned in a furnace with a boiler. The furnace heat converts boiler water to steam, which is then used to spin turbines which turn generators and create electricity. The thermodynamic efficiency of this process varies between about 25% and 50% depending on the pre-combustion treatment, turbine technology (e.g. supercritical steam generator) and the age of the plant.
A few integrated gasification combined cycle (IGCC) power plants have been built, which burn coal more efficiently. Instead of pulverizing the coal and burning it directly as fuel in the steam-generating boiler, the coal is gasified to create syngas, which is burned in a gas turbine to produce electricity (just like natural gas is burned in a turbine). Hot exhaust gases from the turbine are used to raise steam in a heat recovery steam generator which powers a supplemental steam turbine. The overall plant efficiency when used to provide combined heat and power can reach as much as 94%. IGCC power plants emit less local pollution than conventional pulverized coal-fueled plants; however the technology for carbon capture and storage after gasification and before burning has so far proved to be too expensive to use with coal. Other ways to use coal are as coal-water slurry fuel (CWS), which was developed in the Soviet Union, or in an MHD topping cycle. However these are not widely used due to lack of profit.
In 2017 38% of the world's electricity came from coal, the same percentage as 30 years previously. In 2018 global installed capacity was 2TW (of which 1TW is in China) which was 30% of total electricity generation capacity. The most dependent major country is South Africa, with over 80% of its electricity generated by coal.
Maximum use of coal was reached in 2013. In 2018 coal-fired power station capacity factor averaged 51%, that is they operated for about half their available operating hours.
About 8000 Mt of coal are produced annually, about 90% of which is hard coal and 10% lignite. just over half is from underground mines. More accidents occur during underground mining than surface mining. Not all countries publish mining accident statistics so worldwide figures are uncertain, but it is thought that most deaths occur in coal mining accidents in China: in 2017 there were 375 coal mining related deaths in China. Most coal mined is thermal coal (also called steam coal as it is used to make steam to generate electricity) but metallurgical coal (also called "metcoal" or "coking coal" as it is used to make coke to make iron) accounts for 10% to 15% of global coal use.
China mines almost half the world's coal, followed by India with about a tenth. Australia accounts for about a third of world coal exports, followed by Indonesia and Russia; while the largest importers are Japan and India.
The price of metallurgical coal is volatile and much higher than the price of thermal coal because metallurgical coal must be lower in sulfur and requires more cleaning. Coal futures contracts provide coal producers and the electric power industry an important tool for hedging and risk management.
In some countries new onshore wind or solar generation already costs less than coal power from existing plants (see Cost of electricity by source).
However, for China this is forecast for the early 2020s and for south-east Asia not until the late 2020s. In India building new plants is uneconomic and, despite being subsidized, existing plants are losing market share to renewables.
Of the countries which produce coal China mines by far the most, almost half the world's coal, followed by less than 10% by India. China is also by far the largest consumer. Therefore, market trends depend on Chinese energy policy. Although the effort to reduce pollution means that the global long-term trend is to burn less coal, the short and medium term trends may differ, in part due to Chinese financing of new coal-fired power plants in other countries.
Countries with annual production higher than 300 million tonnes are shown.
Countries with annual consumption higher than 500 million tonnes are shown. Shares are based on data expressed in tonnes oil equivalent.
Exporters are at risk of a reduction in import demand from India and China.
The use of coal as fuel causes ill health and deaths. Mining and processing of coal causes air and water pollution. Coal-powered plants emit nitrogen oxides, sulfur dioxide, particulate pollution and heavy metals, which adversely affect human health. Coal bed methane extraction is important to avoid mining accidents.
The deadly London smog was caused primarily by the heavy use of coal. Globally coal is estimated to cause 800,000 premature deaths every year, mostly in India and China.
Burning coal is a major emitter of sulfur dioxide, which creates PM2.5 particulates, the most dangerous form of air pollution.
Coal smokestack emissions cause asthma, strokes, reduced intelligence, artery blockages, heart attacks, congestive heart failure, cardiac arrhythmias, mercury poisoning, arterial occlusion, and lung cancer.
Annual health costs in Europe from use of coal to generate electricity are estimated at up to €43 billion.
In China, improvements to air quality and human health would increase with more stringent climate policies, mainly because the country's energy is so heavily reliant on coal. And there would be a net economic benefit.
A 2017 study in the "Economic Journal" found that for Britain during the period 1851–1860, "a one standard deviation increase in coal use raised infant mortality by 6–8% and that industrial coal use explains roughly one-third of the urban mortality penalty observed during this period."
Breathing in coal dust causes coalworker's pneumoconiosis which is known colloquially as "black lung", so-called because the coal dust literally turns the lungs black from their usual pink color. In the United States alone, it is estimated that 1,500 former employees of the coal industry die every year from the effects of breathing in coal mine dust.
Huge amounts of coal ash and other waste is produced annually. Use of coal generates hundreds of millions of tons of ash and other waste products every year. These include fly ash, bottom ash, and flue-gas desulfurization sludge, that contain mercury, uranium, thorium, arsenic, and other heavy metals, along with non-metals such as selenium.
Around 10% of coal is ash: coal ash is hazardous and toxic to human beings and some other living things. Coal ash contains the radioactive elements uranium and thorium. Coal ash and other solid combustion byproducts are stored locally and escape in various ways that expose those living near coal plants to radiation and environmental toxics.
Coal mining and coal fueling of power stations and industrial processes can cause major environmental damage.
Water systems are affected by coal mining. For example, mining affects groundwater and water table levels and acidity. Spills of fly ash, such as the Kingston Fossil Plant coal fly ash slurry spill, can also contaminate land and waterways, and destroy homes. Power stations that burn coal also consume large quantities of water. This can affect the flows of rivers, and has consequential impacts on other land uses. In areas of water scarcity, such as the Thar Desert in Pakistan, coal mining and coal power plants would use significant quantities of water.
One of the earliest known impacts of coal on the water cycle was acid rain. Approximately 75 Tg/S per year of sulfur dioxide (SO2) is released from burning coal. After release, the sulfur dioxide is oxidized to gaseous H2SO2 which scatters solar radiation, hence its increase in the atmosphere exerts a cooling effect on climate. This beneficially masks some of the warming caused by increased greenhouse gases. However, the sulfur is precipitated out of the atmosphere as acid rain in a matter of weeks, whereas carbon dioxide remains in the atmosphere for hundreds of years. Release of SO2 also contributes to the widespread acidification of ecosystems.
Disused coal mines can also cause issues. Subsidence can occur above tunnels, causing damage to infrastructure or cropland. Coal mining can also cause long lasting fires, and it has been estimated that thousands of coal seam fires are burning at any given time. For example, Brennender Berg has been burning since 1668 and is still burning in the 21st century.
The production of coke from coal produces ammonia, coal tar, and gaseous compounds as by-products which if discharged to land, air or waterways can pollute the environment. The Whyalla steelworks is one example of a coke producing facility where liquid ammonia is discharged to the marine environment.
Thousands of coal fires are burning around the world. Those burning underground can be difficult to locate and many cannot be extinguished. Fires can cause the ground above to subside, their combustion gases are dangerous to life, and breaking out to the surface can initiate surface wildfires. Coal seams can be set on fire by spontaneous combustion or contact with a mine fire or surface fire. Lightning strikes are an important source of ignition. The coal continues to burn slowly back into the seam until oxygen (air) can no longer reach the flame front. A grass fire in a coal area can set dozens of coal seams on fire. Coal fires in China burn an estimated 120 million tons of coal a year, emitting 360 million metric tons of CO2, amounting to 2–3% of the annual worldwide production of CO2 from fossil fuels. In Centralia, Pennsylvania (a borough located in the Coal Region of the United States), an exposed vein of anthracite ignited in 1962 due to a trash fire in the borough landfill, located in an abandoned anthracite strip mine pit. Attempts to extinguish the fire were unsuccessful, and it continues to burn underground to this day. The Australian Burning Mountain was originally believed to be a volcano, but the smoke and ash come from a coal fire that has been burning for some 6,000 years.
At Kuh i Malik in Yagnob Valley, Tajikistan, coal deposits have been burning for thousands of years, creating vast underground labyrinths full of unique minerals, some of them very beautiful.
The reddish siltstone rock that caps many ridges and buttes in the Powder River Basin in Wyoming and in western North Dakota is called "porcelanite", which resembles the coal burning waste "clinker" or volcanic "scoria". Clinker is rock that has been fused by the natural burning of coal. In the Powder River Basin approximately 27 to 54 billion tons of coal burned within the past three million years. Wild coal fires in the area were reported by the Lewis and Clark Expedition as well as explorers and settlers in the area.
The largest and most long-term effect of coal use is the release of carbon dioxide, a greenhouse gas that causes climate change and global warming. Coal-fired power plants were the single largest contributor to the growth in global CO2 emissions in 2018, 40% of the total fossil fuel emissions. Coal mining can emit methane, another greenhouse gas.
In 2016 world gross carbon dioxide emissions from coal usage were 14.5 gigatonnes. For every megawatt-hour generated, coal-fired electric power generation emits around a tonne of carbon dioxide, which is double the approximately 500 kg of carbon dioxide released by a natural gas-fired electric plant. In 2013, the head of the UN climate agency advised that most of the world's coal reserves should be left in the ground to avoid catastrophic global warming. To keep global warming below 1.5 °C or 2 °C hundreds, or possibly thousands, of coal-fired power plants will need to be retired early.
"Clean" coal technology usually addresses atmospheric problems resulting from burning coal. Historically, the primary focus was on SO2 and NOx, the most important gases which caused acid rain; and particulates which cause visible air pollution, illness and premature deaths. SO2 can be removed by flue-gas desulfurization and NO2 by selective catalytic reduction (SCR). Particulates can be removed with electrostatic precipitators. Although perhaps less efficient wet scrubbers can remove both gases and particulates. And mercury emissions can be reduced up to 95%. However capturing carbon dioxide emissions is generally not economically viable.
Local pollution standards include GB13223-2011 (China), India, the Industrial Emissions Directive (EU) and the Clean Air Act (United States).
Satellite monitoring is now used to crosscheck national data, for example Sentinel-5 Precursor has shown that Chinese control of SO2 has only been partially successful. It has also revealed that low use of technology such as SCR has resulted in high NO2 emissions in South Africa and India.
A few Integrated gasification combined cycle (IGCC) coal-fired power plants have been built with coal gasification. Although they burn coal more efficiently and therefore emit less pollution, the technology has not generally proved economically viable for coal, except possibly in Japan although this is controversial.
Although still being intensively researched and considered economically viable for some uses other than with coal; carbon capture and storage has been tested at the Petra Nova and Boundary Dam coal-fired power plants and has been found to be technically feasible but not economically viable for use with coal, due to reductions in the cost of solar PV technology.
In 2018 USD 80 billion was invested in coal supply but almost all for sustaining production levels rather than opening new mines.
In the long term coal and oil could cost the world trillions of dollars per year. Coal alone may cost Australia billions, whereas costs to some smaller companies or cities could be on the scale of millions of dollars. The economies most damaged by coal (via climate change) may be India and the US as they are the countries with the highest social cost of carbon. Bank loans to finance coal are a risk to the Indian economy.
China is the largest producer of coal in the world. It is the world's largest energy consumer, and coal in China supplies 60% of its primary energy. However two fifths of China's coal power stations are estimated to be loss-making.
Air pollution from coal storage and handling costs the USA almost 200 dollars for every extra ton stored, due to PM2.5. Coal pollution costs the EU €43 billion each year. Measures to cut air pollution benefit individuals financially and the economies of countries such as China.
Broadly defined total subsidies for coal in 2015 have been estimated at around US$2.5 trillion, about 3% of global GDP. G20 countries provide at least US$63.9 billion of government support per year for the production of coal, including coal-fired power: many subsidies are impossible to quantify but they include US$27.6 billion in domestic and international public finance, US$15.4 billion in fiscal support, and US$20.9 billion in state-owned enterprise (SOE) investments per year. In the EU state aid to new coal-fired plants is banned from 2020, and to existing coal-fired plants from 2025. However government funding for new coal power plants is being supplied via Exim Bank of China, the Japan Bank for International Cooperation and Indian public sector banks. Coal in Kazakhstan was the main recipient of coal consumption subsidies totalling US$2 billion in 2017. Coal in Turkey benefited from substantial subsidies.
Some coal-fired power stations could become stranded assets, for example China Energy Investment, the world's largest power company, risks losing half its capital. However state owned electricity utilities such as Eskom in South Africa, Perusahaan Listrik Negara in Indonesia, Sarawak Energy in Malaysia, Taipower in Taiwan, EGAT in Thailand, Vietnam Electricity and EÜAŞ in Turkey are building or planning new plants. this may be helping to cause a carbon bubble which could cause financial instability if it bursts.
Countries building or financing new coal-fired power stations, such as China, India, and Japan, face mounting international criticism for obstructing the aims of the Paris Agreement. In 2019, the Pacific Island nations (in particular Vanuatu and Fiji) criticized Australia for failing to cut their emissions at a faster rate than they were, citing concerns about coastal inundation and erosion.
Allegations of corruption are being investigated in India and China.
Opposition to coal pollution was one of the main reasons the modern environmental movement started in the 19th century.
In order to meet global climate goals and provide power to those that don't currently have it coal power must be reduced from nearly 10,000 TWh to less than 2,000 TWh by 2040. Phasing out coal has short-term health and environmental benefits which exceed the costs, but some countries still favor coal, and there is much disagreement about how quickly it should be phased out. However many countries, such as the Powering Past Coal Alliance, have already or are transitioned away from coal; the largest transition announced so far being Germany, which is due to shut down its last coal-fired power station between 2035 and 2038. Some countries use the ideas of a "Just Transition", for example to use some of the benefits of transition to provide early pensions for coal miners. However low-lying Pacific Islands are concerned the transition is not fast enough and that they will be inundated by sea level rise; so they have called for OECD countries to completely phase out coal by 2030 and other countries by 2040.
Although many countries have coal underground not all will be consumed.
Of the three fossil fuels, coal has the most widely distributed resources. Coal is mined on all continents except Antarctica. However many such resources have no economic value (much value has been destroyed by shale gas fracking). The largest resources are found in the United States, Russia, China, Australia and India.
Nowadays "peak coal" means the point in time when consumption of coal reaches a maximum. Coal use peaked in 2013, then rose, but not to the former level, in 2018: the drop in 2019 was only about 1% but is likely to continue to fall.
Coal-fired generation puts out about twice the amount of carbon dioxide—around a tonne for every megawatt hour generated—than electricity generated by burning natural gas at 500 kg of greenhouse gas per megawatt hour. In addition to generating electricity, natural gas is also popular in some countries for heating and as an automotive fuel.
The use of coal in the United Kingdom declined as a result of the development of North Sea oil and the subsequent dash for gas during the 1990s. In Canada some coal power plants, such as the Hearn Generating Station, switched from coal to natural gas. In 2017, coal power in the United States provided 30% of the electricity, down from approximately 49% in 2008, due to plentiful supplies of low cost natural gas obtained by hydraulic fracturing of tight shale formations.
Some coal-mining regions are highly dependent on coal.
Some coal miners are concerned their jobs may be lost in the transition.
The white rot fungus "Trametes versicolor" can grow on and metabolize naturally occurring coal. The bacteria Diplococcus has been found to degrade coal, raising its temperature.
Coal is the official state mineral of Kentucky and the official state rock of Utah; both U.S. states have a historic link to coal mining. Some cultures hold that children who misbehave will receive only a lump of coal from Santa Claus for Christmas in their christmas stockings instead of presents. It is also customary and considered lucky in Scotland and the North of England to give coal as a gift on New Year's Day. This occurs as part of First-Footing and represents warmth for the year to come. | https://en.wikipedia.org/wiki?curid=5987 |
Traditional Chinese medicine
Traditional Chinese medicine (TCM) is a branch of traditional medicine that is said to be based on more than 3,500 years of Chinese medical practice that includes various forms of herbal medicine, acupuncture, cupping therapy, gua sha, massage (tui na), bonesetter (die-da), exercise (qigong), and dietary therapy. In some practices, modern Chinese medicine is a blend of TCM and Western medicine, although there is no high-quality clinical evidence that TCM is safe or effective for treating any disease.
There is no scientific evidence for traditional Chinese concepts such as "qi", meridians, and acupuncture points. The TCM theory and practice are not based upon scientific knowledge, and there is disagreement between TCM practitioners on what diagnosis and treatments should be used for any given person. The effectiveness of Chinese herbal medicine remains poorly researched and supported, and most of its treatments have no logical mechanism of action. There are concerns over a number of potentially toxic plants, animal parts, and mineral Chinese compounds. There are some concerns over illegal trade and transport of endangered species including rhinoceroses and tigers, and the welfare of specially farmed animals including bears. A review of cost-effectiveness research for TCM found that studies had low levels of evidence, with no beneficial outcomes. Pharmaceutical research has explored the potential for creating new drugs from traditional remedies, with few successful results. Proponents suggest that research has so far missed key features of the art of TCM, such as unknown interactions between various ingredients and complex interactive biological systems.
A "Nature" editorial described TCM as "fraught with pseudoscience", and said that the most obvious reason why it has not delivered many cures is that the majority of its treatments have no logical mechanism of action. One of the basic tenets of TCM is that the body's vital energy ("ch'i" or "qi") is circulating through channels called meridians having branches connected to bodily organs and functions. The concept of vital energy is considered pseudoscience. Concepts of the body and of disease used in TCM reflect its ancient origins and its emphasis on dynamic processes over material structure, similar to European humoral theory.
TCM is widely used in the Sinosphere, where it has a long history. The doctrines of Chinese medicine are rooted in books such as the "Yellow Emperor's Inner Canon" and the "Treatise on Cold Damage", as well as in cosmological notions such as yin–yang and the five phases. Starting in the 1950s, these precepts were standardized in the People's Republic of China, including attempts to integrate them with modern notions of anatomy and pathology. In the 1950s, the Chinese government promoted a systematized form of TCM.
Traces of therapeutic activities in China date from the Shang dynasty (14th–11th centuries BCE). Though the Shang did not have a concept of "medicine" as distinct from other fields, their oracular inscriptions on bones and tortoise shells refer to illnesses that affected the Shang royal family: eye disorders, toothaches, bloated abdomen, etc., which Shang elites usually attributed to curses sent by their ancestors. There is currently no evidence that the Shang nobility used herbal remedies. According to a 2006 overview, the "Documentation of Chinese materia medica (CMM) dates back to around 1,100 BCE when only dozens of drugs were first described. By the end of the 16th century, the number of drugs documented had reached close to 1,900. And by the end of the last century, published records of CMM had reached 12,800 drugs."
Stone and bone needles found in ancient tombs led Joseph Needham to speculate that acupuncture might have been carried out in the Shang dynasty. This being said, most historians now make a distinction between medical lancing (or bloodletting) and acupuncture in the narrower sense of using metal needles to attempt to treat illnesses by stimulating points along circulation channels ("meridians") in accordance with beliefs related to the circulation of "Qi". The earliest evidence for acupuncture in this sense dates to the second or first century BCE.
The "Yellow Emperor's Inner Canon (Huangdi Nei Jing)", the oldest received work of Chinese medical theory, was compiled during the Han dynasty around the 1st century BCE on the basis of shorter texts from different medical lineages. Written in the form of dialogues between the legendary Yellow Emperor and his ministers, it offers explanations on the relation between humans, their environment, and the cosmos, on the contents of the body, on human vitality and pathology, on the symptoms of illness, and on how to make diagnostic and therapeutic decisions in light of all these factors. Unlike earlier texts like "Recipes for Fifty-Two Ailments", which was excavated in the 1970s from a tomb that had been sealed in 168 BCE, the "Inner Canon" rejected the influence of spirits and the use of magic. It was also one of the first books in which the cosmological doctrines of Yinyang and the Five Phases were brought to a mature synthesis.
The "Treatise on Cold Damage Disorders and Miscellaneous Illnesses (Shang Han Lun)" was collated by Zhang Zhongjing sometime between 196 and 220 CE; at the end of the Han dynasty. Focusing on drug prescriptions rather than acupuncture, it was the first medical work to combine Yinyang and the Five Phases with drug therapy. This formulary was also the earliest public Chinese medical text to group symptoms into clinically useful "patterns" ("zheng" ) that could serve as targets for therapy. Having gone through numerous changes over time, the formulary now circulates as two distinct books: the "Treatise on Cold Damage Disorders" and the "Essential Prescriptions of the Golden Casket", which were edited separately in the eleventh century, under the Song dynasty.
Nan Jing (Chinese medicine) was originally called "The Yellow Emperor Eighty-one Nan Jing", the book is rumoured to be authored by Bian que in the eastern Han dynasty. This book has been compiled in the form of question and answer explanations. A total of 81 questions have been discussed. Therefore, it is also called "Eighty-One Nan". The book is based on basic theory and has also analyzed some disease certificates. Questions one to twenty-two is about pulse study, questions twenty-three to twenty-nine is about meridian study, questions thirty to forty-seven is related to urgent illnesses, questions forty-eight to sixty-one is related to serious diseases, questions sixty-two to sixty-eight is related to acupuncture points, and questions sixty-nine to eighty-one is related to the needle point methods.
The book is credited as developing its own path, while also inheriting the theories from Huangdi Neijing. The content includes physiology, pathology, diagnosis, treatment contents, and a more essential and specific discussion of pulse diagnosis. It has become one of the four classics for Chinese medicine practitioners to learn from and has impacted the medical development in China.
"Shennong Ben Cao Jing" is one of the earliest written medical books in China. Written during the Eastern Han Dynasty between 200 and 250 CE, it was the combined effort of TCM practitioners in the Qin and Han Dynasties who summarized, collected and compiled the results of pharmacological experience during their time periods. It was the first systematic summary of Chinese herbal medicine. Most of the pharmacological theories and compatibility rules and the proposed "seven emotions and harmony" principle have played a huge role in the practice of medicine for thousands of years in Chinese medicine. Therefore, it has long been a textbook for doctors and pharmacists to learn Chinese pharmacy, and it is also one of the necessary books for medical workers in China. The full text of "Shennong Ben Cao Jing" in English can be found online.
In the centuries that followed, several shorter books tried to summarize or systematize its contents of the book "Yellow Emperor's Inner Canon". The "Canon of Problems" (probably second century CE) tried to reconcile divergent doctrines from the "Inner Canon" and developed a complete medical system centered on needling therapy. The "AB Canon of Acupuncture and Moxibustion" ("Zhenjiu jiayi jing" , compiled by Huangfu Mi sometime between 256 and 282 CE) assembled a consistent body of doctrines concerning acupuncture; whereas the "Canon of the Pulse" ("Maijing" ; ca. 280) presented itself as a "comprehensive handbook of diagnostics and therapy."
In 1950, Chairman Mao Zedong made a speech in support of traditional Chinese medicine (TCM) which was influenced by political necessity. Mao believed he and the Chinese Communist Party should promote TCM but he did not personally believe in TCM and did not use it. In 1952, the president of the Chinese Medical Association said that, "This One Medicine, will possess a basis in modern natural sciences, will have absorbed the ancient and the new, the Chinese and the foreign, all medical achievements—and will be China's New Medicine!"
Then came the Cultural Revolution (1966–1978). Traditional Chinese medicine was strongly affected during this period. The development of traditional medicine in China was part of the pursuit of national identity during the Cultural Revolution. During this period, the Chinese government made large investments in traditional medicine to try to develop affordable medical care and public health facilities. Modernity, cultural identity and China's social and economic reconstruction are the main aspects of the Cultural Revolution. Compared to the colonial and feudal past, this movement tried to define a new and modern China.
The Chinese government has established a grassroots health care system as a step in the search for a new national identity and is trying to revitalize traditional medicine. During the Cultural Revolution, the Ministry of Health directed health care throughout China and established primary care units. Chinese physicians who are trained in Western medicine also learn traditional medicine, while traditional healers receive training in modern methods, dynamically integrate modern medical concepts and methods, and revitalize some of the appropriate aspects of traditional medicine. Therefore, traditional Chinese medicine was re-created in response to Western medicine during the Cultural Revolution.
During the Cultural Revolution in 1968, the Communist Party of China supported a new system of health care delivery for rural areas. Each village is assigned a barefoot doctor (a medical staff with basic medical skills and knowledge to deal with minor illnesses) and is responsible for providing basic medical care. The medical staff combined the values of traditional China with modern methods to provide health and medical care to poor farmers in remote rural areas. The barefoot doctors became a symbol of the Cultural Revolution, for the introduction of modern medicine into villages where traditional Chinese medicine services were used.
These include Zhang Zhongjing, Hua Tuo, Sun Simiao, Tao Hongjing, Zhang Jiegu, and Li Shizhen.
At the beginning of Hong Kong's opening up, Western medicine was not yet popular, and Western medicine doctors were mostly foreigners; local residents mostly relied on Chinese medicine practitioners. In 1841, the British government of Hong Kong issued an announcement pledging to govern Hong Kong residents in accordance with all the original rituals, customs and private legal property rights. As traditional Chinese medicine had always been used in China, the use of traditional Chinese medicine was not regulated.
The establishment in 1870 of the Tung Wah Hospital was the first use of Chinese medicine for the treatment in Chinese hospitals providing free medical services. As the promotion of Western medicine by the British government started from 1940, Western medicine started being popular among Hong Kong population. In 1959, Hong Kong had researched the use of traditional Chinese medicine to replace Western medicine.
Traditional Chinese medicine (TCM) is a broad range of medicine practices sharing common concepts which have been developed in China and are based on a tradition of more than 2,000 years, including various forms of herbal medicine, acupuncture, massage (), exercise (), and dietary therapy. It is primarily used as a complementary alternative medicine approach. TCM is widely used in China and it is also used in the West. Its philosophy is based on Yinyangism (i.e., the combination of Five Phases theory with Yin–Yang theory), which was later absorbed by Daoism. Philosophical texts influenced TCM, mostly by being grounded in the same theories of "qi, yin-yang" and "wuxing" and microcosm-macrocosm analogies.
Yin and yang are ancient Chinese concepts which can be traced back to the Shang dynasty (1600–1100 BCE). They represent two abstract and complementary aspects that every phenomenon in the universe can be divided into. Primordial analogies for these aspects are the sun-facing (yang) and the shady (yin) side of a hill. Two other commonly used representational allegories of yin and yang are water and fire. In the yin–yang theory, detailed attributions are made regarding the yin or yang character of things:
The concept of yin and yang is also applicable to the human body; for example, the upper part of the body and the back are assigned to yang, while the lower part of the body are believed to have the yin character. Yin and yang characterization also extends to the various body functions, and – more importantly – to disease symptoms (e.g., cold and heat sensations are assumed to be yin and yang symptoms, respectively). Thus, yin and yang of the body are seen as phenomena whose lack (or over-abundance) comes with characteristic symptom combinations:
TCM also identifies drugs believed to treat these specific symptom combinations, i.e., to reinforce yin and yang.
Strict rules are identified to apply to the relationships between the Five Phases in terms of sequence, of acting on each other, of counteraction, etc. All these aspects of Five Phases theory constitute the basis of the zàng-fǔ concept, and thus have great influence regarding the TCM model of the body. Five Phase theory is also applied in diagnosis and therapy.
Correspondences between the body and the universe have historically not only been seen in terms of the Five Elements, but also of the "Great Numbers" () For example, the number of acu-points has at times been seen to be 365, corresponding with the number of days in a year; and the number of main meridians–12–has been seen as corresponding with the number of rivers flowing through the ancient Chinese empire.
TCM "holds that the body's vital energy ("chi" or "qi") circulates through channels, called "meridians", that have branches connected to bodily organs and functions." Its view of the human body is only marginally concerned with anatomical structures, but focuses primarily on the body's "functions" (such as digestion, breathing, temperature maintenance, etc.):
These functions are aggregated and then associated with a primary functional entity – for instance, nourishment of the tissues and maintenance of their moisture are seen as connected functions, and the entity postulated to be responsible for these functions is xuě (blood). These functional entities thus constitute "concepts" rather than something with biochemical or anatomical properties.
The primary functional entities used by traditional Chinese medicine are qì, xuě, the five zàng organs, the six fǔ organs, and the meridians which extend through the organ systems. These are all theoretically interconnected: each zàng organ is paired with a fǔ organ, which are nourished by the blood and concentrate qi for a particular function, with meridians being extensions of those functional systems throughout the body.
Concepts of the body and of disease used in TCM are pseudoscientific, similar to European humoral theory. TCM is characterized as full of pseudoscience. Some practitioners no longer consider yin and yang and the idea of an energy flow to apply. Scientific investigation has not found any histological or physiological evidence for traditional Chinese concepts such as "qi", meridians, and acupuncture points. It is a generally held belief within the acupuncture community that acupuncture points and meridians structures are special conduits for electrical signals but no research has established any consistent anatomical structure or function for either acupuncture points or meridians. The scientific evidence for the anatomical existence of either meridians or acupuncture points is not compelling. Stephen Barrett of Quackwatch writes that, "TCM theory and practice are not based upon the body of knowledge related to health, disease, and health care that has been widely accepted by the scientific community. TCM practitioners disagree among themselves about how to diagnose patients and which treatments should go with which diagnoses. Even if they could agree, the TCM theories are so nebulous that no amount of scientific study will enable TCM to offer rational care."
TCM has been the subject of controversy within China. In 2006, the Chinese scholar Zhang Gongyao triggered a national debate when he published an article entitled "Farewell to Traditional Chinese Medicine", arguing that TCM was a pseudoscience that should be abolished in public healthcare and academia. The Chinese government however, interested in the opportunity of export revenues, took the stance that TCM is a science and continued to encourage its development.
TCM distinguishes many kinds of qi (). In a general sense, qi is something that is defined by five "cardinal functions":
Vacuity of qi will be characterized especially by pale complexion, lassitude of spirit, lack of strength, spontaneous sweating, laziness to speak, non-digestion of food, shortness of breath (especially on exertion), and a pale and enlarged tongue.
Qi is believed to be partially generated from food and drink, and partially from air (by breathing). Another considerable part of it is inherited from the parents and will be consumed in the course of life.
TCM uses special terms for qi running inside of the blood vessels and for qi that is distributed in the skin, muscles, and tissues between them. The former is called yíng-qì (); its function is to complement xuè and its nature has a strong yin aspect (although qi in general is considered to be yang). The latter is called weì-qì (); its main function is defence and it has pronounced yang nature.
Qi is said to circulate in the meridians. Just as the qi held by each of the zang-fu organs, this is considered to be part of the 'principal' qi () of the body (also called , "true" qi, or , "original" qi).
In contrast to the majority of other functional entities, xuè (, "blood") is correlated with a physical form – the red liquid running in the blood vessels. Its concept is, nevertheless, defined by its functions: nourishing all parts and tissues of the body, safeguarding an adequate degree of moisture, and sustaining and soothing both consciousness and sleep.
Typical symptoms of a lack of xuě (usually termed "blood vacuity" []) are described as: Pale-white or withered-yellow complexion, dizziness, flowery vision, palpitations, insomnia, numbness of the extremities; pale tongue; "fine" pulse.
Closely related to xuě are the jīnyè (, usually translated as "body fluids"), and just like xuě they are considered to be yin in nature, and defined first and foremost by the functions of nurturing and moisturizing the different structures of the body. Their other functions are to harmonize yin and yang, and to help with the secretion of waste products.
Jīnyè are ultimately extracted from food and drink, and constitute the raw material for the production of xuě; conversely, xuě can also be transformed into jīnyè. Their palpable manifestations are all bodily fluids: tears, sputum, saliva, gastric acid, joint fluid, sweat, urine, etc.
The zàng-fǔ () constitute the centre piece of TCM's systematization of bodily functions. Bearing the names of organs, they are, however, only secondarily tied to (rudimentary) anatomical assumptions (the fǔ a little more, the zàng much less). As they are primarily defined by their functions, they are not equivalent to the anatomical organs–to highlight this fact, their names are usually capitalized.
The term zàng () refers to the five entities considered to be yin in nature–Heart, Liver, Spleen, Lung, Kidney–, while fǔ () refers to the six yang organs–Small Intestine, Large Intestine, Gallbladder, Urinary Bladder, Stomach and Sānjiaō.
The zàng's essential functions consist in production and storage of qì and xuě; they are said to regulate digestion, breathing, water metabolism, the musculoskeletal system, the skin, the sense organs, aging, emotional processes, and mental activity, among other structures and processes. The fǔ organs' main purpose is merely to transmit and digest () substances such as waste and food.
Since their concept was developed on the basis of Wǔ Xíng philosophy, each zàng is paired with a fǔ, and each zàng-fǔ pair is assigned to one of five elemental qualities (i.e., the Five Elements or Five Phases). These correspondences are stipulated as:
The zàng-fǔ are also connected to the twelve standard meridians–each yang meridian is attached to a fǔ organ, and five of the yin meridians are attached to a zàng. As there are only five zàng but six yin meridians, the sixth is assigned to the Pericardium, a peculiar entity almost similar to the Heart zàng.
The meridians (, ) are believed to be channels running from the zàng-fǔ in the interior (, ) of the body to the limbs and joints ("the surface" [, ]), transporting qi and xuĕ. TCM identifies 12 "regular" and 8 "extraordinary" meridians; the Chinese terms being (, lit. "the Twelve Vessels") and () respectively. There's also a number of less customary channels branching from the "regular" meridians.
In traditional clinical encounters, women and men were treated differently. Chinese doctor Cheng Maoxian, born in 1581, lived and practiced medicine in Yangzhou in the 1610s and 1620s. He took diligent care to write case studies of his interactions with his patients and their ailments as well as his prescribed medicines. Doctors such as Maoxian treated men as well as women, however Cheng described each case study with reference to contemporary social structures.
In encounters between sick women and their male doctors, women were often shy about their issues and frequently made the doctor's work more difficult by concealing the extent all symptoms. Even if they did confess all their symptoms, doctors, such as Cheng, would consider the majority of the illness to be linked to some problem concerning the woman's reproductive system or cycle. One of the stories Cheng discussed in his case studies was that of Fan Jisuo's teenage daughter, who could not be diagnosed because she was unwilling to speak on her symptoms. In this particular case, the illness involved discharge from her intimate areas. Cheng insisted on asking more about her illness and the patient's modesty became a frustrating obstacle.
When a woman fell ill, an appropriate adult man was to call the doctor and remain present during the examination, for the woman could not be left alone with the doctor. However, this was not always the case. In particular cases, when a woman dealt with complications of pregnancy or birth, older women became involved and assumed the role of the formal authority. Men in these situations would not have much power to interfere.
However, when a doctor's visit was absolutely necessary, a breaking of norms was sometimes necessary. To properly examine the patient, doctors were faced with the task of going beyond the norm of female modesty. As Cheng describes, there were four standard methods of diagnosis—looking, asking, listening and smelling and touching (for pulse-taking). To maintain some form of modesty, women would often stay hidden behind curtains and screens. The doctor was allowed to touch enough of her body to complete his examination, often just the pulse taking. This would lead to situations where the symptoms and the doctor's diagnosis did not agree and the doctor would have to ask to see more of the patient.
Gender was presumed to influence the movement of energy and a well-trained physician would be expected to read the pulse and should have been able to identify two dozen or more energy flows. Yin and yang concepts were applied to the feminine and masculine aspects of all bodies at large, implying that in nature the differences between men and women begin at the level of this energy flow. According to "Bequeathed Writings of Master Chu" the male's yang pulse movement follows an ascending path in "compliance [with cosmic direction] so that the cycle of circulation in the body and the Vital Gate are felt...The female's yin pulse movement follows a defending path against the direction of cosmic influences, so that the nadir and the Gate of Life are felt at the inch position of the left hand". In sum, classical medicine marked yin and yang as high and low on bodies which in turn would be labeled normal or abnormal and gendered either male or female.
The act of diagnosing women was not as simple as the diagnosing of men in traditional Chinese medicine. This was for several reasons: first, the treatment of sick women was expected to be called in by and take place under male authority. The visiting physician would then discuss the female's problems and diagnosis through the male. Second, women were often silent about their issues with doctors and male figures due to the societal expectation of female modesty and the presence of a male figure in the room. Third, the presence of male authority in the sick room and the patriarchal dominated society also caused doctors to reference their women and children patients "the anonymous category of family members ("Jia Ren") or household ("Ju Jia")" in their journals. This anonymity and lack of conversation between the doctor and woman patient led to the inquiry diagnosis of the Four Diagnostic Methods being the most challenging.
The study of medicine for women was called "Fuke" (known as gynecology and obstetrics in modern science and medicine); however, it has little to no ancient works based on it except for Fu Qing-zhu's "Fu Qing Zhu Nu Ke" (Fu Qing-zhu's "Gynecology"). The most challenging part of a woman's health in Traditional Chinese Medicine was pregnancy and postpartum, this is because there were many definitions of pregnancy in traditional Chinese medicine.
The recognition of pregnancy in the Western medical world has been around since the publication of the Hippocratic Corpus, circa mid-fifth century to the mid-fourth century BCE, in the gynecological treatises On the Nature of the Woman, On the Diseases of Women, Generation, On the Nature of the Child, On Sterile Women, On Fistulae, and On Hemorrhoids. The term Caesarean section derives from an ancient Roman, or Caesarean (from Caesar) law that demanded that when a pregnant woman died, her body could not be buried until the unborn child had been removed. Ancient Roman doctors were forbidden from performing this procedure on living women, however.
Traditional Chinese medicine's attempts to grapple with pregnancy are documented from at least the seventeenth century. According to Charlotte Furth, "a pregnancy (in the seventeenth century) as a known bodily experience emerged [...] out of the liminality of menstrual irregularity, as uneasy digestion, and a sense of fullness". These symptoms were common among other illness as well, so the diagnosis of pregnancy often came later in the term. The "Canon of the Pulse" or the use of pulse in diagnosis stated that pregnancy was "a condition marked by symptoms of disorder in one whose pulse is normal" or "where the pulse and symptoms do not agree". Just as in the normal diagnosis process, women were often silent about suspected pregnancy, this led to many men in the households not knowing their wife or daughter was pregnant until complications arrived.
Complications through the misdiagnosis and silence of pregnancies often involved medically induced abortions, according to Furth's book, Dr.Cheng (her case study) "was unapologetic about endangering a fetus when pregnancy risked a mother's well being". The method of abortion was used through the ingestion of certain herbs and foods. The practice of abortion was contrasted with the families disappointment and disapproval of the loss of the fetus and often led to familial complications down the line.
If the baby and mother survived the term of the pregnancy, childbirth was then the next step. The tools provided for birth were: towels to catch the blood, a container for the placenta, a pregnancy sash to support the belly, and an infant swaddling wrap. With these tools, the baby was born, cleaned, and swaddled; however, the mother was then immediately the focus of the doctor to replenish her "qi." In his writings, Dr.Cheng places a large amount of emphasis on the Four Diagnostic methods to deal with postpartum issues and instructs all physicians to "not neglect any [of the four methods]". The process of birthing was thought to deplete a woman's blood level and "qi" so the most common treatments for postpartum were food (commonly garlic and ginseng), medicine, and rest. This process was followed up by a month check-in with the physician, a practice known as "zuo yuezi".
In TCM, as in many other cultures, the health and medicine of female bodies was less understood than that of male bodies. Women's bodies were often secondary to male bodies, since women were thought of as the weaker, sicklier sex. Yin and Yang were critical to the understanding of women's bodies, yet they were only understood in conjunction with male bodies. Women's diseases were harder to treat and cure in the eyes of many cultures, but especially in TCM. Finally, social and cultural beliefs were often barriers to understanding the female body.
In order to understand the way that TCM looked at female bodies, it is critical to understand the relationship of "yin" and "yang" to the practice of medicine. According to Charlotte Furth and her book, "A Flourishing Yin: Gender in China’s Medical History", yin and yang showed the embodiment of nature within human bodies, and with it the natural phenomena that we understand as gendered differences. Yin and yang ruled the body, the body being a microcosm of the universe and the earth. In addition, gender in the body was understood as homologous, the two genders operating in synchronization.
Operating within yin and yang, bodily functions could be categorized through systems. In many drawings and diagrams, the twelve channels and their visceral systems were organized by yin and yang, and that organization was identical in female and male bodies. This points to the understanding in traditional Chinese medicine that female and male bodies were no different on the plane of yin and yang. Their gendered differences were not acknowledged in diagrams of the human body. Medical journals were filled with illustrations of male bodies or androgynous bodies, those that did not display gendered characteristics.
For example, understanding the womb and its fundamental difference from male bodies was irrelevant. Traditional Chinese medicine did not recognize the womb as the place of reproduction. For the Chinese, the abdominal cavity presented pathologies that were similar in both men and women, which included tumors, growths, hernias, and swellings of the genitals. The "master system" as Charlotte Furth identifies, is the kidney visceral system, which governed reproductive functions in TCM. Therefore, it was not the anatomical structures in women that allowed for pregnancy, but the difference in female processes that allowed for the affliction of pregnancy to occur.
Social and cultural beliefs were often barriers to learning more about female health, with women themselves often being the most formidable barrier. Women were often uncomfortable talking about their illnesses, especially in front of the male chaperones that attended medical examinations. Women would choose to omit certain symptoms as a means of upholding their chastity and honor. One such examples is the case in which a teenage girl was unable to be diagnosed because she failed to mention her symptom of vaginal discharge. Silence was their way of maintaining control in these situations, but it often came at the expense of their health and the advancement of female health and medicine. This silence and control was most obviously seen when the health problem was related to the core of Ming "fuke", or the sexual body. It was often in these diagnostic settings that women would choose silence. In addition, there would be conflict between patient and doctor on the probability of her diagnosis. For example, a woman who thought herself to be past the point of child-bearing age, might not believe a doctor who diagnoses her as pregnant. This only resulted in more conflict.
In conclusion, female health and medicine was a topic that was not well understood by TCM. Although "yin" and "yang" were central to understanding female bodies, it was often only through the lens of male bodies. Women were seen as variations of male bodies, with similar structures but different processes. This was also compounded by the autonomy of women who chose to stay silent during diagnostics, and thus complicate the issue of female health and medicine.
Infertility was also not very well understood in TCM, however it posed serious social and cultural repercussions.
One of the most cited scholar that mentions female health is 7th-century scholar Sun Simiao. He is often quoted to the likes of "those who have prescriptions for women's distinctiveness take their differences of pregnancy, childbirth and [internal] bursting injuries as their basis." Even in contemporary "fu ke" does Sun's observations about female reproductive functions, it stills remains an important function of women's health. The propensity to place more emphasis on reproductive functions, rather than the entire health of the woman, seems to suggest that the main function of "fu ke" is to produce children.
Once again, the kidney visceral system governs the "source "Qi,"" which governs the reproductive systems in both sexes. This source "Qi" was thought to "be slowly depleted through sexual activity, menstruation and childbirth." It was also understood that the depletion of source Qi could result from the movement of an external pathology that moved through the outer visceral systems before causing more permanent damage to the home of source Qi, the kidney system. In addition, the view that only very serious ailments ended in the damage of this system means that those who had trouble with their reproductive systems or fertility were seriously ill.
According to traditional Chinese medical texts, infertility can be summarized into different syndrome types. These were spleen and kidney depletion (yang depletion), liver and kidney depletion (yin depletion), blood depletion, phlegm damp, liver oppression, and damp heat. This is important because, while most other issues were complex in Chinese medical physiology, women's fertility issues were simple. Most syndrome types revolved around menstruation, or lack thereof. The patient was entrusted with recording not only the frequency, but also the "volume, color, consistency, and odor of menstrual flow." This placed responsibility of symptom recording on the patient, and was compounded by the earlier discussed issue of female chastity and honor. This meant that diagnosing female infertility was difficult, because the only symptoms that were recorded and monitored by the physician was the pulse and color of the tongue.
In conclusion, the issue of infertility shows how social and cultural barriers affected the practice of traditional Chinese medicine.
In general, disease is perceived as a disharmony (or imbalance) in the functions or interactions of yin, yang, qi, xuĕ, zàng-fǔ, meridians etc. and/or of the interaction between the human body and the environment. Therapy is based on which "pattern of disharmony" can be identified. Thus, "pattern discrimination" is the most important step in TCM diagnosis. It is also known to be the most difficult aspect of practicing TCM.
To determine which pattern is at hand, practitioners will examine things like the color and shape of the tongue, the relative strength of pulse-points, the smell of the breath, the quality of breathing or the sound of the voice. For example, depending on tongue and pulse conditions, a TCM practitioner might diagnose bleeding from the mouth and nose as: "Liver fire rushes upwards and scorches the Lung, injuring the blood vessels and giving rise to reckless pouring of blood from the mouth and nose." He might then go on to prescribe treatments designed to clear heat or supplement the Lung.
In TCM, a disease has two aspects: "bìng" and "zhèng". The former is often translated as "disease entity", "disease category", "illness", or simply "diagnosis". The latter, and more important one, is usually translated as "pattern" (or sometimes also as "syndrome"). For example, the disease entity of a common cold might present with a pattern of wind-cold in one person, and with the pattern of wind-heat in another.
From a scientific point of view, most of the disease entities () listed by TCM constitute symptoms. Examples include headache, cough, abdominal pain, constipation etc.
Since therapy will not be chosen according to the disease entity but according to the pattern, two people with the same disease entity but different patterns will receive different therapy. Vice versa, people with similar patterns might receive similar therapy even if their disease entities are different. This is called "yì bìng tóng zhì, tóng bìng yì zhì" ().
In TCM, "pattern" () refers to a "pattern of disharmony" or "functional disturbance" within the functional entities the TCM model of the body is composed of. There are disharmony patterns of qi, xuě, the body fluids, the zàng-fǔ, and the meridians. They are ultimately defined by their symptoms and signs (i.e., for example, pulse and tongue findings).
In clinical practice, the identified pattern usually involves a combination of affected entities (compare with typical examples of patterns). The concrete pattern identified should account for "all" the symptoms a person has.
The Six Excesses (, sometimes also translated as "Pathogenic Factors", or "Six Pernicious Influences"; with the alternative term of , – "Six Evils" or "Six Devils") are allegorical terms used to describe disharmony patterns displaying certain typical symptoms. These symptoms resemble the effects of six climatic factors. In the allegory, these symptoms can occur because one or more of those climatic factors (called , "the six qi") were able to invade the body surface and to proceed to the interior. This is sometimes used to draw causal relationships (i.e., prior exposure to wind/cold/etc. is identified as the cause of a disease), while other authors explicitly deny a direct cause-effect relationship between weather conditions and disease, pointing out that the Six Excesses are primarily descriptions of a certain combination of symptoms translated into a pattern of disharmony. It is undisputed, though, that the Six Excesses can manifest inside the body without an external cause. In this case, they might be denoted "internal", e.g., "internal wind" or "internal fire (or heat)".
The Six Excesses and their characteristic clinical signs are:
Six-Excesses-patterns can consist of only one or a combination of Excesses (e.g., wind-cold, wind-damp-heat). They can also transform from one into another.
For each of the functional entities (qi, xuĕ, zàng-fǔ, meridians etc.), typical disharmony patterns are recognized; for example: qi vacuity and qi stagnation in the case of qi; blood vacuity, blood stasis, and blood heat in the case of xuĕ; Spleen qi vacuity, Spleen yang vacuity, Spleen qi vacuity with down-bearing qi, Spleen qi vacuity with lack of blood containment, cold-damp invasion of the Spleen, damp-heat invasion of Spleen and Stomach in case of the Spleen zàng; wind/cold/damp invasion in the case of the meridians.
TCM gives detailed prescriptions of these patterns regarding their typical symptoms, mostly including characteristic tongue and/or pulse findings. For example:
The process of determining which actual pattern is on hand is called (, usually translated as "pattern diagnosis", "pattern identification" or "pattern discrimination"). Generally, the first and most important step in pattern diagnosis is an evaluation of the present signs and symptoms on the basis of the "Eight Principles" (). These eight principles refer to four pairs of fundamental qualities of a disease: exterior/interior, heat/cold, vacuity/repletion, and yin/yang. Out of these, heat/cold and vacuity/repletion have the biggest clinical importance. The yin/yang quality, on the other side, has the smallest importance and is somewhat seen aside from the other three pairs, since it merely presents a general and vague conclusion regarding what other qualities are found. In detail, the Eight Principles refer to the following:
After the fundamental nature of a disease in terms of the Eight Principles is determined, the investigation focuses on more specific aspects. By evaluating the present signs and symptoms against the background of typical disharmony patterns of the various entities, evidence is collected whether or how specific entities are affected. This evaluation can be done
There are also three special pattern diagnosis systems used in case of febrile and infectious diseases only ("Six Channel system" or "six division pattern" []; "Wei Qi Ying Xue system" or "four division pattern" []; "San Jiao system" or "three burners pattern" []).
Although TCM and its concept of disease do not strongly differentiate between cause and effect, pattern discrimination can include considerations regarding the disease cause; this is called (, "disease-cause pattern discrimination").
There are three fundamental categories of disease causes () recognized:
In TCM, there are five major diagnostic methods: inspection, auscultation, olfaction, inquiry, and palpation. These are grouped into what is known as the "Four pillars" of diagnosis, which are Inspection, Auscultation/ Olfaction, Inquiry, and Palpation ().
Examination of the tongue and the pulse are among the principal diagnostic methods in TCM. Details of the tongue, including shape, size, color, texture, cracks, teethmarks, as well as tongue coating are all considered as part of tongue diagnosis. Various regions of the tongue's surface are believed to correspond to the zàng-fŭ organs. For example, redness on the tip of the tongue might indicate heat in the Heart, while redness on the sides of the tongue might indicate a heat in the Liver.
Pulse palpation involves measuring the pulse both at a superficial and at a deep level at three different locations on the radial artery ("Cun, Guan, Chi", located two fingerbreadths from the wrist crease, one fingerbreadth from the wrist crease, and right at the wrist crease, respectively, usually palpated with the index, middle and ring finger) of each arm, for a total of twelve pulses, all of which are thought to correspond with certain zàng-fŭ. The pulse is examined for several characteristics including rhythm, strength and volume, and described with qualities like "floating, slippery, bolstering-like, feeble, thready and quick"; each of these qualities indicate certain disease patterns. Learning TCM pulse diagnosis can take several years.
The term "herbal medicine" is somewhat misleading in that, while plant elements are by far the most commonly used substances in TCM, other, non-botanic substances are used as well: animal, human, and mineral products are also used. Thus, the term "medicinal" (instead of herb) may be used, although there is no scientific evidence that any of these compounds have medicinal effects.
There are roughly 13,000 compounds used in China and over 100,000 TCM recipes recorded in the ancient literature. Plant elements and extracts are by far the most common elements used. In the classic "Handbook of Traditional Drugs" from 1941, 517 drugs were listed – out of these, 45 were animal parts, and 30 were minerals.
Some animal parts used can be considered rather strange such as cow gallstones, hornet nests, leeches, and scorpion. Other examples of animal parts include horn of the antelope or buffalo, deer antlers, testicles and penis bone of the dog, and snake bile. Some TCM textbooks still recommend preparations containing animal tissues, but there has been little research to justify the claimed clinical efficacy of many TCM animal products.
Some compounds can include the parts of endangered species, including tiger bones and rhinoceros horn
which is used for many ailments (though not as an aphrodisiac as is commonly misunderstood in the West).
The black market in rhinoceros horn (driven not just by TCM but also unrelated status-seeking) has reduced the world's rhino population by more than 90 percent over the past 40 years.
Concerns have also arisen over the use of pangolin scales, turtle plastron, seahorses, and the gill plates of mobula and manta rays. Illegal pangolin sales at animal markets are suggested as a possible method of the transmission of SARS-CoV-2 to humans.
Poachers hunt restricted or endangered species to supply the black market with TCM products. There is no scientific evidence of efficacy for tiger medicines. Concern over China considering to legalize the trade in tiger parts prompted the 171-nation Convention on International Trade in Endangered Species (CITES) to endorse a decision opposing the resurgence of trade in tigers. Fewer than 30,000 saiga antelopes remain, which are exported to China for use in traditional fever therapies. Organized gangs illegally export the horn of the antelopes to China. The pressures on seahorses ("Hippocampus" spp.) used in traditional medicine is enormous; tens of millions of animals are unsustainably caught annually. Many species of syngnathid are currently part of the IUCN Red List of Threatened Species or national equivalents.
Since TCM recognizes bear bile as a treatment compound, more than 12,000 asiatic black bears are held in bear farms. The bile is extracted through a permanent hole in the abdomen leading to the gall bladder, which can cause severe pain. This can lead to bears trying to kill themselves. As of 2012, approximately 10,000 bears are farmed in China for their bile. This practice has spurred public outcry across the country. The bile is collected from live bears via a surgical procedure. As of March 2020 bear bile as ingredient of "Tan Re Qing" injection remains on the list of remedies recommended for treatment of "severe cases" of COVID-19 by National Health Commission of China and the National Administration of Traditional Chinese Medicine.
The deer penis is believed to have therapeutic benefits according to traditional Chinese medicine. Tiger parts from poached animals include tiger penis, believed to improve virility, and tiger eyes. The illegal trade for tiger parts in China has driven the species to near-extinction because of its popularity in traditional medicine. Laws protecting even critically endangered species such as the Sumatran tiger fail to stop the display and sale of these items in open markets. Shark fin soup is traditionally regarded in Chinese medicine as beneficial for health in East Asia, and its status as an elite dish has led to huge demand with the increase of affluence in China, devastating shark populations. The shark fins have been a part of traditional Chinese medicine for centuries. Shark finning is banned in many countries, but the trade is thriving in Hong Kong and China, where the fins are part of shark fin soup, a dish considered a delicacy, and used in some types of traditional Chinese medicine.
The tortoise (freshwater turtle, "guiban") and turtle (Chinese softshell turtle, "biejia") species used in traditional Chinese medicine are raised on farms, while restrictions are made on the accumulation and export of other endangered species. However, issues concerning the overexploitation of Asian turtles in China have not been completely solved. Australian scientists have developed methods to identify medicines containing DNA traces of endangered species. Finally, although not an endangered species, sharp rises in exports of donkeys and donkey hide from Africa to China to make the traditional remedy "ejiao" have prompted export restrictions by some African countries.
Traditional Chinese Medicine also includes some human parts: the classic Materia medica (Bencao Gangmu) describes (also criticizes) the use of 35 human body parts and excreta in medicines, including bones, fingernail, hairs, dandruff, earwax, impurities on the teeth, feces, urine, sweat, organs, but most are no longer in use.
Human placenta has been used an ingredient in certain traditional Chinese medicines, including using dried human placenta, known as "Ziheche", to treat infertility, impotence and other conditions. The consumption of the human placenta is a potential source of infection.
The traditional categorizations and classifications that can still be found today are:
there were not enough good-quality trials of herbal therapies to allow their effectiveness to be determined. A high percentage of relevant studies on traditional Chinese medicine are in Chinese databases. Fifty percent of systematic reviews on TCM did not search Chinese databases, which could lead to a bias in the results. Many systematic reviews of TCM interventions published in Chinese journals are incomplete, some contained errors or were misleading. The herbs recommended by traditional Chinese practitioners in the US are unregulated.
With an eye to the enormous Chinese market, pharmaceutical companies have explored the potential for creating new drugs from traditional remedies. A "Nature" editorial described TCM as "fraught with pseudoscience", and stated that having "no rational mechanism of action for most of its therapies" is the "most obvious answer" to why its study didn't provide a "flood of cures", while advocates responded that "researchers are missing aspects of the art, notably the interactions between different ingredients in traditional therapies."
One of the successes was the development in the 1970s of the antimalarial drug artemisinin, which is a processed extract of "Artemisia annua", a herb traditionally used as a fever treatment. Artemisia annua has been used by Chinese herbalists in traditional Chinese medicines for 2,000 years. In 1596, Li Shizhen recommended tea made from qinghao specifically to treat malaria symptoms in his "Compendium of Materia Medica". Researcher Tu Youyou discovered that a low-temperature extraction process could isolate an effective antimalarial substance from the plant. Tu says she was influenced by a traditional Chinese herbal medicine source, "The Handbook of Prescriptions for Emergency Treatments", written in 340 by Ge Hong, which states that this herb should be steeped in cold water. The extracted substance, once subject to detoxification and purification processes, is a usable antimalarial drug – a 2012 review found that artemisinin-based remedies were the most effective drugs for the treatment of malaria. For her work on malaria, Tu received the 2015 Nobel Prize in Physiology or Medicine. Despite global efforts in combating malaria, it remains a large burden for the population. Although WHO recommends artemisinin-based remedies for treating uncomplicated malaria, artemisinin resistance can no longer be ignored.
Also in the 1970s Chinese researcher Zhang TingDong and colleagues investigated the potential use of the traditionally used substance arsenic trioxide to treat acute promyelocytic leukemia (APL). Building on his work, research both in China and the West eventually led to the development of the drug Trisenox, which was approved for leukemia treatment by the FDA in 2000.
Huperzine A, an extract from the herb, "Huperzia serrata", is under preliminary research as a possible therapeutic for Alzheimer's disease, but poor methodological quality of the research restricts conclusions about its effectiveness.
Ephedrine in its natural form, known as "má huáng" () in traditional Chinese medicine, has been documented in China since the Han dynasty (206 BCE – 220 CE) as an antiasthmatic and stimulant. In 1885, the chemical synthesis of ephedrine was first accomplished by Japanese organic chemist Nagai Nagayoshi based on his research on Japanese and Chinese traditional herbal medicines
Pien tze huang was first documented in the Ming dynasty.
A 2012 systematic review found there is a lack of available cost-effectiveness evidence in TCM.
From the earliest records regarding the use of compounds to today, the toxicity of certain substances has been described in all Chinese materiae medicae. Since TCM has become more popular in the Western world, there are increasing concerns about the potential toxicity of many traditional Chinese plants, animal parts and minerals. Traditional Chinese herbal remedies are conveniently available from grocery stores in most Chinese neighborhoods; some of these items may contain toxic ingredients, are imported into the U.S. illegally, and are associated with claims of therapeutic benefit without evidence. For most compounds, efficacy and toxicity testing are based on traditional knowledge rather than laboratory analysis. The toxicity in some cases could be confirmed by modern research (i.e., in scorpion); in some cases it could not (i.e., in "Curculigo"). Traditional herbal medicines can contain extremely toxic chemicals and heavy metals, and naturally occurring toxins, which can cause illness, exacerbate pre-existing poor health or result in death. Botanical misidentification of plants can cause toxic reactions in humans. The description on some plants used in traditional Chinese medicine have changed, leading to unintended intoxication of the wrong plants. A concern is also contaminated herbal medicines with microorganisms and fungal toxins, including aflatoxin. Traditional herbal medicines are sometimes contaminated with toxic heavy metals, including lead, arsenic, mercury and cadmium, which inflict serious health risks to consumers. Also, adulteration of some herbal medicine preparations with conventional drugs which may cause serious adverse effects, such as corticosteroids, phenylbutazone, phenytoin, and glibenclamide, has been reported.
Substances known to be potentially dangerous include "Aconitum", secretions from the Asiatic toad, powdered centipede, the Chinese beetle ("Mylabris phalerata"), certain fungi, "Aristolochia", Arsenic sulfide (Realgar), mercury sulfide, and cinnabar. Asbestos ore (Actinolite, Yang Qi Shi, 阳起石) is used to treat impotence in TCM. Due to galena's (litharge, lead(II) oxide) high lead content, it is known to be toxic. Lead, mercury, arsenic, copper, cadmium, and thallium have been detected in TCM products sold in the U.S. and China.
To avoid its toxic adverse effects "Xanthium sibiricum" must be processed. Hepatotoxicity has been reported with products containing "Reynoutria multiflora" (synonym "Polygonum multiflorum"), glycyrrhizin, "Senecio" and "Symphytum". The herbs indicated as being hepatotoxic included "Dictamnus dasycarpus", "Astragalus membranaceous", and "Paeonia lactiflora". Contrary to popular belief, "Ganoderma lucidum" mushroom extract, as an adjuvant for cancer immunotherapy, appears to have the potential for toxicity. A 2013 review suggested that although the antimalarial herb "Artemisia annua" may not cause hepatotoxicity, haematotoxicity, or hyperlipidemia, it should be used cautiously during pregnancy due to a potential risk of embryotoxicity at a high dose.
However, many adverse reactions are due to misuse or abuse of Chinese medicine. For example, the misuse of the dietary supplement "Ephedra" (containing ephedrine) can lead to adverse events including gastrointestinal problems as well as sudden death from cardiomyopathy. Products adulterated with pharmaceuticals for weight loss or erectile dysfunction are one of the main concerns. Chinese herbal medicine has been a major cause of acute liver failure in China.
Acupuncture is the insertion of needles into superficial structures of the body (skin, subcutaneous tissue, muscles) – usually at acupuncture points (acupoints) – and their subsequent manipulation; this aims at influencing the flow of qi. According to TCM it relieves pain and treats (and prevents) various diseases. The US FDA classifies single-use acupuncture needles as Class II medical devices, under CFR 21.
Acupuncture is often accompanied by moxibustion – the Chinese characters for acupuncture () literally meaning "acupuncture-moxibustion" – which involves burning mugwort on or near the skin at an acupuncture point. According to the American Cancer Society, "available scientific evidence does not support claims that moxibustion is effective in preventing or treating cancer or any other disease".
In electroacupuncture, an electric current is applied to the needles once they are inserted, to further stimulate the respective acupuncture points.
A 2013 editorial by Steven P. Novella and David Colquhoun found that the inconsistency of results of acupuncture studies (i.e. acupuncture relieved pain in some conditions but had no effect in other very similar conditions) suggests false positive results, which may be caused by factors like biased study designs, poor blinding, and the classification of electrified needles (a type of TENS) as a form of acupuncture. The same editorial suggested that given the inability to find consistent results despite more than 3,000 studies of acupuncture, the treatment seems to be a placebo effect and the existing equivocal positive results are noise one expects to see after a large number of studies are performed on an inert therapy. The editorial concluded that the best controlled studies showed a clear pattern, in which the outcome does not rely upon needle location or even needle insertion, and since "these variables are those that define acupuncture, the only sensible conclusion is that acupuncture does not work." According to the US NIH National Cancer Institute, a review of 17,922 patients reported that real acupuncture relieved muscle and joint pain, caused by aromatase inhibitors, much better than sham acupuncture. Regarding cancer patients, The US NIH National Cancer Institute states that acupuncture may cause physical responses in nerve cells, the pituitary gland, and the brain – releasing proteins, hormones, and chemicals that are proposed to affect blood pressure, body temperature, immune activity, and endorphin release.
A 2012 meta-analysis concluded that the mechanisms of acupuncture "are clinically relevant, but that an important part of these total effects is not due to issues considered to be crucial by most acupuncturists, such as the correct location of points and depth of needling ... [but is] ... associated with more potent placebo or context effects". Commenting on this meta-analysis, both Edzard Ernst and David Colquhoun said the results were of negligible clinical significance.
A 2011 overview of Cochrane reviews found evidence that suggests acupuncture is effective for some but not all kinds of pain. A 2010 systematic review found that there is evidence "that acupuncture provides a short-term clinically relevant effect when compared with a waiting list control or when acupuncture is added to another intervention" in the treatment of chronic low back pain. Two review articles discussing the effectiveness of acupuncture, from 2008 and 2009, have concluded that there is not enough evidence to conclude that it is effective beyond the placebo effect.
Acupuncture is generally safe when administered using Clean Needle Technique (CNT). Although serious adverse effects are rare, acupuncture is not without risk. Severe adverse effects, including death, have been reported.
Tui na () is a form of massage akin to acupressure (from which shiatsu evolved). Asian massage is typically administered with the person fully clothed, without the application of grease or oils. Techniques employed may include thumb presses, rubbing, percussion, and assisted stretching.
Qìgōng () is a TCM system of exercise and meditation that combines regulated breathing, slow movement, and focused awareness, purportedly to cultivate and balance qi. One branch of qigong is qigong massage, in which the practitioner combines massage techniques with awareness of the acupuncture channels and points.
Qi is air, breath, energy, or primordial life source that is neither matter or spirit. While Gong is a skillful movement, work, or exercise of the qi.
Cupping () is a type of Chinese massage, consisting of placing several glass "cups" (open spheres) on the body. A match is lit and placed inside the cup and then removed before placing the cup against the skin. As the air in the cup is heated, it expands, and after placing in the skin, cools, creating lower pressure inside the cup that allows the cup to stick to the skin via suction. When combined with massage oil, the cups can be slid around the back, offering "reverse-pressure massage".
Gua sha () is abrading the skin with pieces of smooth jade, bone, animal tusks or horns or smooth stones; until red spots then bruising cover the area to which it is done. It is believed that this treatment is for almost any ailment. The red spots and bruising take three to ten days to heal, there is often some soreness in the area that has been treated.
Diē-dá () or Dit Da, is a traditional Chinese bone-setting technique, usually practiced by martial artists who know aspects of Chinese medicine that apply to the treatment of trauma and injuries such as bone fractures, sprains, and bruises. Some of these specialists may also use or recommend other disciplines of Chinese medical therapies if serious injury is involved. Such practice of bone-setting () is not common in the West.
Traditional Chinese characters and for the words "yin" and "yang" denote different classes of foods, and it is important to consume them in a balanced fashion. The meal sequence should also observe these classes:
In the Orient, it is traditional to eat yang before yin. Miso soup (yang – fermented soybean protein) for breakfast; raw fish (more yang protein); and then the vegetables which are yin.
Many governments have enacted laws to regulate TCM practice.
From 1 July 2012 Chinese medicine practitioners must be registered under the national registration and accreditation scheme with the Chinese Medicine Board of Australia and meet the Board's Registration Standards, to practice in Australia.
TCM is regulated in five provinces in Canada: Alberta, British Columbia, Ontario, Quebec, and Newfoundland.
China has maintained a policy of supporting the adoption of traditional Chinese medicine since its early days under Mao. A National Administration of Traditional Chinese Medicine was created to absorb existing TCM management in 1986 with major changes in 1998.
China's National People's Congress Standing Committee passed the country's first law on TCM in 2016, scheduled to take effect 1 July 2017. The new law standardized TCM certifications by requiring TCM practitioners to (i) pass exams administered by provincial-level TCM authorities, and (ii) obtain recommendations from two certified practitioners. TCM products and services can be advertised only with approval from the local TCM authority.
The Chinese Medicine Council of Hong Kong was established in 1999. It regulates the compounds and professional standards for TCM practitioners. All TCM practitioners in Hong Kong are required to register with the Council. The eligibility for registration includes a recognised 5-year university degree of TCM, a 30-week minimum supervised clinical internship, and passing the licensing exam.
All traditional medicines, including TCM, are regulated by Indonesian Minister of Health Regulation of 2013 on traditional medicine. Traditional medicine license ("Surat Izin Pengobatan Tradisional" – SIPT) is granted to the practitioners whose methods are recognized as safe and may benefit health. The TCM clinics are registered but there is no explicit regulation for it. The only TCM method which is accepted by medical logic and is empirically proofed is acupuncture. The acupuncturists can get SIPT and participate on health care facilities.
The Traditional and Complementary Medicine Bill was passed by parliament in 2012 establishing the Traditional and Complementary Medicine Council to register and regulate traditional and complementary medicine practitioners, including traditional Chinese medicine practitioners as well as other traditional and complementary medicine practitioners such as those in traditional Malay medicine and traditional Indian medicine.
There are no specific regulations in the Netherlands on traditional Chinese medicine, traditional Chinese medicine is neither prohibited nor recognised by the government of the Netherlands. Chinese herbs as well as Chinese herbal products that are used in traditional Chinese medicine are classified as foods and food supplements and these Chinese herbs can be imported into the Netherlands as well as marketed as such without any type registration or notification to the government.
Despite its status, some private health insurance companies reimburse a certain amount of annual costs for acupuncture treatments, this depends on one's insurance policy, as not all insurance policies cover it, and if the acupuncture practitioner is or isn't a member of one the professional organisations that are recognised by private health insurance companies. The recognized professional organizations include the Nederlandse Vereniging voor Acupunctuur (NVA), Nederlandse Artsen Acupunctuur Vereniging (NAAV), ZHONG, (Nederlandse Vereniging voor Traditionele Chinese Geneeskunde), Nederlandse Beroepsvereniging Chinese Geneeswijzen Yi (NBCG Yi), and Wetenschappelijke Artsen Vereniging voor Acupunctuur in Nederland (WAVAN).
The TCM Practitioners Act was passed by Parliament in 2000 and the TCM Practitioners Board was established in 2001 as a statutory board under the Ministry of Health, to register and regulate TCM practitioners. The requirements for registration include possession of a diploma or degree from a TCM educational institution/university on a gazetted list, either structured TCM clinical training at an approved local TCM educational institution or foreign TCM registration together with supervised TCM clinical attachment/practice at an approved local TCM clinic, and upon meeting these requirements, passing the Singapore TCM Physicians Registration Examination (STRE) conducted by the TCM Practitioners Board.
As of July 2012, only six states do not have existing legislation to regulate the professional practice of TCM. These six states are Alabama, Kansas, North Dakota, South Dakota, Oklahoma, and Wyoming. In 1976, California established an Acupuncture Board and became the first state licensing professional acupuncturists. | https://en.wikipedia.org/wiki?curid=5992 |
Chemical bond
A chemical bond is a lasting attraction between atoms, ions or molecules that enables the formation of chemical compounds. The bond may result from the electrostatic force of attraction between oppositely charged ions as in ionic bonds or through the sharing of electrons as in covalent bonds. The strength of chemical bonds varies considerably; there are "strong bonds" or "primary bonds" such as covalent, ionic and metallic bonds, and "weak bonds" or "secondary bonds" such as dipole–dipole interactions, the London dispersion force and hydrogen bonding.
Since opposite charges attract via a simple electromagnetic force, the negatively charged electrons that are orbiting the nucleus and the positively charged protons in the nucleus attract each other. An electron positioned between two nuclei will be attracted to both of them, and the nuclei will be attracted toward electrons in this position. This attraction constitutes the chemical bond. Due to the matter wave nature of electrons and their smaller mass, they must occupy a much larger amount of volume compared with the nuclei, and this volume occupied by the electrons keeps the atomic nuclei in a bond relatively far apart, as compared with the size of the nuclei themselves.
In general, strong chemical bonding is associated with the sharing or transfer of electrons between the participating atoms. The atoms in molecules, crystals, metals and diatomic gases—indeed most of the physical environment around us—are held together by chemical bonds, which dictate the structure and the bulk properties of matter.
All bonds can be explained by quantum theory, but, in practice, simplification rules allow chemists to predict the strength, directionality, and polarity of bonds. The octet rule and VSEPR theory are two examples. More sophisticated theories are valence bond theory, which includes orbital hybridization and resonance, and molecular orbital theory which includes linear combination of atomic orbitals and ligand field theory. Electrostatics are used to describe bond polarities and the effects they have on chemical substances.
A chemical bond is an attraction between atoms. This attraction may be seen as the result of different behaviors of the outermost or valence electrons of atoms. These behaviors merge into each other seamlessly in various circumstances, so that there is no clear line to be drawn between them. However it remains useful and customary to differentiate between different types of bond, which result in different properties of condensed matter.
In the simplest view of a covalent bond, one or more electrons (often a pair of electrons) are drawn into the space between the two atomic nuclei. Energy is released by bond formation. This is not as a result of reduction in potential energy, because the attraction of the two electrons to the two protons is offset by the electron-electron and proton-proton repulsions. Instead, the release of energy (and hence stability of the bond) arises from the reduction in kinetic energy due to the electrons being in a more spatially distributed (i.e. longer de Broglie wavelength) orbital compared with each electron being confined closer to its respective nucleus. These bonds exist between two particular identifiable atoms and have a direction in space, allowing them to be shown as single connecting lines between atoms in drawings, or modeled as sticks between spheres in models.
In a polar covalent bond, one or more electrons are unequally shared between two nuclei. Covalent bonds often result in the formation of small collections of better-connected atoms called molecules, which in solids and liquids are bound to other molecules by forces that are often much weaker than the covalent bonds that hold the molecules internally together. Such weak intermolecular bonds give organic molecular substances, such as waxes and oils, their soft bulk character, and their low melting points (in liquids, molecules must cease most structured or oriented contact with each other). When covalent bonds link long chains of atoms in large molecules, however (as in polymers such as nylon), or when covalent bonds extend in networks through solids that are not composed of discrete molecules (such as diamond or quartz or the silicate minerals in many types of rock) then the structures that result may be both strong and tough, at least in the direction oriented correctly with networks of covalent bonds. Also, the melting points of such covalent polymers and networks increase greatly.
In a simplified view of an "ionic" bond, the bonding electron is not shared at all, but transferred. In this type of bond, the outer atomic orbital of one atom has a vacancy which allows the addition of one or more electrons. These newly added electrons potentially occupy a lower energy-state (effectively closer to more nuclear charge) than they experience in a different atom. Thus, one nucleus offers a more tightly bound position to an electron than does another nucleus, with the result that one atom may transfer an electron to the other. This transfer causes one atom to assume a net positive charge, and the other to assume a net negative charge. The "bond" then results from electrostatic attraction between the positive and negatively charged ions. Ionic bonds may be seen as extreme examples of polarization in covalent bonds. Often, such bonds have no particular orientation in space, since they result from equal electrostatic attraction of each ion to all ions around them. Ionic bonds are strong (and thus ionic substances require high temperatures to melt) but also brittle, since the forces between ions are short-range and do not easily bridge cracks and fractures. This type of bond gives rise to the physical characteristics of crystals of classic mineral salts, such as table salt.
A less often mentioned type of bonding is "metallic" bonding. In this type of bonding, each atom in a metal donates one or more electrons to a "sea" of electrons that reside between many metal atoms. In this sea, each electron is free (by virtue of its wave nature) to be associated with a great many atoms at once. The bond results because the metal atoms become somewhat positively charged due to loss of their electrons while the electrons remain attracted to many atoms, without being part of any given atom. Metallic bonding may be seen as an extreme example of delocalization of electrons over a large system of covalent bonds, in which every atom participates. This type of bonding is often very strong (resulting in the tensile strength of metals). However, metallic bonding is more collective in nature than other types, and so they allow metal crystals to more easily deform, because they are composed of atoms attracted to each other, but not in any particularly-oriented ways. This results in the malleability of metals. The cloud of electrons in metallic bonding causes the characteristically good electrical and thermal conductivity of metals, and also their shiny lustre that reflects most frequencies of white light.
Early speculations about the nature of the chemical bond, from as early as the 12th century, supposed that certain types of chemical species were joined by a type of chemical affinity. In 1704, Sir Isaac Newton famously outlined his atomic bonding theory, in "Query 31" of his "Opticks", whereby atoms attach to each other by some "force". Specifically, after acknowledging the various popular theories in vogue at the time, of how atoms were reasoned to attach to each other, i.e. "hooked atoms", "glued together by rest", or "stuck together by conspiring motions", Newton states that he would rather infer from their cohesion, that "particles attract one another by some force, which in immediate contact is exceedingly strong, at small distances performs the chemical operations, and reaches not far from the particles with any sensible effect."
In 1819, on the heels of the invention of the voltaic pile, Jöns Jakob Berzelius developed a theory of chemical combination stressing the electronegative and electropositive characters of the combining atoms. By the mid 19th century, Edward Frankland, F.A. Kekulé, A.S. Couper, Alexander Butlerov, and Hermann Kolbe, building on the theory of radicals, developed the theory of valency, originally called "combining power", in which compounds were joined owing to an attraction of positive and negative poles. In 1916, chemist Gilbert N. Lewis developed the concept of the electron-pair bond, in which two atoms may share one to six electrons, thus forming the single electron bond, a single bond, a double bond, or a triple bond; in Lewis's own words, "An electron may form a part of the shell of two different atoms and cannot be said to belong to either one exclusively."
That same year, Walther Kossel put forward a theory similar to Lewis' only his model assumed complete transfers of electrons between atoms, and was thus a model of ionic bonding. Both Lewis and Kossel structured their bonding models on that of Abegg's rule (1904).
Niels Bohr proposed a model of the atom and a model of the chemical bond. According to his model for a diatomic molecule, the electrons of the atoms of the molecule form a rotating ring whose plane is perpendicular to the axis of the molecule and equidistant from the atomic nuclei. The dynamic equilibrium of the molecular system is achieved through the balance of forces between the forces of attraction of nuclei to the plane of the ring of electrons and the forces of mutual repulsion of the nuclei. The Bohr model of the chemical bond took into account the Coulomb repulsion – the electrons in the ring are at the maximum distance from each other.
In 1927, the first mathematically complete quantum description of a simple chemical bond, i.e. that produced by one electron in the hydrogen molecular ion, H2+, was derived by the Danish physicist Øyvind Burrau. This work showed that the quantum approach to chemical bonds could be fundamentally and quantitatively correct, but the mathematical methods used could not be extended to molecules containing more than one electron. A more practical, albeit less quantitative, approach was put forward in the same year by Walter Heitler and Fritz London. The Heitler–London method forms the basis of what is now called valence bond theory. In 1929, the linear combination of atomic orbitals molecular orbital method (LCAO) approximation was introduced by Sir John Lennard-Jones, who also suggested methods to derive electronic structures of molecules of F2 (fluorine) and O2 (oxygen) molecules, from basic quantum principles. This molecular orbital theory represented a covalent bond as an orbital formed by combining the quantum mechanical Schrödinger atomic orbitals which had been hypothesized for electrons in single atoms. The equations for bonding electrons in multi-electron atoms could not be solved to mathematical perfection (i.e., "analytically"), but approximations for them still gave many good qualitative predictions and results. Most quantitative calculations in modern quantum chemistry use either valence bond or molecular orbital theory as a starting point, although a third approach, density functional theory, has become increasingly popular in recent years.
In 1933, H. H. James and A. S. Coolidge carried out a calculation on the dihydrogen molecule that, unlike all previous calculation which used functions only of the distance of the electron from the atomic nucleus, used functions which also explicitly added the distance between the two electrons. With up to 13 adjustable parameters they obtained a result very close to the experimental result for the dissociation energy. Later extensions have used up to 54 parameters and gave excellent agreement with experiments. This calculation convinced the scientific community that quantum theory could give agreement with experiment. However this approach has none of the physical pictures of the valence bond and molecular orbital theories and is difficult to extend to larger molecules.
Because atoms and molecules are three-dimensional, it is difficult to use a single method to indicate orbitals and bonds. In molecular formulas the chemical bonds (binding orbitals) between atoms are indicated in different ways depending on the type of discussion. Sometimes, some details are neglected. For example, in organic chemistry one is sometimes concerned only with the functional group of the molecule. Thus, the molecular formula of ethanol may be written in conformational form, three-dimensional form, full two-dimensional form (indicating every bond with no three-dimensional directions), compressed two-dimensional form (CH3–CH2–OH), by separating the functional group from another part of the molecule (C2H5OH), or by its atomic constituents (C2H6O), according to what is discussed. Sometimes, even the non-bonding valence shell electrons (with the two-dimensional approximate directions) are marked, e.g. for elemental carbon .'C'. Some chemists may also mark the respective orbitals, e.g. the hypothetical ethene−4 anion (\/C=C/\ −4) indicating the possibility of bond formation.
Strong chemical bonds are the "intramolecular" forces that hold atoms together in molecules. A strong chemical bond is formed from the transfer or sharing of electrons between atomic centers and relies on the electrostatic attraction between the protons in nuclei and the electrons in the orbitals.
The types of strong bond differ due to the difference in electronegativity of the constituent elements. A large difference in electronegativity leads to more polar (ionic) character in the bond.
Ionic bonding is a type of electrostatic interaction between atoms that have a large electronegativity difference. There is no precise value that distinguishes ionic from covalent bonding, but an electronegativity difference of over 1.7 is likely to be ionic while a difference of less than 1.7 is likely to be covalent. Ionic bonding leads to separate positive and negative ions. Ionic charges are commonly between −3e to +3e. Ionic bonding commonly occurs in metal salts such as sodium chloride (table salt). A typical feature of ionic bonds is that the species form into ionic crystals, in which no ion is specifically paired with any single other ion in a specific directional bond. Rather, each species of ion is surrounded by ions of the opposite charge, and the spacing between it and each of the oppositely charged ions near it is the same for all surrounding atoms of the same type. It is thus no longer possible to associate an ion with any specific other single ionized atom near it. This is a situation unlike that in covalent crystals, where covalent bonds between specific atoms are still discernible from the shorter distances between them, as measured via such techniques as X-ray diffraction.
Ionic crystals may contain a mixture of covalent and ionic species, as for example salts of complex acids such as sodium cyanide, NaCN. X-ray diffraction shows that in NaCN, for example, the bonds between sodium cations (Na+) and the cyanide anions (CN−) are "ionic", with no sodium ion associated with any particular cyanide. However, the bonds between C and N atoms in cyanide are of the "covalent" type, so that each carbon is strongly bound to "just one" nitrogen, to which it is physically much closer than it is to other carbons or nitrogens in a sodium cyanide crystal.
When such crystals are melted into liquids, the ionic bonds are broken first because they are non-directional and allow the charged species to move freely. Similarly, when such salts dissolve into water, the ionic bonds are typically broken by the interaction with water but the covalent bonds continue to hold. For example, in solution, the cyanide ions, still bound together as single CN− ions, move independently through the solution, as do sodium ions, as Na+. In water, charged ions move apart because each of them are more strongly attracted to a number of water molecules than to each other. The attraction between ions and water molecules in such solutions is due to a type of weak dipole-dipole type chemical bond. In melted ionic compounds, the ions continue to be attracted to each other, but not in any ordered or crystalline way.
Covalent bonding is a common type of bonding in which two or more atoms share valence electrons more or less equally. The simplest and most common type is a single bond in which two atoms share two electrons. Other types include the double bond, the triple bond, one- and three-electron bonds, the three-center two-electron bond and three-center four-electron bond.
In non-polar covalent bonds, the electronegativity difference between the bonded atoms is small, typically 0 to 0.3. Bonds within most organic compounds are described as covalent. The figure shows methane (CH4), in which each hydrogen forms a covalent bond with the carbon. See sigma bonds and pi bonds for LCAO descriptions of such bonding.
Molecules that are formed primarily from non-polar covalent bonds are often immiscible in water or other polar solvents, but much more soluble in non-polar solvents such as hexane.
A polar covalent bond is a covalent bond with a significant ionic character. This means that the two shared electrons are closer to one of the atoms than the other, creating an imbalance of charge. Such bonds occur between two atoms with moderately different electronegativities and give rise to dipole–dipole interactions. The electronegativity difference between the two atoms in these bonds is 0.3 to 1.7.
A single bond between two atoms corresponds to the sharing of one pair of electrons. The Hydrogen (H) atom has one valence electron. Two Hydrogen atoms can then form a molecule, held together by the shared pair of electrons. Each H atom now has the noble gas electron configuration of helium (He). The pair of shared electrons forms a single covalent bond. The electron density of these two bonding electrons in the region between the two atoms increases from the density of two non-interacting H atoms.
A double bond has two shared pairs of electrons, one in a sigma bond and one in a pi bond with electron density concentrated on two opposite sides of the internuclear axis. A triple bond consists of three shared electron pairs, forming one sigma and two pi bonds. An example is nitrogen. Quadruple and higher bonds are very rare and occur only between certain transition metal atoms.
A coordinate covalent bond is a covalent bond in which the two shared bonding electrons are from the same one of the atoms involved in the bond. For example, boron trifluoride (BF3) and ammonia (NH3) form an adduct or coordination complex F3B←NH3 with a B–N bond in which a lone pair of electrons on N is shared with an empty atomic orbital on B. BF3 with an empty orbital is described as an electron pair acceptor or Lewis acid, while NH3 with a lone pair that can be shared is described as an electron-pair donor or Lewis base. The electrons are shared roughly equally between the atoms in contrast to ionic bonding. Such bonding is shown by an arrow pointing to the Lewis acid.
Transition metal complexes are generally bound by coordinate covalent bonds. For example, the ion Ag+ reacts as a Lewis acid with two molecules of the Lewis base NH3 to form the complex ion Ag(NH3)2+, which has two Ag←N coordinate covalent bonds.
In metallic bonding, bonding electrons are delocalized over a lattice of atoms. By contrast, in ionic compounds, the locations of the binding electrons and their charges are static. The free movement or delocalization of bonding electrons leads to classical metallic properties such as luster (surface light reflectivity), electrical and thermal conductivity, ductility, and high tensile strength.
There are four basic types of bonds that can be formed between two or more (otherwise non-associated) molecules, ions or atoms. Intermolecular forces cause molecules to be attracted or repulsed by each other. Often, these define some of the physical characteristics (such as the melting point) of a substance.
In the (unrealistic) limit of "pure" ionic bonding, electrons are perfectly localized on one of the two atoms in the bond. Such bonds can be understood by classical physics. The forces between the atoms are characterized by isotropic continuum electrostatic potentials. Their magnitude is in simple proportion to the charge difference.
Covalent bonds are better understood by valence bond (VB) theory or molecular orbital (MO) theory. The properties of the atoms involved can be understood using concepts such as oxidation number, formal charge, and electronegativity. The electron density within a bond is not assigned to individual atoms, but is instead delocalized between atoms. In valence bond theory, bonding is conceptualized as being built up from electron pairs that are localized and shared by two atoms via the overlap of atomic orbitals. The concepts of orbital hybridization and resonance augment this basic notion of the electron pair bond. In molecular orbital theory, bonding is viewed as being delocalized and apportioned in orbitals that extend throughout the molecule and are adapted to its symmetry properties, typically by considering linear combinations of atomic orbitals (LCAO). Valence bond theory is more chemically intuitive by being spatially localized, allowing attention to be focused on the parts of the molecule undergoing chemical change. In contrast, molecular orbitals are more "natural" from a quantum mechanical point of view, with orbital energies being physically significant and directly linked to experimental ionization energies from photoelectron spectroscopy. Consequently, valence bond theory and molecular orbital theory are often viewed as competing but complementary frameworks that offer different insights into chemical systems. As approaches for electronic structure theory, both MO and VB methods can give approximations to any desired level of accuracy, at least in principle. However, at lower levels, the approximations differ, and one approach may be better suited for computations involving a particular system or property than the other.
Unlike the spherically symmetrical Coulombic forces in pure ionic bonds, covalent bonds are generally directed and anisotropic. These are often classified based on their symmetry with respect to a molecular plane as sigma bonds and pi bonds. In the general case, atoms form bonds that are intermediate between ionic and covalent, depending on the relative electronegativity of the atoms involved. Bonds of this type are known as polar covalent bonds. | https://en.wikipedia.org/wiki?curid=5993 |
Climate
Climate is the long-term average of weather, typically averaged over a period of 30 years. Some of the meteorological variables that are commonly measured are temperature, humidity, atmospheric pressure, wind, and precipitation. In a broader sense, climate is the state of the components of the climate system, which includes the ocean and ice on Earth. The climate of a location is affected by its latitude, terrain, and altitude, as well as nearby water bodies and their currents.
More generally, the "climate" of a region is the general state of the climate system at that location at the current time.
Climates can be classified according to the average and the typical ranges of different variables, most commonly temperature and precipitation. The most commonly used classification scheme was the Köppen climate classification. The Thornthwaite system, in use since 1948, incorporates evapotranspiration along with temperature and precipitation information and is used in studying biological diversity and how climate change affects it. The Bergeron and Spatial Synoptic Classification systems focus on the origin of air masses that define the climate of a region.
Paleoclimatology is the study of ancient climates. Since very few direct observations of climate are available before the 19th century, paleoclimates are inferred from proxy variables that include non-biotic evidence such as sediments found in lake beds and ice cores, and biotic evidence such as tree rings and coral. Climate models are mathematical models of past, present and future climates. Climate change may occur over long and short timescales from a variety of factors; recent warming is discussed in global warming. Global warming results in redistributions. For example, "a 3°C change in mean annual temperature corresponds to a shift in isotherms of approximately 300–400 km in latitude (in the temperate zone) or 500 m in elevation. Therefore, species are expected to move upwards in elevation or towards the poles in latitude in response to shifting climate zones".
Climate (from Ancient Greek "klima", meaning "inclination") is commonly defined as the weather averaged over a long period. The standard averaging period is 30 years, but other periods may be used depending on the purpose. Climate also includes statistics other than the average, such as the magnitudes of day-to-day or year-to-year variations. The Intergovernmental Panel on Climate Change (IPCC) 2001 glossary definition is as follows:
The World Meteorological Organization (WMO) describes climate "normals" as "reference points used by climatologists to compare current climatological trends to that of the past or what is considered 'normal'. A Normal is defined as the arithmetic average of a climate element (e.g. temperature) over a 30-year period. A 30 year period is used, as it is long enough to filter out any interannual variation or anomalies, but also short enough to be able to show longer climatic trends." The WMO originated from the International Meteorological Organization which set up a technical commission for climatology in 1929. At its 1934 Wiesbaden meeting the technical commission designated the thirty-year period from 1901 to 1930 as the reference time frame for climatological standard normals. In 1982 the WMO agreed to update climate normals, and these were subsequently completed on the basis of climate data from 1 January 1961 to 31 December 1990.
The difference between climate and weather is usefully summarized by the popular phrase "Climate is what you expect, weather is what you get." Over historical time spans, there are a number of nearly constant variables that determine climate, including latitude, altitude, proportion of land to water, and proximity to oceans and mountains. All of these variables change only over periods of millions of years due to processes such as plate tectonics. Other climate determinants are more dynamic: the thermohaline circulation of the ocean leads to a 5 °C (9 °F) warming of the northern Atlantic Ocean compared to other ocean basins. Other ocean currents redistribute heat between land and water on a more regional scale. The density and type of vegetation coverage affects solar heat absorption, water retention, and rainfall on a regional level. Alterations in the quantity of atmospheric greenhouse gases determines the amount of solar energy retained by the planet, leading to global warming or global cooling. The variables which determine climate are numerous and the interactions complex, but there is general agreement that the broad outlines are understood, at least insofar as the determinants of historical climate change are concerned.
There are several ways to classify climates into similar regimes. Originally, climes were defined in Ancient Greece to describe the weather depending upon a location's latitude. Modern climate classification methods can be broadly divided into "genetic" methods, which focus on the causes of climate, and "empiric" methods, which focus on the effects of climate. Examples of genetic classification include methods based on the relative frequency of different air mass types or locations within synoptic weather disturbances. Examples of empiric classifications include climate zones defined by plant hardiness, evapotranspiration, or more generally the Köppen climate classification which was originally designed to identify the climates associated with certain biomes. A common shortcoming of these classification schemes is that they produce distinct boundaries between the zones they define, rather than the gradual transition of climate properties more common in nature.
The simplest classification is that involving air masses. The Bergeron classification is the most widely accepted form of air mass classification. Air mass classification involves three letters. The first letter describes its moisture properties, with c used for continental air masses (dry) and m for maritime air masses (moist). The second letter describes the thermal characteristic of its source region: T for tropical, P for polar, A for Arctic or Antarctic, M for monsoon, E for equatorial, and S for superior air (dry air formed by significant downward motion in the atmosphere). The third letter is used to designate the stability of the atmosphere. If the air mass is colder than the ground below it, it is labeled k. If the air mass is warmer than the ground below it, it is labeled w. While air mass identification was originally used in weather forecasting during the 1950s, climatologists began to establish synoptic climatologies based on this idea in 1973.
Based upon the Bergeron classification scheme is the Spatial Synoptic Classification system (SSC). There are six categories within the SSC scheme: Dry Polar (similar to continental polar), Dry Moderate (similar to maritime superior), Dry Tropical (similar to continental tropical), Moist Polar (similar to maritime polar), Moist Moderate (a hybrid between maritime polar and maritime tropical), and Moist Tropical (similar to maritime tropical, maritime monsoon, or maritime equatorial).
The Köppen classification depends on average monthly values of temperature and precipitation. The most commonly used form of the Köppen classification has five primary types labeled A through E. These primary types are A) tropical, B) dry, C) mild mid-latitude, D) cold mid-latitude, and E) polar. The five primary classifications can be further divided into secondary classifications such as rainforest, monsoon, tropical savanna, humid subtropical, humid continental, oceanic climate, Mediterranean climate, desert, steppe, subarctic climate, tundra, and polar ice cap.
Rainforests are characterized by high rainfall, with definitions setting minimum normal annual rainfall between and . Mean monthly temperatures exceed during all months of the year.
A monsoon is a seasonal prevailing wind which lasts for several months, ushering in a region's rainy season. Regions within North America, South America, Sub-Saharan Africa, Australia and East Asia are monsoon regimes.
A tropical savanna is a grassland biome located in semi-arid to semi-humid climate regions of subtropical and tropical latitudes, with average temperatures remaining at or above all year round, and rainfall between and a year. They are widespread on Africa, and are found in India, the northern parts of South America, Malaysia, and Australia.
The humid subtropical climate zone where winter rainfall (and sometimes snowfall) is associated with large storms that the westerlies steer from west to east. Most summer rainfall occurs during thunderstorms and from occasional tropical cyclones. Humid subtropical climates lie on the east side of continents, roughly between latitudes 20° and 40° degrees away from the equator.
A humid continental climate is marked by variable weather patterns and a large seasonal temperature variance. Places with more than three months of average daily temperatures above and a coldest month temperature below and which do not meet the criteria for an arid or semi-arid climate, are classified as continental.
An oceanic climate is typically found along the west coasts at the middle latitudes of all the world's continents, and in southeastern Australia, and is accompanied by plentiful precipitation year-round.
The Mediterranean climate regime resembles the climate of the lands in the Mediterranean Basin, parts of western North America, parts of Western and South Australia, in southwestern South Africa and in parts of central Chile. The climate is characterized by hot, dry summers and cool, wet winters.
A steppe is a dry grassland with an annual temperature range in the summer of up to and during the winter down to .
A subarctic climate has little precipitation, and monthly temperatures which are above for one to three months of the year, with permafrost in large parts of the area due to the cold winters. Winters within subarctic climates usually include up to six months of temperatures averaging below .
Tundra occurs in the far Northern Hemisphere, north of the taiga belt, including vast areas of northern Russia and Canada.
A polar ice cap, or polar ice sheet, is a high-latitude region of a planet or moon that is covered in ice. Ice caps form because high-latitude regions receive less energy as solar radiation from the sun than equatorial regions, resulting in lower surface temperatures.
A desert is a landscape form or region that receives very little precipitation. Deserts usually have a large diurnal and seasonal temperature range, with high or low, depending on location daytime temperatures (in summer up to ), and low nighttime temperatures (in winter down to ) due to extremely low humidity. Many deserts are formed by rain shadows, as mountains block the path of moisture and precipitation to the desert.
Devised by the American climatologist and geographer C. W. Thornthwaite, this climate classification method monitors the soil water budget using evapotranspiration. It monitors the portion of total precipitation used to nourish vegetation over a certain area. It uses indices such as a humidity index and an aridity index to determine an area's moisture regime based upon its average temperature, average rainfall, and average vegetation type. The lower the value of the index in any given area, the drier the area is.
The moisture classification includes climatic classes with descriptors such as hyperhumid, humid, subhumid, subarid, semi-arid (values of −20 to −40), and arid (values below −40). Humid regions experience more precipitation than evaporation each year, while arid regions experience greater evaporation than precipitation on an annual basis. A total of 33 percent of the Earth's landmass is considered either arid or semi-arid, including southwest North America, southwest South America, most of northern and a small part of southern Africa, southwest and portions of eastern Asia, as well as much of Australia. Studies suggest that precipitation effectiveness (PE) within the Thornthwaite moisture index is overestimated in the summer and underestimated in the winter. This index can be effectively used to determine the number of herbivore and mammal species numbers within a given area. The index is also used in studies of climate change.
Thermal classifications within the Thornthwaite scheme include microthermal, mesothermal, and megathermal regimes. A microthermal climate is one of low annual mean temperatures, generally between and which experiences short summers and has a potential evaporation between and . A mesothermal climate lacks persistent heat or persistent cold, with potential evaporation between and . A megathermal climate is one with persistent high temperatures and abundant rainfall, with potential annual evaporation in excess of .
Paleoclimatology is the study of past climate over a great period of the Earth's history. It uses evidence from ice sheets, tree rings, sediments, coral, and rocks to determine the past state of the climate. It demonstrates periods of stability and periods of change and can indicate whether changes follow patterns such as regular cycles.
Details of the modern climate record are known through the taking of measurements from such weather instruments as thermometers, barometers, and anemometers during the past few centuries. The instruments used to study weather over the modern time scale, their known error, their immediate environment, and their exposure have changed over the years, which must be considered when studying the climate of centuries past.
Climate variability is the term to describe variations in the mean state and other characteristics of climate (such as chances or possibility of extreme weather, etc.) "on all spatial and temporal scales beyond that of individual weather events." Some of the variability does not appear to be caused systematically and occurs at random times. Such variability is called "random variability" or "noise". On the other hand, periodic variability occurs relatively regularly and in distinct modes of variability or climate patterns.
There are close correlations between Earth's climate oscillations and astronomical factors (barycenter changes, solar variation, cosmic ray flux, cloud albedo feedback, Milankovic cycles), and modes of heat distribution between the ocean-atmosphere climate system. In some cases, current, historical and paleoclimatological natural oscillations may be masked by significant volcanic eruptions, impact events, irregularities in climate proxy data, positive feedback processes or anthropogenic emissions of substances such as greenhouse gases.
Over the years, the definitions of "climate variability" and the related term "climate change" have shifted. While the term "climate change" now implies change that is both long-term and of human causation, in the 1960s the word climate change was used for what we now describe as climate variability, that is, climatic inconsistencies and anomalies.
Climate change is the variation in global or regional climates over time. It reflects changes in the variability or average state of the atmosphere over time scales ranging from decades to millions of years. These changes can be caused by processes internal to the Earth, external forces (e.g. variations in sunlight intensity) or, more recently, human activities.
In recent usage, especially in the context of environmental policy, the term "climate change" often refers only to changes in modern climate, including the rise in average surface temperature known as global warming. In some cases, the term is also used with a presumption of human causation, as in the United Nations Framework Convention on Climate Change (UNFCCC). The UNFCCC uses "climate variability" for non-human caused variations.
Earth has undergone periodic climate shifts in the past, including four major ice ages. These consisting of glacial periods where conditions are colder than normal, separated by interglacial periods. The accumulation of snow and ice during a glacial period increases the surface albedo, reflecting more of the Sun's energy into space and maintaining a lower atmospheric temperature. Increases in greenhouse gases, such as by volcanic activity, can increase the global temperature and produce an interglacial period. Suggested causes of ice age periods include the positions of the continents, variations in the Earth's orbit, changes in the solar output, and volcanism.
Climate models use quantitative methods to simulate the interactions of the atmosphere, oceans, land surface and ice. They are used for a variety of purposes; from the study of the dynamics of the weather and climate system, to projections of future climate. All climate models balance, or very nearly balance, incoming energy as short wave (including visible) electromagnetic radiation to the earth with outgoing energy as long wave (infrared) electromagnetic radiation from the earth. Any imbalance results in a change in the average temperature of the earth.
The most talked-about applications of these models in recent years have been their use to infer the consequences of increasing greenhouse gases in the atmosphere, primarily carbon dioxide (see greenhouse gas). These models predict an upward trend in the global mean surface temperature, with the most rapid increase in temperature being projected for the higher latitudes of the Northern Hemisphere.
Models can range from relatively simple to quite complex: | https://en.wikipedia.org/wiki?curid=5999 |
History of the Comoros
The history of the Comoros extends to about 1000 BC when the archipelago was first inhabited. The Comoros have been inhabited by various groups throughout this time. France colonised the islands in the 19th century, and they became independent in 1975.
The earliest inhabitants of the islands were likely people from the Swahili culture from as early as 1000 BC, small trading communities of Bantu-speakers from the coasts of East Africa.
In the 8th to 13th centuries AD they were followed by an influx of Austronesian sailors from Southeast Asia, who had earlier settled nearby Madagascar. They are the source for the earliest archeological evidence of farming in the islands. Crops from archeological sites in Sima are predominantly rice strains of both "indica" and "japonica" varieties from Southeast Asia, as well as various other Asian crops like mung bean and cotton. Only a minority of the examined crops were African-derived, like finger millet, African sorghum, and cowpea. The Comoros are believed to be the first site of contact and subsequent admixture between African and Asian populations (earlier than Madagascar). Comorians today still display at most 20% Austronesian admixture.
From around the 15th century AD, Shirazi slave traders established trading ports and brought in slaves from the mainland. In the 16th century, social changes on the East African coast probably linked to the arrival of the Portuguese saw the arrival of a number of Arabs of Hadrami who established alliances with the Shirazis and founded several royal clans.
Over the centuries, the Comoros have been settled by a succession of diverse groups from the coast of Africa, the Persian Gulf, Southeast Asia and Madagascar.
Portuguese explorers first visited the archipelago in 1505.
Apart from a visit by the French Parmentier brothers in 1529, for much of the 16th century the only Europeans to visit the islands were Portuguese. British and Dutch ships began arriving around the start of the 17th century and the island of Ndzwani soon became a major supply point on the route to the East Indies. Ndzwani was generally ruled by a single sultan, who occasionally attempted to extend his authority to Mayotte and Mwali; Ngazidja was more fragmented, on occasion being divided into as many as 12 small kingdoms.
Sir James Lancaster's voyage to the Indian Ocean in 1591 was the first attempt by the English to break into the spice trade, which was dominated by the Portuguese. Only one of his four ships made it back from the Indies on that voyage, and that one with a decimated crew of 5 men and a boy. Lancaster himself was marooned by a cyclone on the Comoros. Many of his crew were speared to death by angry islanders although Lancaster found his way home in 1594. (Dalrymple W. 2019; Bloomsbury Publishing ).
Both the British and the French turned their attention to the Comoros islands in the middle of the 19th century. The French finally acquired the islands through a cunning mixture of strategies, including the policy of "divide and conquer", chequebook politics and a serendipitous affair between a sultana and a French trader that was put to good use by the French, who kept control of the islands, quelling unrest and the occasional uprising.
William Sunley, a planter and British Consul from 1848–1866, was an influence on Anjouan.
On 25 March 1841, France purchased the island of Maore — the name of the island was corrupted in French to "Mayotte" — which became a colony upon ratification on 13 June 1843.
In 1850 Sultan Selim of Johanna island seized the American whaler "Maria" and imprisoned her commander, Captain Moores. In response, the United States Navy launched the Johanna Expedition in February 1852 to gain the release of Moores and extract compensation. Initially, the sultan did not meet the demands, and the sloop-of-war USS "Dale" bombarded the island's fortifications; ultimately Selim paid US$1,000 and released Captain Moores.
In 1886 Said Ali bin Said Omar, Sultan of Bambao, signed an agreement with the French government that allowed France to establish a protectorate over the entire island of Ngazidja (Grande Comore); protectorates were also established over Ndzwani (Anjouan), and Mwali (Mohéli island in French) the same year. Résidents were posted on the three islands.
On 9 April 1908, France declared the protectorates and Mayotte a single colony, Mayotte and dependencies.
On 25 July 1912, it was annexed to Madagascar as a province of that colony.
From 16 June 1940 to 1942 the colonial administration remained loyal to Vichy France (from 1942, under Free French), but 25 September 1942 to 13 October 1946 they were, like Madagascar, under British occupation.
Until the opening of the Suez Canal, the islands used to be an important refuelling and provisioning station for ships from Europe to the Indian Ocean.
Independence came gradually for the Comoros. During the middle of the 20th century the French reluctantly began to accede to requests for constitutional changes and in 1946 the Comoros had become a separately administered colony from Madagascar.
After World War II, the islands became a French overseas territory and were represented in France's National Assembly. Internal political autonomy was granted in 1961. Agreement was reached with France in 1973 for the Comoros to become independent in 1978. On July 6, 1975, however, the Comorian parliament passed a resolution declaring unilateral independence. The deputies of Mayotte abstained.
In 1961 the Comoros was granted autonomous rule and, in 1975, it broke all ties with France and established itself as an independent republic. From the very beginning Mayotte refused to join the new republic and aligned itself even more firmly to the French Republic, but the other islands remained committed to independence. The first president of the Comoros, Ahmed Abdallah Abderemane, did not last long before being ousted in a coup d'état by Ali Soilih, an atheist with an Islamic background.
Soilih began with a set of solid socialist ideals designed to modernize the country. However, the regime faced problems. A French mercenary by the name of Bob Denard, arrived in the Comoros at dawn on 13 May 1978, and removed Soilih from power. Solih was shot and killed during the coup. Abdallah returned to govern the country and the mercenaries were given key positions in government.
In two referendums, in December 1974 and February 1976, the population of Mayotte voted against independence from France (by 63.8% and 99.4% respectively). Mayotte thus remains under French administration, and the Comorian Government has effective control over only Grande Comore, Anjouan, and Mohéli.
Later, French settlers, French-owned companies, and Arab merchants established a plantation-based economy that now uses about one-third of the land for export crops.
In 1978, president Ali Soilih, who had a firm anti-French line, was killed and Ahmed Abdallah came to power. Under the reign of Abdallah, Denard was commander of the Presidential Guard (PG) and "de facto" ruler of the country. He was trained, supported and funded by the white regimes in South Africa (SA) and Rhodesia (now Zimbabwe) in return for permission to set up a secret listening post on the islands. South-African agents kept an ear on the important ANC bases in Lusaka and Dar es Salaam and watched the war in Mozambique, in which SA played an active role. The Comoros were also used for the evasion of arms sanctions.
When in 1981 François Mitterrand was elected president Denard lost the support of the French intelligence service, but he managed to strengthen the link between SA and the Comoros. Besides the military, Denard established his own company SOGECOM, for both the security and construction, and seemed to profit by the arrangement. Between 1985 and 1987 the relationship of the PG with the local Comorians became worse.
At the end of the 1980s the South Africans did not wish to continue to support the mercenary regime and France was in agreement. Also President Abdallah wanted the mercenaries to leave. Their response was a (third) coup resulting in the death of President Abdallah, in which Denard and his men were probably involved. South Africa and the French government subsequently forced Denard and his mercenaries to leave the islands in 1989.
Said Mohamed Djohar became president. His time in office was turbulent, including an impeachment attempt in 1991 and a coup attempt in 1992.
On September 28, 1995 Bob Denard and a group of mercenaries took over the Comoros islands in a coup (named operation Kaskari by the mercenaries) against President Djohar. France immediately severely denounced the coup, and backed by the 1978 defense agreement with the Comoros, President Jacques Chirac ordered his special forces to retake the island. Bob Denard began to take measures to stop the coming invasion. A new presidential guard was created. Strong points armed with heavy machine guns were set up around the island, particularly around the island's two airports.
On October 3, 1995, 11 p.m., the French deployed 600 men against a force of 33 mercenaries and a 300-man dissident force. Denard however ordered his mercenaries not to fight. Within 7 hours the airports at Iconi and Hahaya and the French Embassy in Moroni were secured. By 3:00 p.m. the next day Bob Denard and his mercenaries had surrendered. This (response) operation, codenamed "Azalée", was remarkable, because there were no casualties, and just in seven days, plans were drawn up and soldiers were deployed. Denard was taken to France and jailed. Prime minister Caambi El-Yachourtu became acting president until Djohar returned from exile in January, 1996. In March 1996, following presidential elections, Mohamed Taki Abdoulkarim, a member of the civilian government that Denard had tried to set up in October 1995, became president. On 23 November 1996, Ethiopian Airlines Flight 961 crashed near a beach on the island after it was hijacked and ran out of fuel killing 125 people and leaving 50 survivors.
In 1997, the islands of Anjouan and Mohéli declared their independence from the Comoros. A subsequent attempt by the government to re-establish control over the rebellious islands by force failed, and presently the African Union is brokering negotiations to effect a reconciliation. This process is largely complete, at least in theory. According to some sources, Mohéli did return to government control in 1998. In 1999, Anjouan had internal conflicts and on August 1 of that year, the 80-year-old first president Foundi Abdallah Ibrahim resigned, transferring power to a national coordinator, Said Abeid. The government was overthrown in a coup by army and navy officers on August 9, 2001. Mohamed Bacar soon rose to leadership of the junta that took over and by the end of the month he was the leader of the country. Despite two coup attempts in the following three months, including one by Abeid, Bacar's government remained in power, and was apparently more willing to negotiate with the Comoros. Presidential elections were held for all of the Comoros in 2002, and presidents have been chosen for all three islands as well, which have become a confederation. Most notably, Mohammed Bacar was elected for a 5-year term as president of Anjouan. Grande Comore had experienced troubles of its own in the late 1990s, when President Taki died on November 6, 1998. Colonel Azali Assoumani became president following a military coup in 1999. There have been several coup attempts since, but he gained firm control of the country after stepping down temporarily and winning a presidential election in 2002.
In May 2006, Ahmed Abdallah Sambi was elected from the island of Anjouan to be the president of the Union of the Comoros. He is a well-respected Sunni cleric who studied in the Sudan, Iran and Saudi Arabia. He is respectfully called "Ayatollah" by his supporters but is considered, and is, a moderate Islamist. He has been quoted as stating that the Comoros is not ready to become an Islamic state, nor shall the veil be forced upon any women in the Comoros. | https://en.wikipedia.org/wiki?curid=6000 |
Demographics of the Comoros
The Comorians () inhabiting Grande Comore, Anjouan, and Mohéli (86% of the population) share African-Arab origins. Islam is the dominant religion, and Quranic schools for children reinforce its influence. Although Islamic culture is firmly established throughout, a small minority are Christian.
The most common language is Comorian, related to Swahili. French and Arabic also are spoken. About 89% of the population is literate.
The Comoros have had seven censuses since World War II:
Population density figures conceal a great disparity between the republic's most crowded island, Nzwani, which had a density of 470 persons per square kilometer in 1991; Ngazidja, which had a density of 250 persons per square kilometer in 1991; and Mwali, where the 1991 population density figure was 120 persons per square kilometer. Overall population density increased to about 285 persons per square kilometer by 1994.
By comparison, estimates of the population density per square kilometer of the Indian Ocean's other island microstates ranged from 241 (Seychelles) to 690 (Maldives) in 1993. Given the rugged terrain of Ngazidja and Nzwani, and the dedication of extensive tracts to agriculture on all three islands, population pressures on the Comoros are becoming increasingly critical.
The age structure of the population of the Comoros is similar to that of many developing countries, in that the republic has a very large proportion of young people. In 1989, 46.4 percent of the population was under fifteen years of age, an above-average proportion even for sub-Saharan Africa. The population's rate of growth was a relatively high 3.5 percent per annum in the mid 1980s, up substantially from 2.0 percent in the mid-1970s and 2.1 percent in the mid-1960s.
In 1983 the Abdallah regime borrowed US$2.85 million from the International Development Association to devise a national family planning program. However, Islamic reservations about contraception made forthright advocacy and implementation of birth control programs politically hazardous, and consequently little was done in the way of public policy.
The Comorian population has become increasingly urbanized in recent years. In 1991 the percentage of Comorians residing in cities and towns of more than 5,000 persons was about 30 percent, up from 25 percent in 1985 and 23 percent in 1980. The Comoros' largest cities were the capital, Moroni, with about 30,000 people, and the port city of Mutsamudu, on the island of Nzwani, with about 20,000 people.
Migration among the various islands is important. Natives of Nzwani have settled in significant numbers on less crowded Mwali, causing some social tensions, and many Nzwani also migrate to Maore. In 1977 Maore expelled peasants from Ngazidja and Nzwani who had recently settled in large numbers on the island. Some were allowed to reenter starting in 1981 but solely as migrant labor.
The number of Comorians living abroad has been estimated at between 80,000 and 100,000; during the colonial period, most of them lived in Tanzania, Madagascar, and other parts of Southeast Africa. The number of Comorians residing in Madagascar was drastically reduced after anti-Comorian rioting in December 1976 in Mahajanga, in which at least 1,400 Comorians were killed. As many as 17,000 Comorians left Madagascar to seek refuge in their native land in 1977 alone. About 100,000 Comorians live in France; many of them had gone there for a university education and never returned. Small numbers of Indians, Malagasy, South Africans, and Europeans (mostly French) live on the islands and play an important role in the economy. Most French left after independence in 1975.
Some Persian Gulf countries started buying Comorian citizenship for their stateless bidoon residents and deporting them to Comoros.
90% of the people living in the Comoros are black, and 10% are mixed race, mostly black and white.
UN medium variant projections.
Statistics :
Total Fertility Rate (TFR) (Wanted Fertility Rate) and Crude Birth Rate (CBR):
Structure of the population (DHS 2012) (Males 11 088, Females 12 284 = 23 373) :
Fertility data as of 2012 (DHS Program):
Demographic statistics according to the World Population Review in 2019.
The following demographic statistics are from the CIA World Factbook.
total dependency ratio: 75.5 (2015 est.)
youth dependency ratio: 70.5 (2015 est.)
elderly dependency ratio: 5.1 (2015 est.)
potential support ratio: 19.7 (2015 est.) | https://en.wikipedia.org/wiki?curid=6002 |
Politics of the Comoros
The Union of the Comoros consists of the three islands Njazidja (Grande Comoros), Mwali (Moheli) and Nzwani (Anjouan) while the island of Mayotte remains under French administration. The Politics of the Union of the Comoros take place in a framework of a federal presidential republic, whereby the President of the Comoros is both head of state and head of government, and of a multi-party system. Executive power is exercised by the government. Federal legislative power is vested in both the government and parliament. The precolonial legacies of the sultanates linger while the political situation in Comoros has been extremely fluid since the country's independence in 1975, subject to the volatility of coups and political insurrection.
As of 2008, Comoros and Mauritania were considered by US-based organization Freedom House as the only real “electoral democracies” of the Arab World.
Sultanates in the late nineteenth century used a cyclic age system and hierarchical lineage membership to provide the foundation for participation in the political process. In the capital, "the sultan was assisted by his ministers and by a madjelis, an advisory council composed of elders, whom he consulted regularly". Apart from local administration, the age system was used to include the population in decision making, depending on the scope of the decision being made. For example, the elders of the island of Njazidja held considerable influence on the authority of the sultan. Though sultanates granted rights to their free inhabitants, were provided with warriors during war and taxed the towns under their authority, their definition as a state is open to debate. The islands' incorporation as a province of the colony of Madagascar into the French colonial empire marked the end of the sultanates.
Despite French colonization, Comorans identify first with kinship or regional ties and rarely ever with the central government. This is a lingering effect of the sovereign sultanates of pre-colonial times. French colonial administration was based on a misconception that the sultanates operated as absolute monarchs: district boundaries were the same as the sultanates', multiple new taxes forced men into wage labor on colonial plantations and was reinforced through a compulsory public labor system that had little effect on infrastructure. French policy was hampered an by absence of settlers, effective communication across islands, rough geographical terrain and hostility towards the colonial government. Policies were made to apply to Madagascar as a whole and seldom to the nuances of each province: civil servants were typically Christian, unaware of local customs and unable to speak the local language. The French established the Ouatou Akouba in 1915, a local form of governance based on "customary structures" already in place that attempted to model itself after the age system in place under the sultanates. Their understanding of the elders' council as a corporate group bypassed the reality that there were men "who had accomplished the necessary customary rituals to be accorded the status of elder and thus be eligible to participate in the political process in the village", which effectively rendered the French elders' council ineffective. Though the Ouatou Akouba was disbanded, it resulted in the consolidation and formalization of the age system as access to power in the customary and local government spheres. The French failure to establish a functioning state in the Comoros has had repercussions in the post-independence era.
At independence there were five main political parties: OUDZIMA, UMMA, the Comoro People's Democratic Rally, the Comoro National Liberation Movement and the Socialist Objective Party. The political groups previously known simply as the 'green' and 'white' party became the Rassemblement Démocratique du Peuple Comorien (RDPC) and the Union Démocratique des Comores (UDC), headed by Sayyid Muhammad Cheikh and Sayyid Ibrahim. Members from both parties later merged to form OUDZIMA under the leadership of first president Ahmad Abdallah while dissidents from both created UMMA under the leadership of future president Ali Soilih.
Prince Said Ibrahim took power in 1970 but was democratically elected out of office in 1972 in favor of former French senator Ahmed Abdallah. President Abdallah declared independence for all islands, except Mayotte which remained under French administration, in 1975. The threat of renewed socioeconomic marginalization following the transfer of the capital to Ngazidja in 1962, more than social or cultural differences, underlay the island's subsequent rejection of independence. France withdrew all economic and technical support for the now independent state, which would encourage a revolutionary regime under future president Ali Soilih. French military and financial aid to mercenaries brought Prince Said Mohammed Jaffar to power after the United National Front of the Comoros (FNU) party toppled Abdallah's government. This mercenary coup was unique in that, unlike other coups on the continent, it was "uninspired by any ideological convictions". The Jaffar regime's inefficient distribution of resources and poor mismanagement was shown through the expulsion of French civil servants as well as endemic unemployment and food shortages. The regime used famine as "an opportunity to switch food patronage from France to the World Food Programme's emergency aid".
President Jaffar's ousting by Minister of Defense and Justice, Ali Soilih, brought about the "periode noire" (dark period) of the country; you could vote at 14, most civil servants were dismissed and there was a ban on some Islamic customs. He implemented revolutionary social reforms such as replacing French with Shikomoro, burning down the national archives and nationalizing land. His government received support from Egypt, Iraq and Sudan. Soilih's attacks on religious and customary authority contributed to his eventual ousting through a French-backed coup consisting of mercenaries and ex-politicians who together formed the Politico Military Doctorate.
Abdallah was reinstated and constructed a mercantile state by resuscitating the structures of the colonial era. His establishment of a one party state and intolerance for dissent further alienated civil society from the state. In May 1978 the Comoros were renamed the Islamic Republic of the Comoros and continued strengthening ties with the Arab world which resulted in their joining the Arab League. Abdallah's government sought to reverse Soilih's 'de-sacralization' by re-introducing the grand marriage, declaring Arabic the second official language behind French, and creating the office of the Grand Mufti. The doctorate & compromise government was dissolved, constitutional changes removed succession from a politician and neutralized the post of another possible challenger in abolishing the position of Prime Minister, which effectively cemented a client-patron network by making the civil service position dependent on Abdallah's political base. The Democratic Front's (DF) internal opposition to Abdallah was suppressed through the incarceration of over 600 people allegedly involved in a failed coup attempt. Abdallah then stocked the House of Assembly with loyal clientelist supporters through rigged parliamentary elections. All of these actions effectively consolidated Abdallah's position.
Muhammed Djohar succeeded president Abdallah after his assassination in 1989 but was evacuated by French troops after a failed coup attempt in 1996. The Comoros were led by Muhammed Taki Abd al-Karim beginning in 1996 and he was followed by interim president Said Massunde who eventually gave way to Assoumani Azali. Taki's lack of Arab heritage led to his lack of understanding Nzwani's cultural differences and economic problems, as seen by the establishment of the elders council with only loyal Taki supporters. As a result, the council was ignored by the true elders of the island. After Taki's death, a military coup in 1999, the nation's eighteenth since independence in 1975, installed Azali in to power. Colonel Azali Assoumani seized power in a bloodless coup in April 1999, overthrowing Interim President Tadjidine Ben Said Massounde, who himself had held the office since the death of democratically elected President Mohamed Taki Abdoulkarim in November, 1998. In May 1999, Azali decreed a constitution that gave him both executive and legislative powers. Bowing somewhat to international criticism, Azali appointed a civilian Prime Minister, Bainrifi Tarmidi, in December 1999; however, Azali retained the mantle of Head of State and army Commander. In December 2000, Azali named a new civilian Prime Minister, Hamada Madi, and formed a new civilian Cabinet. When Azali took power he also pledged to step down in April 2000 and relinquish control to a democratically elected president—a pledge with mixed results. Under Mohammed Taki and Assoumani Azali, access to the state was used to support client networks which led to crumbling infrastructure that cultivated in the islands of Nzwani and Mwali declaring independence only to be stopped by French troops. Azali lacked the social obligations required to address the elders and when combined with his gross mismanagement and increasing economic and social dependence on foreign entities, made managing daily life near nonexistent in the state. Therefore, local administrative structures began popping up and drifting away from reliance on the state, funded by remittances from the expatriate community in France.
The Comoros Islands have experienced five different constitutions.
In a separate nod to pressure to restore civilian rule, the government organized several committees to compose a new constitution, including the August 2000 National Congress and November 2000 Tripartite Commission. The opposition parties initially refused to participate in the Tripartite Commission, but on 17 February, representatives of the government, the Anjouan separatists, the political opposition, and civil society organizations signed a "Framework Accord for Reconciliation in Comoros," brokered by the Organization for African Unity
The accord called for the creation of a new Tripartite Commission for National Reconciliation to develop a "New Comorian Entity" with a new constitution. The new federal Constitution came into effect in 2002; it included elements of consociationalism, including a presidency that rotates every four years among the islands and extensive autonomy for each island. Presidential elections were held in 2002, at which Azali Assoumani was elected President. In April 2004 legislative elections were held, completing the implementation of the new constitution.
The new Union of the Comoros consists of three islands, Grande Comore, Anjouan and Mohéli. Each island has a president, who shares the presidency of the Union on a rotating basis. The president and his vice-presidents are elected for a term of four years. The constitution states that, "the islands enjoy financial autonomy, freely draw up and manage their budgets".
President Assoumani Azali of Grande Comore is the first Union president. President Mohamed Bacar of Anjouan formed his 13-member government at the end of April, 2003.
On 15 May 2006, Ahmed Abdallah Sambi, a cleric and successful businessman educated in Iran, Saudi Arabia and Sudan, was declared the winner of elections for President of the Republic. He is considered a moderate Islamist and is called Ayatollah by his supporters. He beat out retired French air force officer Mohamed Djaanfari and long-time politician Ibrahim Halidi, whose candidacy was backed by Azali Assoumani, the outgoing president.
A referendum took place on May 16, 2009 to decide whether to cut down the government's unwieldy political bureaucracy. 52.7% of those eligible voted, and 93.8% of votes were cast in approval of the referendum. The referendum would cause each island's president to become a governor and the ministers to become councilors.
The constitution gives Grande Comore, Anjouan and Mohéli the right to govern most of their own affairs with their own presidents, except the activities assigned to the Union of the Comoros like Foreign Policy, Defense, Nationality, Banking and others. Comoros considers Mayotte, an overseas collectivity of France, to be part of its territory, with an autonomous status
As of 2011, the three autonomous islands are subdivided into 16 prefectures, 54 communes, and 318 villes or villages.
The federal presidency is rotated between the islands' presidents.
The Union of Comoros abolished the position of Prime Minister in 2002. The position of Vice-President of the Comoros was used 2002–2019.
The Assembly of the Union has 33 seats, 24 elected in single seat constituencies and 9 representatives of the regional assemblies.
The Supreme Court or Cour Supreme, has two members appointed by the president, two members elected by the Federal Assembly, one by the Council of each island, and former presidents of the republic.
The Comoros are member of the ACCT, ACP, AfDB, AMF, African Union, FAO, G-77, IBRD, ICAO, ICCt (signatory), ICRM, IDA, IDB, IFAD, IFC, IFRCS, ILO, IMF, InOC, Interpol, IOC, ITU, LAS, NAM, OIC, OPCW (signatory), United Nations, UNCTAD, UNESCO, UNIDO, UPU, WCO, WHO, WMO. | https://en.wikipedia.org/wiki?curid=6003 |
Telecommunications in the Comoros
In large part thanks to international aid programs, Moroni has international telecommunications service. Telephone service, however, is largely limited to the islands' few towns.
Telephones - main lines in use:
5,000 (1995)
Telephones - mobile cellular:
0 (1995)
Telephone system:
sparse system of microwave radio relay and HF radiotelephone communication stations
"domestic:"
HF radiotelephone communications and microwave radio relay
CMDA mobile network (Huri, operated by Comores Telecom)
"international:"
HF radiotelephone communications to Madagascar and Réunion
Radio broadcast stations:
AM 1, FM 2, shortwave 1 (1998)
Radios:
90,000 (1997)
Television broadcast stations:
0 (1998)
Televisions:
1,000 (1997)
Internet Service Providers (ISPs): 1 (1999)
Country code (Top-level domain): KM
In October 2011 the State of Qatar launched a special program for the construction of a wireless network to interconnect the three islands of the archipelago, by means of low cost, repeatable technology. The project has been developed by Qatar University and Politecnico di Torino, under the supervision of prof. Mazen Hasna and prof. Daniele Trinchero, with a major participation of local actors (Comorian Government, NRTIC, University of the Comoros). The project has been referred as an example of technology transfer and Sustainable Inclusion in developing countries | https://en.wikipedia.org/wiki?curid=6005 |
Transport in the Comoros
There are a number of systems of transport in the Comoros. The Comoros possesses of road, of which are paved. It has three seaports: Fomboni, Moroni and Moutsamoudou, but does not have a merchant marine, and no longer has any railway network. It has four airports, all with paved runways, one with runways over long, with the others having runways shorter than .
The isolation of the Comoros had made air traffic a major means of transportation. One of President Abdallah's accomplishments was to make the Comoros more accessible by air. During his administration, he negotiated agreements to initiate or enhance commercial air links with Tanzania and Madagascar. The Djohar regime reached an agreement in 1990 to link Moroni and Brussels by air. By the early 1990s, commercial flights connected the Comoros with France, Mauritius, Kenya, South Africa, Tanzania, and Madagascar. The national airline was Air Comores. Daily flights linked the three main islands, and air service was also available to Mahoré; each island had airstrips. In 1986 the republic received a grant from the French government's CCCE to renovate and expand Hahaya airport, near Moroni. Because of the absence of scheduled sea transport between the islands, nearly all interisland passenger traffic is by air.
More than 99% of freight is transported by sea. Both Moroni on Njazidja and Mutsamudu on Nzwani have artificial harbors. There is also a harbor at Fomboni, on Mwali. Despite extensive internationally financed programs to upgrade the harbors at Moroni and Mutsamudu, by the early 1990s only Mutsamudu was operational as a deepwater facility. Its harbor could accommodate vessels of up to eleven meters' draught. At Moroni, ocean-going vessels typically lie offshore and are loaded or unloaded by smaller craft, a costly and sometimes dangerous procedure. Most freight continues to be sent to Kenya, Reunion, or Madagascar for transshipment to the Comoros. Use of Comoran ports is further restricted by the threat of cyclones from December through March. The privately operated Comoran Navigation Company ("Société Comorienne de Navigation") is based in Moroni, and provides services to Madagascar.
Roads serve the coastal areas, rather than the interior, and the mountainous terrain makes surface travel difficult. | https://en.wikipedia.org/wiki?curid=6006 |
Military of the Comoros
The Comorian Armed Forces () consist of a small standing army and a 500-member police force, as well as a 500-member defense force. A defense treaty with France provides naval resources for protection of territorial waters, training of Comorian military personnel, and air surveillance. France maintains a small troop presence in the Comoros at government request. France maintains a small maritime base and a Foreign Legion Detachment (DLEM) on Mayotte.
NOTE: The last comprehensive aircraft inventory list was from "Aviation Week & Space Technology" in 2007. | https://en.wikipedia.org/wiki?curid=6008 |
Computer worm
A computer worm is a standalone malware computer program that replicates itself in order to spread to other computers. It often uses a computer network to spread itself, relying on security failures on the target computer to access it. It will use this machine as a host to scan and infect other computers. When these new worm-invaded computers are controlled, the worm will continue to scan and infect other computers using these computers as hosts, and this behavior will continue. Computer worms use recursive method to copy themselves without host program and distribute themselves based on the law of exponential growth, and then controlling and infecting more and more computers in a short time. Worms almost always cause at least some harm to the network, even if only by consuming bandwidth, whereas viruses almost always corrupt or modify files on a targeted computer.
Many worms are designed only to spread, and do not attempt to change the systems they pass through. However, as the Morris worm and Mydoom showed, even these "payload-free" worms can cause major disruption by increasing network traffic and other unintended effects.
The actual term "worm" was first used in John Brunner's 1975 novel, "The Shockwave Rider". In that novel, Nichlas Haflinger designs and sets off a data-gathering worm in an act of revenge against the powerful men who run a national electronic information web that induces mass conformity. "You have the biggest-ever worm loose in the net, and it automatically sabotages any attempt to monitor it. There's never been a worm with that tough a head or that long a tail!"
On November 2, 1988, Robert Tappan Morris, a Cornell University computer science graduate student, unleashed what became known as the Morris worm, disrupting many computers then on the Internet, guessed at the time to be one tenth of all those connected. During the Morris appeal process, the U.S. Court of Appeals estimated the cost of removing the worm from each installation at between $200 and $53,000; this work prompted the formation of the CERT Coordination Center and Phage mailing list. Morris himself became the first person tried and convicted under the 1986 Computer Fraud and Abuse Act.
Independence
Computer viruses generally require a host program. The virus writes its own code into the host program. When the program runs, the written virus program is executed first, causing infection and damage. A worm does not need a host program, as it is an independent program or code chunk. Therefore, it is not restricted by the host program, but can run independently and actively carry out attacks.
Exploit attacks
Because a worm is not limited by the host program, worms can take advantage of various operating system vulnerabilities to carry out active attacks. For example, the "Nimda" virus exploits vulnerabilities to attack.
Complexity
Some worms are combined with web page scripts, and are hidden in HTML pages using VBScript, ActiveX and other technologies. When a user accesses a webpage containing a virus, the virus automatically resides in memory and waits to be triggered. There are also some worms that are combined with backdoor programs or Trojan horses, such as "Code Red".
Contagiousness
Worms are more infectious than traditional viruses. They not only infect local computers, but also all servers and clients on the network based on the local computer. Worms can easily spread through shared folders, e-mails, malicious web pages, and servers with a large number of vulnerabilities in the network.
Any code designed to do more than spread the worm is typically referred to as the "payload". Typical malicious payloads might delete files on a host system (e.g., the ExploreZip worm), encrypt files in a ransomware attack, or exfiltrate data such as confidential documents or passwords.
Probably the most common payload for worms is to install a backdoor. This allows the computer to be remotely controlled by the worm author as a "zombie". Networks of such machines are often referred to as botnets and are very commonly used for a range of malicious purposes, including sending spam or performing DoS attacks.
Some special worms attack industrial systems in a targeted manner. Stuxnet does not need to be transmitted through a network connection. This virus can destroy the core production control computer software used by chemical, power generation and power transmission companies in various countries around the world, and instead "issue orders" to other computers in the factory. They can also use multiple vulnerabilities in Windows systems and Siemens SIMATICWinCC systems to attack. Although these systems operate independently from the network, if the operator inserts a virus-infected U disk into the system's USB interface, the virus will be able to gain control of the system without any other operational requirements or prompts.
Worms spread by exploiting vulnerabilities in operating systems.
Vendors with security problems supply regular security updates (see "Patch Tuesday"), and if these are installed to a machine, then the majority of worms are unable to spread to it. If a vulnerability is disclosed before the security patch released by the vendor, a zero-day attack is possible.
Users need to be wary of opening unexpected email, and should not run attached files or programs, or visit web sites that are linked to such emails. However, as with the ILOVEYOU worm, and with the increased growth and efficiency of phishing attacks, it remains possible to trick the end-user into running malicious code.
Anti-virus and anti-spyware software are helpful, but must be kept up-to-date with new pattern files at least every few days. The use of a firewall is also recommended.
Mitigation techniques include:
Infections can sometimes be detected by their behavior - typically scanning the Internet randomly, looking for vulnerable hosts to infect. In addition, machine learning techniques can be used to detect new worms, by analyzing the behavior of the suspected computer.
A helpful worm or anti-worm is a worm designed to do something that its author feels is helpful, though not necessarily with the permission of the executing computer's owner. Beginning with the very first research into worms at Xerox PARC, there have been attempts to create useful worms. Those worms allowed John Shoch and Jon Hupp to test the Ethernet principles on their network of Xerox Alto computers. Similarly, the Nachi family of worms tried to download and install patches from Microsoft's website to fix vulnerabilities in the host system by exploiting those same vulnerabilities. In practice, although this may have made these systems more secure, it generated considerable network traffic, rebooted the machine in the course of patching it, and did its work without the consent of the computer's owner or user. Regardless of their payload or their writers' intentions, most security experts regard all worms as malware.
Several worms, including some XSS worms, have been written to research how worms spread, such as the effects of changes in social activity or user behavior. One study proposed what seems to be the first computer worm that operates on the second layer of the OSI model (Data link Layer), utilizing topology information such as Content-addressable memory (CAM) tables and Spanning Tree information stored in switches to propagate and probe for vulnerable nodes until the enterprise network is covered.
Anti-worms have been used to combat the effects of the Code Red, Blaster, and Santy worms. Welchia is an example of a helpful worm. Utilizing the same deficiencies exploited by the Blaster worm, Welchia infected computers and automatically began downloading Microsoft security updates for Windows without the users' consent. Welchia automatically reboots the computers it infects after installing the updates. One of these updates was the patch that fixed the exploit.
Other examples of helpful worms are "Den_Zuko", "Cheeze", "CodeGreen", and "Millenium". | https://en.wikipedia.org/wiki?curid=6010 |
Chomsky hierarchy
In formal language theory, computer science and linguistics, the Chomsky hierarchy (occasionally referred to as the Chomsky–Schützenberger hierarchy) is a containment hierarchy of classes of formal grammars.
This hierarchy of grammars was described by Noam Chomsky in 1956. It is also named after Marcel-Paul Schützenberger, who played a crucial role in the development of the theory of formal languages.
A formal grammar of this type consists of a finite set of "production rules" ("left-hand side" → "right-hand side"), where each side consists of a finite sequence of the following symbols:
A formal grammar provides an axiom schema for (or "generates") a "formal language", which is a (usually infinite) set of finite-length sequences of symbols that may be constructed by applying production rules to another sequence of symbols (which initially contains just the start symbol). A rule may be applied by replacing an occurrence of the symbols on its left-hand side with those that appear on its right-hand side. A sequence of rule applications is called a "derivation". Such a grammar defines the formal language: all words consisting solely of terminal symbols which can be reached by a derivation from the start symbol.
Nonterminals are often represented by uppercase letters, terminals by lowercase letters, and the start symbol by . For example, the grammar with terminals , nonterminals , production rules
and start symbol , defines the language of all words of the form formula_1 (i.e. copies of followed by copies of ).
The following is a simpler grammar that defines the same language:
Terminals , Nonterminals , Start symbol , Production rules
As another example, a grammar for a toy subset of English language is given by:
and start symbol . An example derivation is
Other sequences that can be derived from this grammar are: ""ideas hate great linguists"", and ""ideas generate"". While these sentences are nonsensical, they are syntactically correct. A syntactically incorrect sentence ( e.g. ""ideas ideas great hate"") cannot be derived from this grammar. See "Colorless green ideas sleep furiously" for a similar example given by Chomsky in 1957; see Phrase structure grammar and Phrase structure rules for more natural language examples and the problems of formal grammar in that area.
The following table summarizes each of Chomsky's four types of grammars, the class of language it generates, the type of automaton that recognizes it, and the form its rules must have.
Note that the set of grammars corresponding to recursive languages is not a member of this hierarchy; these would be properly between Type-0 and Type-1.
Every regular language is context-free, every context-free language is context-sensitive, every context-sensitive language is recursive and every recursive language is recursively enumerable. These are all proper inclusions, meaning that there exist recursively enumerable languages that are not context-sensitive, context-sensitive languages that are not context-free and context-free languages that are not regular.
Type-0 grammars include all formal grammars. They generate exactly all languages that can be recognized by a Turing machine. These languages are also known as the "recursively enumerable" or "Turing-recognizable" languages. Note that this is different from the recursive languages, which can be "decided" by an always-halting Turing machine.
Type-1 grammars generate context-sensitive languages. These grammars have rules of the form formula_2 with formula_3 a nonterminal and formula_4, formula_5 and formula_6 strings of terminals and/or nonterminals. The strings formula_4 and formula_5 may be empty, but formula_6 must be nonempty. The rule formula_10 is allowed if formula_11 does not appear on the right side of any rule. The languages described by these grammars are exactly all languages that can be recognized by a linear bounded automaton (a nondeterministic Turing machine whose tape is bounded by a constant times the length of the input.)
Type-2 grammars generate the context-free languages. These are defined by rules of the form formula_12 with formula_3 being a nonterminal and formula_4 being a string of terminals and/or nonterminals. These languages are exactly all languages that can be recognized by a non-deterministic pushdown automaton. Context-free languages—or rather its subset of deterministic context-free language—are the theoretical basis for the phrase structure of most programming languages, though their syntax also includes context-sensitive name resolution due to declarations and scope. Often a subset of grammars is used to make parsing easier, such as by an LL parser.
Type-3 grammars generate the regular languages. Such a grammar restricts its rules to a single nonterminal on the left-hand side and a right-hand side consisting of a single terminal, possibly followed by a single nonterminal (right regular). Alternatively, the right-hand side of the grammar can consist of a single terminal, possibly preceded by a single nonterminal (left regular). These generate the same languages. However, if left-regular rules and right-regular rules are combined, the language need no longer be regular. The rule formula_10 is also allowed here if formula_11 does not appear on the right side of any rule. These languages are exactly all languages that can be decided by a finite state automaton. Additionally, this family of formal languages can be obtained by regular expressions. Regular languages are commonly used to define search patterns and the lexical structure of programming languages. | https://en.wikipedia.org/wiki?curid=6011 |
Cathode-ray tube
The cathode-ray tube (CRT) is a vacuum tube that contains one or more electron guns and a phosphorescent screen and is used to display images. It modulates, accelerates, and deflects electron beam(s) onto the screen to create the images. The images may represent electrical waveforms (oscilloscope), pictures (television, computer monitor), radar targets, or other phenomena. CRTs have also been used as memory devices, in which case the visible light emitted from the fluorescent material (if any) is not intended to have significant meaning to a visual observer (though the visible pattern on the tube face may cryptically represent the stored data).
In television sets and computer monitors, the entire front area of the tube is scanned repetitively and systematically in a fixed pattern called a raster. In color devices, an image is produced by controlling the intensity of each of the three electron beams, one for each additive primary color (red, green, and blue) with a video signal as a reference. In all modern CRT monitors and televisions, the beams are bent by "magnetic deflection", a varying magnetic field generated by coils and driven by electronic circuits around the neck of the tube, although electrostatic deflection is commonly used in oscilloscopes, a type of electronic test instrument.
A CRT is constructed from a glass envelope which is large, deep (i.e., long from front screen face to rear end), fairly heavy, and relatively fragile. The interior of a CRT is evacuated to approximately to , evacuation being necessary to facilitate the free flight of electrons from the gun(s) to the tube's face. The fact that it is evacuated makes handling an intact CRT potentially dangerous due to the risk of breaking the tube and causing a violent implosion that can hurl shards of glass at great velocity. As a matter of safety, the face is typically made of thick lead glass so as to be highly shatter-resistant and to block most X-ray emissions, particularly if the CRT is used in a consumer product.
Since the late 2000s, CRTs have been largely superseded by newer "flat panel" display technologies such as LCD, plasma display, and OLED displays, which have lower manufacturing costs and power consumption, as well as significantly less weight and bulk. Flat-panel displays can also be made in very large sizes; whereas was about the largest size of a CRT television, flat panels are available in and even larger sizes.
Cathode rays were discovered by Julius Plücker and Johann Wilhelm Hittorf. Hittorf observed that some unknown rays were emitted from the cathode (negative electrode) which could cast shadows on the glowing wall of the tube, indicating the rays were traveling in straight lines. In 1890, Arthur Schuster demonstrated cathode rays could be deflected by electric fields, and William Crookes showed they could be deflected by magnetic fields. In 1897, J. J. Thomson succeeded in measuring the charge-mass-ratio of cathode rays, showing that they consisted of negatively charged particles smaller than atoms, the first "subatomic particles", which had already been named "electrons" by Irish physicist, George Johnstone Stoney in 1891. The earliest version of the CRT was known as the "Braun tube", invented by the German physicist Ferdinand Braun in 1897. It was a cold-cathode diode, a modification of the Crookes tube with a phosphor-coated screen.
The first cathode-ray tube to use a hot cathode was developed by John Bertrand Johnson (who gave his name to the term Johnson noise) and Harry Weiner Weinhart of Western Electric, and became a commercial product in 1922.
In 1926, Kenjiro Takayanagi demonstrated a CRT television that received images with a 40-line resolution. By 1927, he improved the resolution to 100 lines, which was unrivaled until 1931. By 1928, he was the first to transmit human faces in half-tones on a CRT display. By 1935, he had invented an early all-electronic CRT television.
It was named in 1929 by inventor Vladimir K. Zworykin, who was influenced by Takayanagi's earlier work. RCA was granted a trademark for the term (for its cathode-ray tube) in 1932; it voluntarily released the term to the public domain in 1950.
The first commercially made electronic television sets with cathode-ray tubes were manufactured by Telefunken in Germany in 1934.
Flat panel displays dropped in price and started significantly displacing cathode-ray tubes in the 2000s, with LCD screens exceeding CRTs in 2008.
The last known manufacturer of (in this case, recycled) CRTs, Videocon, ceased in 2015.
In oscilloscope CRTs, electrostatic deflection is used, rather than the magnetic deflection commonly used with television and other large CRTs. The beam is deflected horizontally by applying an electric field between a pair of plates to its left and right, and vertically by applying an electric field to plates above and below. Televisions use magnetic rather than electrostatic deflection because the deflection plates obstruct the beam when the deflection angle is as large as is required for tubes that are relatively short for their size.
Various phosphors are available depending upon the needs of the measurement or display application. The brightness, color, and persistence of the illumination depends upon the type of phosphor used on the CRT screen. Phosphors are available with persistences ranging from less than one microsecond to several seconds. For visual observation of brief transient events, a long persistence phosphor may be desirable. For events which are fast and repetitive, or high frequency, a short-persistence phosphor is generally preferable.
When displaying fast one-shot events, the electron beam must deflect very quickly, with few electrons impinging on the screen, leading to a faint or invisible image on the display. Oscilloscope CRTs designed for very fast signals can give a brighter display by passing the electron beam through a micro-channel plate just before it reaches the screen. Through the phenomenon of secondary emission, this plate multiplies the number of electrons reaching the phosphor screen, giving a significant improvement in writing rate (brightness) and improved sensitivity and spot size as well.
Most oscilloscopes have a graticule as part of the visual display, to facilitate measurements. The graticule may be permanently marked inside the face of the CRT, or it may be a transparent external plate made of glass or acrylic plastic. An internal graticule eliminates parallax error, but cannot be changed to accommodate different types of measurements. Oscilloscopes commonly provide a means for the graticule to be illuminated from the side, which improves its visibility.
These are found in "analog phosphor storage oscilloscopes". These are distinct from "digital storage oscilloscopes" which rely on solid state digital memory to store the image.
Where a single brief event is monitored by an oscilloscope, such an event will be displayed by a conventional tube only while it actually occurs. The use of a long persistence phosphor may allow the image to be observed after the event, but only for a few seconds at best. This limitation can be overcome by the use of a direct view storage cathode-ray tube (storage tube). A storage tube will continue to display the event after it has occurred until such time as it is erased. A storage tube is similar to a conventional tube except that it is equipped with a metal grid coated with a dielectric layer located immediately behind the phosphor screen. An externally applied voltage to the mesh initially ensures that the whole mesh is at a constant potential. This mesh is constantly exposed to a low velocity electron beam from a 'flood gun' which operates independently of the main gun. This flood gun is not deflected like the main gun but constantly 'illuminates' the whole of the storage mesh. The initial charge on the storage mesh is such as to repel the electrons from the flood gun which are prevented from striking the phosphor screen.
When the main electron gun writes an image to the screen, the energy in the main beam is sufficient to create a 'potential relief' on the storage mesh. The areas where this relief is created no longer repel the electrons from the flood gun which now pass through the mesh and illuminate the phosphor screen. Consequently, the image that was briefly traced out by the main gun continues to be displayed after it has occurred. The image can be 'erased' by resupplying the external voltage to the mesh restoring its constant potential. The time for which the image can be displayed was limited because, in practice, the flood gun slowly neutralises the charge on the storage mesh. One way of allowing the image to be retained for longer is temporarily to turn off the flood gun. It is then possible for the image to be retained for several days. The majority of storage tubes allow for a lower voltage to be applied to the storage mesh which slowly restores the initial charge state. By varying this voltage a variable persistence is obtained. Turning off the flood gun and the voltage supply to the storage mesh allows such a tube to operate as a conventional oscilloscope tube.
Color tubes use three different phosphors which emit red, green, and blue light respectively. They are packed together in stripes (as in aperture grille designs) or clusters called "triads" (as in shadow mask CRTs). Color CRTs have three electron guns, one for each primary color, arranged either in a straight line or in an equilateral triangular configuration (the guns are usually constructed as a single unit). (The triangular configuration is often called "delta-gun", based on its relation to the shape of the Greek letter delta Δ.) A grille or mask absorbs the electrons that would otherwise hit the wrong phosphor. A shadow mask tube uses a metal plate with tiny holes, placed so that the electron beam only illuminates the correct phosphors on the face of the tube; the holes are tapered so that the electrons that strike the inside of any hole will be reflected back, if they are not absorbed (e.g. due to local charge accumulation), instead of bouncing through the hole to strike a random (wrong) spot on the screen. Another type of color CRT uses an aperture grille of tensioned vertical wires to achieve the same result.
Due to limitations in the dimensional precision with which CRTs can be manufactured economically, it has not been practically possible to build color CRTs in which three electron beams could be aligned to hit phosphors of respective color in acceptable coordination, solely on the basis of the geometric configuration of the electron gun axes and gun aperture positions, shadow mask apertures, etc. The shadow mask ensures that one beam will only hit spots of certain colors of phosphors, but minute variations in physical alignment of the internal parts among individual CRTs will cause variations in the exact alignment of the beams through the shadow mask, allowing some electrons from, for example, the red beam to hit, say, blue phosphors, unless some individual compensation is made for the variance among individual tubes.
Color convergence and color purity are two aspects of this single problem. Firstly, for correct color rendering it is necessary that regardless of where the beams are deflected on the screen, all three hit the same spot (and nominally pass through the same hole or slot) on the shadow mask. This is called convergence. More specifically, the convergence at the center of the screen (with no deflection field applied by the yoke) is called static convergence, and the convergence over the rest of the screen area is called dynamic convergence. The beams may converge at the center of the screen and yet stray from each other as they are deflected toward the edges; such a CRT would be said to have good static convergence but poor dynamic convergence. Secondly, each beam must only strike the phosphors of the color it is intended to strike and no others. This is called purity. Like convergence, there is static purity and dynamic purity, with the same meanings of "static" and "dynamic" as for convergence. Convergence and purity are distinct parameters; a CRT could have good purity but poor convergence, or vice versa. Poor convergence causes color "shadows" or "ghosts" along displayed edges and contours, as if the image on the screen were intaglio printed with poor registration. Poor purity causes objects on the screen to appear off-color while their edges remain sharp. Purity and convergence problems can occur at the same time, in the same or different areas of the screen or both over the whole screen, and either uniformly or to greater or lesser degrees over different parts of the screen.
The solution to the static convergence and purity problems is a set of color alignment magnets installed around the neck of the CRT. These movable weak permanent magnets are usually mounted on the back end of the deflection yoke assembly and are set at the factory to compensate for any static purity and convergence errors that are intrinsic to the unadjusted tube. Typically there are two or three pairs of two magnets in the form of rings made of plastic impregnated with a magnetic material, with their magnetic fields parallel to the planes of the magnets, which are perpendicular to the electron gun axes. Each pair of magnetic rings forms a single effective magnet whose field vector can be fully and freely adjusted (in both direction and magnitude). By rotating a pair of magnets relative to each other, their relative field alignment can be varied, adjusting the effective field strength of the pair. (As they rotate relative to each other, each magnet's field can be considered to have two opposing components at right angles, and these four components [two each for two magnets] form two pairs, one pair reinforcing each other and the other pair opposing and canceling each other. Rotating away from alignment, the magnets' mutually reinforcing field components decrease as they are traded for increasing opposed, mutually cancelling components.) By rotating a pair of magnets together, preserving the relative angle between them, the direction of their collective magnetic field can be varied. Overall, adjusting all of the convergence/purity magnets allows a finely tuned slight electron beam deflection or lateral offset to be applied, which compensates for minor static convergence and purity errors intrinsic to the uncalibrated tube. Once set, these magnets are usually glued in place, but normally they can be freed and readjusted in the field (e.g. by a TV repair shop) if necessary.
On some CRTs, additional fixed adjustable magnets are added for dynamic convergence or dynamic purity at specific points on the screen, typically near the corners or edges. Further adjustment of dynamic convergence and purity typically cannot be done passively, but requires active compensation circuits.
Dynamic color convergence and purity are one of the main reasons why until late in their history, CRTs were long-necked (deep) and had biaxially curved faces; these geometric design characteristics are necessary for intrinsic passive dynamic color convergence and purity. Only starting around the 1990s did sophisticated active dynamic convergence compensation circuits become available that made short-necked and flat-faced CRTs workable. These active compensation circuits use the deflection yoke to finely adjust beam deflection according to the beam target location. The same techniques (and major circuit components) also make possible the adjustment of display image rotation, skew, and other complex raster geometry parameters through electronics under user control.
If the shadow mask or aperture grille becomes magnetized, its magnetic field alters the paths of the electron beams. This causes errors of "color purity" as the electrons no longer follow only their intended paths, and some will hit some phosphors of colors other than the one intended. For example, some electrons from the red beam may hit blue or green phosphors, imposing a magenta or yellow tint to parts of the image that are supposed to be pure red. (This effect is localized to a specific area of the screen if the magnetization is localized.) Therefore, it is important that the shadow mask or aperture grille not be magnetized.
Most color CRT displays, i.e. television sets and computer monitors, each have a built-in degaussing (demagnetizing) circuit, the primary component of which is a degaussing coil which is mounted around the perimeter of the CRT face inside the bezel. Upon power-up of the CRT display, the degaussing circuit produces a brief, alternating current through the degaussing coil which smoothly decays in strength (fades out) to zero over a period of a few seconds, producing a decaying alternating magnetic field from the coil. This degaussing field is strong enough to remove shadow mask magnetization in most cases. In unusual cases of strong magnetization where the internal degaussing field is not sufficient, the shadow mask may be degaussed externally with a stronger portable degausser or demagnetizer. However, an excessively strong magnetic field, whether alternating or constant, may mechanically deform (bend) the shadow mask, causing a permanent color distortion on the display which looks very similar to a magnetization effect.
The degaussing circuit is often built of a thermo-electric (not electronic) device containing a small ceramic heating element and a positive thermal coefficient (PTC) resistor, connected directly to the switched AC power line with the resistor in series with the degaussing coil. When the power is switched on, the heating element heats the PTC resistor, increasing its resistance to a point where degaussing current is minimal, but not actually zero. In older CRT displays, this low-level current (which produces no significant degaussing field) is sustained along with the action of the heating element as long as the display remains switched on. To repeat a degaussing cycle, the CRT display must be switched off and left off for at least several seconds to reset the degaussing circuit by allowing the PTC resistor to cool to the ambient temperature; switching the display-off and immediately back on will result in a weak degaussing cycle or effectively no degaussing cycle.
This simple design is effective and cheap to build, but it wastes some power continuously. Later models, especially Energy Star rated ones, use a relay to switch the entire degaussing circuit on and off, so that the degaussing circuit uses energy only when it is functionally active and needed. The relay design also enables degaussing on user demand through the unit's front panel controls, without switching the unit off and on again. This relay can often be heard clicking off at the end of the degaussing cycle a few seconds after the monitor is turned on, and on and off during a manually initiated degaussing cycle.
At high refresh rates and resolutions, the deflection coil/yoke starts to produce large amounts of heat, due to the need to quickly move the electron beam (because the electron beam has to scan more lines per second), which in turn requires large amounts of power, for quickly generating strong magnetic fields. This makes CRTs beyond certain resolutions and refresh rates impractical, since the coils would need active cooling, to prevent the heat from the coils from melting the glue that is used to attach them to the neck of the CRT.
Vector monitors were used in early computer aided design systems and are in some late-1970s to mid-1980s arcade games such as "Asteroids".
They draw graphics point-to-point, rather than scanning a raster. Either monochrome or color CRTs can be used in vector displays, and the essential principles of CRT design and operation are the same for either type of display; the main difference is in the beam deflection patterns and circuits.
Dot pitch defines the maximum resolution of the display, assuming delta-gun CRTs. In these, as the scanned resolution approaches the dot pitch resolution, moiré appears, as the detail being displayed is finer than what the shadow mask can render. Aperture grille monitors do not suffer from vertical moiré; however, because their phosphor stripes have no vertical detail. In smaller CRTs, these strips maintain position by themselves, but larger aperture-grille CRTs require one or two crosswise (horizontal) support strips.
CRTs have a pronounced triode characteristic, which results in significant gamma (a nonlinear relationship in an electron gun between applied video voltage and beam intensity).
The Williams tube or Williams-Kilburn tube was a cathode-ray tube used to electronically store binary data. It was used in computers of the 1940s as a random-access digital storage device. In contrast to other CRTs in this article, the Williams tube was not a display device, and in fact could not be viewed since a metal plate covered its screen.
In some vacuum tube radio sets, a "Magic Eye" or "Tuning Eye" tube was provided to assist in tuning the receiver. Tuning would be adjusted until the width of a radial shadow was minimized. This was used instead of a more expensive electromechanical meter, which later came to be used on higher-end tuners when transistor sets lacked the high voltage required to drive the device. The same type of device was used with tape recorders as a recording level meter, and for various other applications including electrical test equipment.
Some displays for early computers (those that needed to display more text than was practical using vectors, or that required high speed for photographic output) used Charactron CRTs. These incorporate a perforated metal character mask (stencil), which shapes a wide electron beam to form a character on the screen. The system selects a character on the mask using one set of deflection circuits, but that causes the extruded beam to be aimed off-axis, so a second set of deflection plates has to re-aim the beam so it is headed toward the center of the screen. A third set of plates places the character wherever required. The beam is unblanked (turned on) briefly to draw the character at that position. Graphics could be drawn by selecting the position on the mask corresponding to the code for a space (in practice, they were simply not drawn), which had a small round hole in the center; this effectively disabled the character mask, and the system reverted to regular vector behavior. Charactrons had exceptionally long necks, because of the need for three deflection systems.
Nimo was the trademark of a family of small specialised CRTs manufactured by Industrial Electronics Engineers. These had 10 electron guns which produced electron beams in the form of digits in a manner similar to that of the charactron. The tubes were either simple single-digit displays or more complex 4- or 6- digit displays produced by means of a suitable magnetic deflection system. Having little of the complexities of a standard CRT, the tube required a relatively simple driving circuit, and as the image was projected on the glass face, it provided a much wider viewing angle than competitive types (e.g., nixie tubes).
Flood beam CRTs are small tubes that are arranged as pixels for large screens like Jumbotrons. The first screen using this technology was introduced by Mitsubishi Electric for the 1980 Major League Baseball All-Star Game. It differs from a normal CRT in that the electron gun within does not produce a focused controllable beam. Instead, electrons are sprayed in a wide cone across the entire front of the phosphor screen, basically making each unit act as a single light bulb. Each one is coated with a red, green or blue phosphor, to make up the color sub-pixels. This technology has largely been replaced with light emitting diode displays. Unfocused and undeflected CRTs were used as grid-controlled stroboscope lamps since 1958.
CRTs with an unphosphored front glass but with fine wires embedded in it were used as electrostatic print heads in the 1960s. The wires would pass the electron beam current through the glass onto a sheet of paper where the desired content was therefore deposited as an electrical charge pattern. The paper was then passed near a pool of liquid ink with the opposite charge. The charged areas of the paper attract the ink and thus form the image.
In the late 1990s and early 2000s Philips Research Laboratories experimented with a type of thin CRT known as the "Zeus" display which contained CRT-like functionality in a flat panel display. The devices were demonstrated but never marketed.
Some CRT manufacturers, both LG Display and Samsung Display, innovated CRT technology by creating a slimmer tube. Slimmer CRT has a trade name Superslim and Ultraslim. A flat CRT has a depth. The depth of Superslim was and Ultraslim was .
Despite being a mainstay of display technology for decades, CRT-based computer monitors and televisions are now virtually a dead technology. Demand for CRT screens dropped in the late 2000s. The rapid advances and falling prices of LCD flat panel technology — first for computer monitors, and then for televisions — spelled doom for competing display technologies such as CRT, rear-projection, and plasma display.
Most high-end CRT production had ceased by around 2010, including high-end Sony and Panasonic product lines. In Canada and the United States, the sale and production of high-end CRT TVs ( screens) in these markets had all but ended by 2007. Just a couple of years later, inexpensive "combo" CRT TVs ( screens with an integrated VHS player) disappeared from discount stores.
Electronics retailers such as Best Buy steadily reduced store spaces for CRTs. In 2005, Sony announced that they would stop the production of CRT computer displays. Samsung did not introduce any CRT models for the 2008 model year at the 2008 Consumer Electronics Show; on 4 February 2008, they removed their 30" wide screen CRTs from their North American website and did not replace them with new models.
In the United Kingdom, DSG (Dixons), the largest retailer of domestic electronic equipment, reported that CRT models made up 80–90% of the volume of televisions sold at Christmas 2004 and 15–20% a year later, and that they were expected to be less than 5% at the end of 2006. Dixons ceased selling CRT televisions in 2006.
However, the demise of CRTs has been happening more slowly in the developing world. According to iSupply, production in units of CRTs was not surpassed by LCDs production until 4Q 2007, owing largely to CRT production at factories in China.
CRTs, despite advances, remained relatively heavy and cumbersome in comparison to other display technologies. CRT screens had much deeper cabinets compared to flat panels and rear-projection displays for a given screen size; it was largely impractical to build or use CRTs larger than . The CRT disadvantages became especially significant in light of rapid technological advancements in LCD and plasma flat-panels, which allowed them to easily surpass these size limitations while also being thin and wall-mountable (two key features that were increasingly being preferred by customers).
CRTs can emit a small amount of X-ray radiation as a result of the electron beam's bombardment of the shadow mask/aperture grille and phosphors. The amount of radiation escaping the front of the monitor is widely considered not to be harmful. The Food and Drug Administration regulations in are used to strictly limit, for instance, television receivers to 0.5 milliroentgens per hour (mR/h) (0.13 µC/(kg·h) or 36 pA/kg) at a distance of from any external surface; since 2007, most CRTs have emissions that fall well below this limit.
Older color and monochrome CRTs may have been manufactured with toxic substances, such as cadmium, in the phosphors. The rear glass tube of modern CRTs may be made from leaded glass, which represent an environmental hazard if disposed of improperly. By the time personal computers were produced, glass in the front panel (the viewable portion of the CRT) used barium rather than lead, though the rear of the CRT was still produced from leaded glass. Monochrome CRTs typically do not contain enough leaded glass to fail EPA TCLP tests. While the TCLP process grinds the glass into fine particles in order to expose them to weak acids to test for leachate, intact CRT glass does not leach (The lead is vitrified, contained inside the glass itself, similar to leaded glass crystalware).
Due to the toxins contained in CRT monitors the United States Environmental Protection Agency created rules (in October 2001) stating that CRTs must be brought to special e-waste recycling facilities. In November 2002, the EPA began fining companies that disposed of CRTs through landfills or incineration. Regulatory agencies, local and statewide, monitor the disposal of CRTs and other computer equipment.
Various states participate in the recycling of CRTs, each with their reporting requirements for collectors and recycling facilities. For example, in California the recycling of CRTs is governed by CALRecycle, the California Department of Resources Recycling and Recovery through their Payment System.. Recycling facilities that accept CRT devices from business and residential sector must obtain contact information such as address and phone number to ensure the CRTs come from a California source in order to participate in the CRT Recycling Payment System.
In Europe, disposal of CRT televisions and monitors is covered by the WEEE Directive.
At low refresh rates (60 Hz and below), the periodic scanning of the display may produce a flicker that some people perceive more easily than others, especially when viewed with peripheral vision. Flicker is commonly associated with CRT as most televisions run at 50 Hz (PAL) or 60 Hz (NTSC), although there are some 100 Hz PAL televisions that are flicker-free. Typically only low-end monitors run at such low frequencies, with most computer monitors supporting at least 75 Hz and high-end monitors capable of 100 Hz or more to eliminate any perception of flicker. Though the 100 Hz PAL was often achieved using interleaved scanning, dividing the circuit and scan into two beams of 50 Hz. Non-computer CRTs or CRT for sonar or radar may have long persistence phosphor and are thus flicker free. If the persistence is too long on a video display, moving images will be blurred.
50 Hz/60 Hz CRTs used for television operate with horizontal scanning frequencies of 15,734 Hz (for NTSC systems) or 15,625 Hz (for PAL systems). These frequencies are at the upper range of human hearing and are inaudible to many people; however, some people (especially children) will perceive a high-pitched tone near an operating television CRT. The sound is due to magnetostriction in the magnetic core and periodic movement of windings of the flyback transformer.
This problem does not occur on 100/120 Hz TVs and on non-CGA (Color Graphics Adapter) computer displays, because they use much higher horizontal scanning frequencies that produce sound which is inaudible to humans (22 kHz to over 100 kHz).
High vacuum inside glass-walled cathode-ray tubes permits electron beams to fly freely—without colliding into molecules of air or other gas. If the glass is damaged, atmospheric pressure can collapse the vacuum tube into dangerous fragments which accelerate inward and then spray at high speed in all directions. The implosion energy is proportional to the evacuated volume of the CRT. Although modern cathode-ray tubes used in televisions and computer displays have epoxy-bonded face-plates or other measures to prevent shattering of the envelope, CRTs must be handled carefully to avoid personal injury.
To accelerate the electrons from the cathode to the screen with sufficient velocity, a very high voltage (EHT or extra-high tension) is required, from a few thousand volts for a small oscilloscope CRT to tens of kV for a larger screen color TV. This is many times greater than household power supply voltage. Even after the power supply is turned off, some associated capacitors and the CRT itself may retain a charge for some time and therefore dissipate that charge suddenly through a ground such as an inattentive human grounding a capacitor discharge lead.
Under some circumstances, the signal radiated from the electron guns, scanning circuitry, and associated wiring of a CRT can be captured remotely and used to reconstruct what is shown on the CRT using a process called Van Eck phreaking. Special TEMPEST shielding can mitigate this effect. Such radiation of a potentially exploitable signal, however, occurs also with other display technologies and with electronics in general.
As electronic waste, CRTs are considered one of the hardest types to recycle. CRTs have relatively high concentration of lead and phosphors (not phosphorus), both of which are necessary for the display. There are several companies in the United States that charge a small fee to collect CRTs, then subsidize their labor by selling the harvested copper, wire, and printed circuit boards. The United States Environmental Protection Agency (EPA) includes discarded CRT monitors in its category of "hazardous household waste" but considers CRTs that have been set aside for testing to be commodities if they are not discarded, speculatively accumulated, or left unprotected from weather and other damage.
Leaded CRT glass was sold to be remelted into other CRTs, or even broken down and used in road construction.
Basics of cathode rays and discharge in low-pressure gas:
Light production by cathode rays:
Manipulating the electron beam:
Applying CRT in different display-purpose:
Miscellaneous phenomena:
Historical aspects:
Safety and precautions: | https://en.wikipedia.org/wiki?curid=6014 |
Crystal
A crystal or crystalline solid is a solid material whose constituents (such as atoms, molecules, or ions) are arranged in a highly ordered microscopic structure, forming a crystal lattice that extends in all directions. In addition, macroscopic single crystals are usually identifiable by their geometrical shape, consisting of flat faces with specific, characteristic orientations. The scientific study of crystals and crystal formation is known as crystallography. The process of crystal formation via mechanisms of crystal growth is called crystallization or solidification.
The word "crystal" derives from the Ancient Greek word (), meaning both "ice" and "rock crystal", from (), "icy cold, frost".
Examples of large crystals include snowflakes, diamonds, and table salt. Most inorganic solids are not crystals but polycrystals, i.e. many microscopic crystals fused together into a single solid. Examples of polycrystals include most metals, rocks, ceramics, and ice. A third category of solids is amorphous solids, where the atoms have no periodic structure whatsoever. Examples of amorphous solids include glass, wax, and many plastics.
Despite the name, lead crystal, crystal glass, and related products are "not" crystals, but rather types of glass, i.e. amorphous solids.
Crystals are often used in pseudoscientific practices such as crystal therapy, and, along with gemstones, are sometimes associated with spellwork in Wiccan beliefs and related religious movements.
The scientific definition of a "crystal" is based on the microscopic arrangement of atoms inside it, called the crystal structure. A crystal is a solid where the atoms form a periodic arrangement. (Quasicrystals are an exception, see below).
Not all solids are crystals. For example, when liquid water starts freezing, the phase change begins with small ice crystals that grow until they fuse, forming a "polycrystalline" structure. In the final block of ice, each of the small crystals (called "crystallites" or "grains") is a true crystal with a periodic arrangement of atoms, but the whole polycrystal does "not" have a periodic arrangement of atoms, because the periodic pattern is broken at the grain boundaries. Most macroscopic inorganic solids are polycrystalline, including almost all metals, ceramics, ice, rocks, etc. Solids that are neither crystalline nor polycrystalline, such as glass, are called "amorphous solids", also called glassy, vitreous, or noncrystalline. These have no periodic order, even microscopically. There are distinct differences between crystalline solids and amorphous solids: most notably, the process of forming a glass does not release the latent heat of fusion, but forming a crystal does.
A crystal structure (an arrangement of atoms in a crystal) is characterized by its "unit cell", a small imaginary box containing one or more atoms in a specific spatial arrangement. The unit cells are stacked in three-dimensional space to form the crystal.
The symmetry of a crystal is constrained by the requirement that the unit cells stack perfectly with no gaps. There are 219 possible crystal symmetries, called crystallographic space groups. These are grouped into 7 crystal systems, such as cubic crystal system (where the crystals may form cubes or rectangular boxes, such as Halite (mineral) shown at right) or hexagonal crystal system (where the crystals may form hexagons, such as ordinary water ice).
Crystals are commonly recognized by their shape, consisting of flat faces with sharp angles. These shape characteristics are not "necessary" for a crystal—a crystal is scientifically defined by its microscopic atomic arrangement, not its macroscopic shape—but the characteristic macroscopic shape is often present and easy to see.
Euhedral crystals are those with obvious, well-formed flat faces. Anhedral crystals do not, usually because the crystal is one grain in a polycrystalline solid.
The flat faces (also called facets) of a euhedral crystal are oriented in a specific way relative to the underlying atomic arrangement of the crystal: they are planes of relatively low Miller index. This occurs because some surface orientations are more stable than others (lower surface energy). As a crystal grows, new atoms attach easily to the rougher and less stable parts of the surface, but less easily to the flat, stable surfaces. Therefore, the flat surfaces tend to grow larger and smoother, until the whole crystal surface consists of these plane surfaces. (See diagram on right.)
One of the oldest techniques in the science of crystallography consists of measuring the three-dimensional orientations of the faces of a crystal, and using them to infer the underlying crystal symmetry.
A crystal's habit is its visible external shape. This is determined by the crystal structure (which restricts the possible facet orientations), the specific crystal chemistry and bonding (which may favor some facet types over others), and the conditions under which the crystal formed.
By volume and weight, the largest concentrations of crystals in the Earth are part of its solid bedrock. Crystals found in rocks typically range in size from a fraction of a millimetre to several centimetres across, although exceptionally large crystals are occasionally found. , the world's largest known naturally occurring crystal is a crystal of beryl from Malakialina, Madagascar, long and in diameter, and weighing .
Some crystals have formed by magmatic and metamorphic processes, giving origin to large masses of crystalline rock. The vast majority of igneous rocks are formed from molten magma and the degree of crystallization depends primarily on the conditions under which they solidified. Such rocks as granite, which have cooled very slowly and under great pressures, have completely crystallized; but many kinds of lava were poured out at the surface and cooled very rapidly, and in this latter group a small amount of amorphous or glassy matter is common. Other crystalline rocks, the metamorphic rocks such as marbles, mica-schists and quartzites, are recrystallized. This means that they were at first fragmental rocks like limestone, shale and sandstone and have never been in a molten condition nor entirely in solution, but the high temperature and pressure conditions of metamorphism have acted on them by erasing their original structures and inducing recrystallization in the solid state.
Other rock crystals have formed out of precipitation from fluids, commonly water, to form druses or quartz veins.
Evaporites such as Halite (mineral), gypsum and some limestones have been deposited from aqueous solution, mostly owing to evaporation in arid climates.
Water-based ice in the form of snow, sea ice, and glaciers are common crystalline/polycrystalline structures on Earth and other planets. A single snowflake is a single crystal or a collection of crystals, while an ice cube is a polycrystal.
Many living organisms are able to produce crystals, for example calcite and aragonite in the case of most molluscs or hydroxylapatite in the case of vertebrates.
The same group of atoms can often solidify in many different ways. Polymorphism is the ability of a solid to exist in more than one crystal form. For example, water ice is ordinarily found in the hexagonal form Ice Ih, but can also exist as the cubic Ice Ic, the rhombohedral ice II, and many other forms. The different polymorphs are usually called different "phases".
In addition, the same atoms may be able to form noncrystalline phases. For example, water can also form amorphous ice, while SiO2 can form both fused silica (an amorphous glass) and quartz (a crystal). Likewise, if a substance can form crystals, it can also form polycrystals.
For pure chemical elements, polymorphism is known as allotropy. For example, diamond and graphite are two crystalline forms of carbon, while amorphous carbon is a noncrystalline form. Polymorphs, despite having the same atoms, may have wildly different properties. For example, diamond is among the hardest substances known, while graphite is so soft that it is used as a lubricant.
Polyamorphism is a similar phenomenon where the same atoms can exist in more than one amorphous solid form.
Crystallization is the process of forming a crystalline structure from a fluid or from materials dissolved in a fluid. (More rarely, crystals may be deposited directly from gas; see thin-film deposition and epitaxy.)
Crystallization is a complex and extensively-studied field, because depending on the conditions, a single fluid can solidify into many different possible forms. It can form a single crystal, perhaps with various possible phases, stoichiometries, impurities, defects, and habits. Or, it can form a polycrystal, with various possibilities for the size, arrangement, orientation, and phase of its grains. The final form of the solid is determined by the conditions under which the fluid is being solidified, such as the chemistry of the fluid, the ambient pressure, the temperature, and the speed with which all these parameters are changing.
Specific industrial techniques to produce large single crystals (called "boules") include the Czochralski process and the Bridgman technique. Other less exotic methods of crystallization may be used, depending on the physical properties of the substance, including hydrothermal synthesis, sublimation, or simply solvent-based crystallization.
Large single crystals can be created by geological processes. For example, selenite crystals in excess of 10 meters are found in the Cave of the Crystals in Naica, Mexico. For more details on geological crystal formation, see above.
Crystals can also be formed by biological processes, see above. Conversely, some organisms have special techniques to "prevent" crystallization from occurring, such as antifreeze proteins.
An "ideal" crystal has every atom in a perfect, exactly repeating pattern. However, in reality, most crystalline materials have a variety of crystallographic defects, places where the crystal's pattern is interrupted. The types and structures of these defects may have a profound effect on the properties of the materials.
A few examples of crystallographic defects include vacancy defects (an empty space where an atom should fit), interstitial defects (an extra atom squeezed in where it does not fit), and dislocations (see figure at right). Dislocations are especially important in materials science, because they help determine the mechanical strength of materials.
Another common type of crystallographic defect is an impurity, meaning that the "wrong" type of atom is present in a crystal. For example, a perfect crystal of diamond would only contain carbon atoms, but a real crystal might perhaps contain a few boron atoms as well. These boron impurities change the diamond's color to slightly blue. Likewise, the only difference between ruby and sapphire is the type of impurities present in a corundum crystal.
In semiconductors, a special type of impurity, called a dopant, drastically changes the crystal's electrical properties. Semiconductor devices, such as transistors, are made possible largely by putting different semiconductor dopants into different places, in specific patterns.
Twinning is a phenomenon somewhere between a crystallographic defect and a grain boundary. Like a grain boundary, a twin boundary has different crystal orientations on its two sides. But unlike a grain boundary, the orientations are not random, but related in a specific, mirror-image way.
Mosaicity is a spread of crystal plane orientations. A mosaic crystal is supposed to consist of smaller crystalline units that are somewhat misaligned with respect to each other.
In general, solids can be held together by various types of chemical bonds, such as metallic bonds, ionic bonds, covalent bonds, van der Waals bonds, and others. None of these are necessarily crystalline or non-crystalline. However, there are some general trends as follows.
Metals are almost always polycrystalline, though there are exceptions like amorphous metal and single-crystal metals. The latter are grown synthetically. (A microscopically-small piece of metal may naturally form into a single crystal, but larger pieces generally do not.) Ionic compound materials are usually crystalline or polycrystalline. In practice, large salt crystals can be created by solidification of a molten fluid, or by crystallization out of a solution. Covalently bonded solids (sometimes called covalent network solids) are also very common, notable examples being diamond and quartz. Weak van der Waals forces also help hold together certain crystals, such as crystalline molecular solids, as well as the interlayer bonding in graphite. Polymer materials generally will form crystalline regions, but the lengths of the molecules usually prevent complete crystallization—and sometimes polymers are completely amorphous.
A quasicrystal consists of arrays of atoms that are ordered but not strictly periodic. They have many attributes in common with ordinary crystals, such as displaying a discrete pattern in x-ray diffraction, and the ability to form shapes with smooth, flat faces.
Quasicrystals are most famous for their ability to show five-fold symmetry, which is impossible for an ordinary periodic crystal (see crystallographic restriction theorem).
The International Union of Crystallography has redefined the term "crystal" to include both ordinary periodic crystals and quasicrystals ("any solid having an essentially discrete diffraction diagram").
Quasicrystals, first discovered in 1982, are quite rare in practice. Only about 100 solids are known to form quasicrystals, compared to about 400,000 periodic crystals known in 2004. The 2011 Nobel Prize in Chemistry was awarded to Dan Shechtman for the discovery of quasicrystals.
Crystals can have certain special electrical, optical, and mechanical properties that glass and polycrystals normally cannot. These properties are related to the anisotropy of the crystal, i.e. the lack of rotational symmetry in its atomic arrangement. One such property is the piezoelectric effect, where a voltage across the crystal can shrink or stretch it. Another is birefringence, where a double image appears when looking through a crystal. Moreover, various properties of a crystal, including electrical conductivity, electrical permittivity, and Young's modulus, may be different in different directions in a crystal. For example, graphite crystals consist of a stack of sheets, and although each individual sheet is mechanically very strong, the sheets are rather loosely bound to each other. Therefore, the mechanical strength of the material is quite different depending on the direction of stress.
Not all crystals have all of these properties. Conversely, these properties are not quite exclusive to crystals. They can appear in glasses or polycrystals that have been made anisotropic by working or stress—for example, stress-induced birefringence.
"Crystallography" is the science of measuring the crystal structure (in other words, the atomic arrangement) of a crystal. One widely used crystallography technique is X-ray diffraction. Large numbers of known crystal structures are stored in crystallographic databases. | https://en.wikipedia.org/wiki?curid=6015 |
Cytosine
Cytosine (; C) is one of the four main bases found in DNA and RNA, along with adenine, guanine, and thymine (uracil in RNA). It is a pyrimidine derivative, with a heterocyclic aromatic ring and two substituents attached (an amine group at position 4 and a keto group at position 2). The nucleoside of cytosine is cytidine. In Watson-Crick base pairing, it forms three hydrogen bonds with guanine.
Cytosine was discovered and named by Albrecht Kossel and Albert Neumann in 1894 when it was hydrolyzed from calf thymus tissues. A structure was proposed in 1903, and was synthesized (and thus confirmed) in the laboratory in the same year.
In 1998 cytosine was used in an early demonstration of quantum information processing when Oxford University researchers implemented the Deutsch-Jozsa algorithm on a two qubit nuclear magnetic resonance quantum computer (NMRQC).
In March 2015, NASA scientists reported the formation of cytosine, along with uracil and thymine, from pyrimidine under the space-like laboratory conditions, which is of interest because pyrimidine has been found in meteorites although its origin is unknown.
Cytosine can be found as part of DNA, as part of RNA, or as a part of a nucleotide. As cytidine triphosphate (CTP), it can act as a co-factor to enzymes, and can transfer a phosphate to convert adenosine diphosphate (ADP) to adenosine triphosphate (ATP).
In DNA and RNA, cytosine is paired with guanine. However, it is inherently unstable, and can change into uracil (spontaneous deamination). This can lead to a point mutation if not repaired by the DNA repair enzymes such as uracil glycosylase, which cleaves a uracil in DNA.
Cytosine can also be methylated into 5-methylcytosine by an enzyme called DNA methyltransferase or be methylated and hydroxylated to make 5-hydroxymethylcytosine. The difference in rates of deamination of cytosine and 5-methylcytosine (to uracil and thymine) forms the basis of bisulfite sequencing.
When found third in a codon of RNA, cytosine is synonymous with uracil, as they are interchangeable as the third base.
When found as the second base in a codon, the third is always interchangeable. For example, UCU, UCC, UCA and UCG are all serine, regardless of the third base.
Active enzymatic deamination of cytosine or 5-methylcytosine by the APOBEC family of cytosine deaminases could have both beneficial and detrimental implications on various cellular processes as well as on organismal evolution. The implications of deamination on 5-hydroxymethylcytosine, on the other hand, remains less understood.
Cytosine has not been found in meteorites, which suggests the first strands of RNA and DNA had to look elsewhere to obtain this building block. Cytosine likely formed within some meteorite parent bodies, however did not persist within these bodies due to an effective deamination reaction into uracil. | https://en.wikipedia.org/wiki?curid=6016 |
Computational chemistry
Computational chemistry is a branch of chemistry that uses computer simulation to assist in solving chemical problems. It uses methods of theoretical chemistry, incorporated into efficient computer programs, to calculate the structures and properties of molecules and solids. It is necessary because, apart from relatively recent results concerning the hydrogen molecular ion (dihydrogen cation, see references therein for more details), the quantum many-body problem cannot be solved analytically, much less in closed form. While computational results normally complement the information obtained by chemical experiments, it can in some cases predict hitherto unobserved chemical phenomena. It is widely used in the design of new drugs and materials.
Examples of such properties are structure (i.e., the expected positions of the constituent atoms), absolute and relative (interaction) energies, electronic charge density distributions, dipoles and higher multipole moments, vibrational frequencies, reactivity, or other spectroscopic quantities, and cross sections for collision with other particles.
The methods used cover both static and dynamic situations. In all cases, the computer time and other resources (such as memory and disk space) increase rapidly with the size of the system being studied. That system can be one molecule, a group of molecules, or a solid. Computational chemistry methods range from very approximate to highly accurate; the latter are usually feasible for small systems only. "Ab initio" methods are based entirely on quantum mechanics and basic physical constants. Other methods are called empirical or semi-empirical because they use additional empirical parameters.
Both "ab initio" and semi-empirical approaches involve approximations. These range from simplified forms of the first-principles equations that are easier or faster to solve, to approximations limiting the size of the system (for example, periodic boundary conditions), to fundamental approximations to the underlying equations that are required to achieve any solution to them at all. For example, most "ab initio" calculations make the Born–Oppenheimer approximation, which greatly simplifies the underlying Schrödinger equation by assuming that the nuclei remain in place during the calculation. In principle, "ab initio" methods eventually converge to the exact solution of the underlying equations as the number of approximations is reduced. In practice, however, it is impossible to eliminate all approximations, and residual error inevitably remains. The goal of computational chemistry is to minimize this residual error while keeping the calculations tractable.
In some cases, the details of electronic structure are less important than the long-time phase space behavior of molecules. This is the case in conformational studies of proteins and protein-ligand binding thermodynamics. Classical approximations to the potential energy surface are used, typically with molecular mechanics force fields, as they are computationally less intensive than electronic calculations, to enable longer simulations of molecular dynamics. Furthermore, cheminformatics uses even more empirical (and computationally cheaper) methods like machine learning based on physicochemical properties. One typical problem in cheminformatics is to predict the binding affinity of drug molecules to a given target. Other problems include predicting binding specificity, off-target effects, toxicity, and pharmacokinetic properties.
Building on the founding discoveries and theories in the history of quantum mechanics, the first theoretical calculations in chemistry were those of Walter Heitler and Fritz London in 1927. The books that were influential in the early development of computational quantum chemistry include Linus Pauling and E. Bright Wilson's 1935 "Introduction to Quantum Mechanics – with Applications to Chemistry", Eyring, Walter and Kimball's 1944 "Quantum Chemistry", Heitler's 1945 "Elementary Wave Mechanics – with Applications to Quantum Chemistry", and later Coulson's 1952 textbook "Valence", each of which served as primary references for chemists in the decades to follow.
With the development of efficient computer technology in the 1940s, the solutions of elaborate wave equations for complex atomic systems began to be a realizable objective. In the early 1950s, the first semi-empirical atomic orbital calculations were performed. Theoretical chemists became extensive users of the early digital computers. One major advance came with the 1951 paper in Reviews of Modern Physics by Clemens C. J. Roothaan in 1951, largely on the "LCAO MO" approach (Linear Combination of Atomic Orbitals Molecular Orbitals), for many years the second-most cited paper in that journal. A very detailed account of such use in the United Kingdom is given by Smith and Sutcliffe. The first "ab initio" Hartree–Fock method calculations on diatomic molecules were performed in 1956 at MIT, using a basis set of Slater orbitals. For diatomic molecules, a systematic study using a minimum basis set and the first calculation with a larger basis set were published by Ransil and Nesbet respectively in 1960. The first polyatomic calculations using Gaussian orbitals were performed in the late 1950s. The first configuration interaction calculations were performed in Cambridge on the EDSAC computer in the 1950s using Gaussian orbitals by Boys and coworkers. By 1971, when a bibliography of "ab initio" calculations was published, the largest molecules included were naphthalene and azulene. Abstracts of many earlier developments in "ab initio" theory have been published by Schaefer.
In 1964, Hückel method calculations (using a simple linear combination of atomic orbitals (LCAO) method to determine electron energies of molecular orbitals of π electrons in conjugated hydrocarbon systems) of molecules, ranging in complexity from butadiene and benzene to ovalene, were generated on computers at Berkeley and Oxford. These empirical methods were replaced in the 1960s by semi-empirical methods such as CNDO.
In the early 1970s, efficient "ab initio" computer programs such as ATMOL, Gaussian, IBMOL, and POLYAYTOM, began to be used to speed "ab initio" calculations of molecular orbitals. Of these four programs, only Gaussian, now vastly expanded, is still in use, but many other programs are now in use. At the same time, the methods of molecular mechanics, such as MM2 force field, were developed, primarily by Norman Allinger.
One of the first mentions of the term "computational chemistry" can be found in the 1970 book "Computers and Their Role in the Physical Sciences" by Sidney Fernbach and Abraham Haskell Taub, where they state "It seems, therefore, that 'computational chemistry' can finally be more and more of a reality." During the 1970s, widely different methods began to be seen as part of a new emerging discipline of "computational chemistry". The "Journal of Computational Chemistry" was first published in 1980.
Computational chemistry has featured in several Nobel Prize awards, most notably in 1998 and 2013. Walter Kohn, "for his development of the density-functional theory", and John Pople, "for his development of computational methods in quantum chemistry", received the 1998 Nobel Prize in Chemistry. Martin Karplus, Michael Levitt and Arieh Warshel received the 2013 Nobel Prize in Chemistry for "the development of multiscale models for complex chemical systems".
The term "theoretical chemistry" may be defined as a mathematical description of chemistry, whereas "computational chemistry" is usually used when a mathematical method is sufficiently well developed that it can be automated for implementation on a computer. In theoretical chemistry, chemists, physicists, and mathematicians develop algorithms and computer programs to predict atomic and molecular properties and reaction paths for chemical reactions. Computational chemists, in contrast, may simply apply existing computer programs and methodologies to specific chemical questions.
Computational chemistry has two different aspects:
Thus, computational chemistry can assist the experimental chemist or it can challenge the experimental chemist to find entirely new chemical objects.
Several major areas may be distinguished within computational chemistry:
The words "exact" and "perfect" do not apply here, as very few aspects of chemistry can be computed exactly. However, almost every aspect of chemistry can be described in a qualitative or approximate quantitative computational scheme.
Molecules consist of nuclei and electrons, so the methods of quantum mechanics apply. Computational chemists often attempt to solve the non-relativistic Schrödinger equation, with relativistic corrections added, although some progress has been made in solving the fully relativistic Dirac equation. In principle, it is possible to solve the Schrödinger equation in either its time-dependent or time-independent form, as appropriate for the problem in hand; in practice, this is not possible except for very small systems. Therefore, a great number of approximate methods strive to achieve the best trade-off between accuracy and computational cost.
Accuracy can always be improved with greater computational cost. Significant errors can present themselves in ab initio models comprising many electrons, due to the computational cost of full relativistic-inclusive methods. This complicates the study of molecules interacting with high atomic mass unit atoms, such as transitional metals and their catalytic properties. Present algorithms in computational chemistry can routinely calculate the properties of small molecules that contain up to about 40 electrons with errors for energies less than a few kJ/mol. For geometries, bond lengths can be predicted within a few picometres and bond angles within 0.5 degrees. The treatment of larger molecules that contain a few dozen atoms is computationally tractable by more approximate methods such as density functional theory (DFT).
There is some dispute within the field whether or not the latter methods are sufficient to describe complex chemical reactions, such as those in biochemistry. Large molecules can be studied by semi-empirical approximate methods. Even larger molecules are treated by classical mechanics methods that use what are called molecular mechanics (MM). In QM-MM methods, small parts of large complexes are treated quantum mechanically (QM), and the remainder is treated approximately (MM).
One molecular formula can represent more than one molecular isomer: a set of isomers. Each isomer is a local minimum on the energy surface (called the potential energy surface) created from the total energy (i.e., the electronic energy, plus the repulsion energy between the nuclei) as a function of the coordinates of all the nuclei. A stationary point is a geometry such that the derivative of the energy with respect to all displacements of the nuclei is zero. A local (energy) minimum is a stationary point where all such displacements lead to an increase in energy. The local minimum that is lowest is called the global minimum and corresponds to the most stable isomer. If there is one particular coordinate change that leads to a decrease in the total energy in both directions, the stationary point is a transition structure and the coordinate is the reaction coordinate. This process of determining stationary points is called geometry optimization.
The determination of molecular structure by geometry optimization became routine only after efficient methods for calculating the first derivatives of the energy with respect to all atomic coordinates became available. Evaluation of the related second derivatives allows the prediction of vibrational frequencies if harmonic motion is estimated. More importantly, it allows for the characterization of stationary points. The frequencies are related to the eigenvalues of the Hessian matrix, which contains second derivatives. If the eigenvalues are all positive, then the frequencies are all real and the stationary point is a local minimum. If one eigenvalue is negative (i.e., an imaginary frequency), then the stationary point is a transition structure. If more than one eigenvalue is negative, then the stationary point is a more complex one, and is usually of little interest. When one of these is found, it is necessary to move the search away from it if the experimenter is looking solely for local minima and transition structures.
The total energy is determined by approximate solutions of the time-dependent Schrödinger equation, usually with no relativistic terms included, and by making use of the Born–Oppenheimer approximation, which allows for the separation of electronic and nuclear motions, thereby simplifying the Schrödinger equation. This leads to the evaluation of the total energy as a sum of the electronic energy at fixed nuclei positions and the repulsion energy of the nuclei. A notable exception are certain approaches called direct quantum chemistry, which treat electrons and nuclei on a common footing. Density functional methods and semi-empirical methods are variants on the major theme. For very large systems, the relative total energies can be compared using molecular mechanics. The ways of determining the total energy to predict molecular structures are:
The programs used in computational chemistry are based on many different quantum-chemical methods that solve the molecular Schrödinger equation associated with the molecular Hamiltonian. Methods that do not include any empirical or semi-empirical parameters in their equations – being derived directly from theoretical principles, with no inclusion of experimental data – are called "ab initio methods". This does not imply that the solution is an exact one; they are all approximate quantum mechanical calculations. It means that a particular approximation is rigorously defined on first principles (quantum theory) and then solved within an error margin that is qualitatively known beforehand. If numerical iterative methods must be used, the aim is to iterate until full machine accuracy is obtained (the best that is possible with a finite word length on the computer, and within the mathematical and/or physical approximations made).
The simplest type of "ab initio" electronic structure calculation is the Hartree–Fock method (HF), an extension of molecular orbital theory, in which the correlated electron-electron repulsion is not specifically taken into account; only its average effect is included in the calculation. As the basis set size is increased, the energy and wave function tend towards a limit called the Hartree–Fock limit. Many types of calculations (termed post-Hartree–Fock methods) begin with a Hartree–Fock calculation and subsequently correct for electron-electron repulsion, referred to also as electronic correlation. As these methods are pushed to the limit, they approach the exact solution of the non-relativistic Schrödinger equation. To obtain exact agreement with experiment, it is necessary to include relativistic and spin orbit terms, both of which are far more important for heavy atoms. In all of these approaches, along with choice of method, it is necessary to choose a basis set. This is a set of functions, usually centered on the different atoms in the molecule, which are used to expand the molecular orbitals with the linear combination of atomic orbitals (LCAO) molecular orbital method ansatz. Ab initio methods need to define a level of theory (the method) and a basis set.
The Hartree–Fock wave function is a single configuration or determinant. In some cases, particularly for bond breaking processes, this is inadequate, and several configurations must be used. Here, the coefficients of the configurations, and of the basis functions, are optimized together.
The total molecular energy can be evaluated as a function of the molecular geometry; in other words, the potential energy surface. Such a surface can be used for reaction dynamics. The stationary points of the surface lead to predictions of different isomers and the transition structures for conversion between isomers, but these can be determined without a full knowledge of the complete surface.
A particularly important objective, called computational thermochemistry, is to calculate thermochemical quantities such as the enthalpy of formation to chemical accuracy. Chemical accuracy is the accuracy required to make realistic chemical predictions and is generally considered to be 1 kcal/mol or 4 kJ/mol. To reach that accuracy in an economic way it is necessary to use a series of post-Hartree–Fock methods and combine the results. These methods are called quantum chemistry composite methods.
Density functional theory (DFT) methods are often considered to be "ab initio methods" for determining the molecular electronic structure, even though many of the most common functionals use parameters derived from empirical data, or from more complex calculations. In DFT, the total energy is expressed in terms of the total one-electron density rather than the wave function. In this type of calculation, there is an approximate Hamiltonian and an approximate expression for the total electron density. DFT methods can be very accurate for little computational cost. Some methods combine the density functional exchange functional with the Hartree–Fock exchange term and are termed hybrid functional methods.
Semi-empirical quantum chemistry methods are based on the Hartree–Fock method formalism, but make many approximations and obtain some parameters from empirical data. They were very important in computational chemistry from the 60s to the 90s, especially for treating large molecules where the full Hartree–Fock method without the approximations were too costly. The use of empirical parameters appears to allow some inclusion of correlation effects into the methods.
Primitive semi-empirical methods were designed even before, where the two-electron part of the Hamiltonian is not explicitly included. For π-electron systems, this was the Hückel method proposed by Erich Hückel, and for all valence electron systems, the extended Hückel method proposed by Roald Hoffmann. Sometimes, Hückel methods are referred to as "completely emprirical" because they do not derive from a Hamiltonian.Yet, the term "empirical methods", or "empirical force fields" is usually used to describe Molecular Mechanics.
In many cases, large molecular systems can be modeled successfully while avoiding quantum mechanical calculations entirely. Molecular mechanics simulations, for example, use one classical expression for the energy of a compound, for instance the harmonic oscillator. All constants appearing in the equations must be obtained beforehand from experimental data or "ab initio" calculations.
The database of compounds used for parameterization, i.e., the resulting set of parameters and functions is called the force field, is crucial to the success of molecular mechanics calculations. A force field parameterized against a specific class of molecules, for instance proteins, would be expected to only have any relevance when describing other molecules of the same class.
These methods can be applied to proteins and other large biological molecules, and allow studies of the approach and interaction (docking) of potential drug molecules.
Computational chemical methods can be applied to solid state physics problems. The electronic structure of a crystal is in general described by a band structure, which defines the energies of electron orbitals for each point in the Brillouin zone. Ab initio and semi-empirical calculations yield orbital energies; therefore, they can be applied to band structure calculations. Since it is time-consuming to calculate the energy for a molecule, it is even more time-consuming to calculate them for the entire list of points in the Brillouin zone.
Once the electronic and nuclear variables are separated (within the Born–Oppenheimer representation), in the time-dependent approach, the wave packet corresponding to the nuclear degrees of freedom is propagated via the time evolution operator (physics) associated to the time-dependent Schrödinger equation (for the full molecular Hamiltonian). In the complementary energy-dependent approach, the time-independent Schrödinger equation is solved using the scattering theory formalism. The potential representing the interatomic interaction is given by the potential energy surfaces. In general, the potential energy surfaces are coupled via the vibronic coupling terms.
The most popular methods for propagating the wave packet associated to the molecular geometry are:
Molecular dynamics (MD) use either quantum mechanics, molecular mechanics or a mixture of both to calculate forces which are then used to solve Newton's laws of motion to examine the time-dependent behaviour of systems. The result of a molecular dynamics simulation is a trajectory that describes how the position and velocity of particles varies with time.
QM/MM is a hybrid method that attempts to combine the accuracy of quantum mechanics with the speed of molecular mechanics. It is useful for simulating very large molecules such as enzymes.
The atoms in molecules (QTAIM) model of Richard Bader was developed to effectively link the quantum mechanical model of a molecule, as an electronic wavefunction, to chemically useful concepts such as atoms in molecules, functional groups, bonding, the theory of Lewis pairs, and the valence bond model. Bader has demonstrated that these empirically useful chemistry concepts can be related to the topology of the observable charge density distribution, whether measured or calculated from a quantum mechanical wavefunction. QTAIM analysis of molecular wavefunctions is implemented, for example, in the AIMAll software package.
Many self-sufficient exist. Some include many methods covering a wide range, while others concentrate on a very specific range or even on one method. Details of most of them can be found in: | https://en.wikipedia.org/wiki?curid=6019 |
Crash (Ballard novel)
Crash is a novel by English author J. G. Ballard, first published in 1973. It is a story about symphorophilia; specifically car-crash sexual fetishism: its protagonists become sexually aroused by staging and participating in real car-crashes.
In 1996, the novel was made into a film of the same name by David Cronenberg.
The story is told through the eyes of narrator James Ballard, named after the author himself, but it centers on the sinister figure of Dr. Robert Vaughan, a "former TV-scientist, turned nightmare angel of the expressways". Ballard meets Vaughan after being involved in a car accident himself near London Airport. Gathering around Vaughan is a group of alienated people, all of them former crash victims, who follow him in his pursuit to re-enact the crashes of celebrities and experience what the narrator calls "a new sexuality, born from a perverse technology". Vaughan's ultimate fantasy is to die in a head-on collision with movie star Elizabeth Taylor.
The novel received divided reviews when originally published. One publisher's reader returned the verdict "This author is beyond psychiatric help. Do Not Publish!" A 1973 review in The New York Times was equally horrified: ""Crash" is, hands-down, the most repulsive book I've yet to come across."
However, retrospective opinion now considers "Crash" to be one of Ballard's best and most challenging works. Reassessing "Crash" in "The Guardian", Zadie Smith wrote, "C"rash" is an existential book about how "everybody uses everything". How everything uses everybody. And yet it is not a hopeless vision." On Ballard's legacy, she writes: "In Ballard's work there is always this mix of futuristic dread and excitement, a sweet spot where dystopia and utopia converge. For we cannot say we haven't got precisely what we dreamed of, what we always wanted, so badly."
The Papers of J.G. Ballard at the British Library include two revised drafts of "Crash" (Add MS 88938/3/8). Scanned extracts from Ballard's drafts are included in "Crash: The Collector's Edition," ed. Chris Beckett.
The Normal's 1978 song "Warm Leatherette" was inspired by the novel as was "Miss the Girl," a 1983 single by The Creatures.
The Manic Street Preachers' song "Mausoleum" from 1994's "The Holy Bible" contains the famous Ballard quote about his reasons for writing the book, "I wanted to rub the human face in its own vomit. I wanted to force it to look in the mirror."
An apparently unauthorized adaptation of Crash called "Nightmare Angel" was filmed in 1986 by Susan Emerling and Zoe Beloff. This short film bears the credit "Inspired by J.G. Ballard". | https://en.wikipedia.org/wiki?curid=6020 |
C (programming language)
C (, as in the letter "c") is a general-purpose, procedural computer programming language supporting structured programming, lexical variable scope, and recursion, with a static type system. By design, C provides constructs that map efficiently to typical machine instructions. It has found lasting use in applications previously coded in assembly language. Such applications include operating systems and various application software for computers architectures that range from supercomputers to PLCs and embedded systems.
A successor to the programming language "B", C was originally developed at Bell Labs by Dennis Ritchie between 1972 and 1973 to construct utilities running on Unix. It was applied to re-implementing the kernel of the Unix operating system. During the 1980s, C gradually gained popularity. It has become one of the most widely used programming languages, with C compilers from various vendors available for the majority of existing computer architectures and operating systems. C has been standardized by the ANSI since 1989 (ANSI C) and by the International Organization for Standardization (ISO).
C is an imperative procedural language. It was designed to be compiled to provide low-level access to memory and language constructs that map efficiently to machine instructions, all with minimal runtime support. Despite its low-level capabilities, the language was designed to encourage cross-platform programming. A standards-compliant C program written with portability in mind can be compiled for a wide variety of computer platforms and operating systems with few changes to its source code.
Like most procedural languages in the ALGOL tradition, C has facilities for structured programming and allows lexical variable scope and recursion. Its static type system prevents unintended operations. In C, all executable code is contained within subroutines (also called "functions", though not strictly in the sense of functional programming). Function parameters are always passed by value. Pass-by-reference is simulated in C by explicitly passing pointer values. C program source text is free-format, using the semicolon as a statement terminator and curly braces for grouping blocks of statements.
The C language also exhibits the following characteristics:
While C does not include certain features found in other languages (such as object orientation and garbage collection), these can be implemented or emulated, often through the use of external libraries (e.g., the GLib Object System or the Boehm garbage collector).
Many later languages have borrowed directly or indirectly from C, including C++, C#, Unix's C shell, D, Go, Java, JavaScript (including transpilers), Limbo, LPC, Objective-C, Perl, PHP, Python, Rust, Swift, Verilog and SystemVerilog (hardware description languages). These languages have drawn many of their control structures and other basic features from C. Most of them (Python being a dramatic exception) also express highly similar syntax to C, and they tend to combine the recognizable expression and statement syntax of C with underlying type systems, data models, and semantics that can be radically different.
The origin of C is closely tied to the development of the Unix operating system, originally implemented in assembly language on a PDP-7 by Dennis Ritchie and Ken Thompson, incorporating several ideas from colleagues. Eventually, they decided to port the operating system to a PDP-11. The original PDP-11 version of Unix was also developed in assembly language.
Thompson desired a programming language to make utilities for the new platform. At first, he tried to make a Fortran compiler, but soon gave up the idea. Instead, he created a cut-down version of the recently developed BCPL systems programming language. The official description of BCPL was not available at the time, and Thompson modified the syntax to be less wordy, producing the similar but somewhat simpler B. However, few utilities were ultimately written in B because it was too slow, and B could not take advantage of PDP-11 features such as byte addressability.
In 1972, Ritchie started to improve B, which resulted in creating a new language C. The C compiler and some utilities made with it were included in Version 2 Unix.
At Version 4 Unix, released in November 1973, the Unix kernel was extensively re-implemented in C. By this time, the C language had acquired some powerful features such as codice_11 types.
Unix was one of the first operating system kernels implemented in a language other than assembly. Earlier instances include the Multics system (which was written in PL/I) and Master Control Program (MCP) for the Burroughs B5000 (which was written in ALGOL) in 1961. In around 1977, Ritchie and Stephen C. Johnson made further changes to the language to facilitate portability of the Unix operating system. Johnson's Portable C Compiler served as the basis for several implementations of C on new platforms.
In 1978, Brian Kernighan and Dennis Ritchie published the first edition of "The C Programming Language". This book, known to C programmers as "K&R", served for many years as an informal specification of the language. The version of C that it describes is commonly referred to as "K&R C". The second edition of the book covers the later ANSI C standard, described below.
"K&R" introduced several language features:
Even after the publication of the 1989 ANSI standard, for many years K&R C was still considered the "lowest common denominator" to which C programmers restricted themselves when maximum portability was desired, since many older compilers were still in use, and because carefully written K&R C code can be legal Standard C as well.
In early versions of C, only functions that return types other than codice_29 must be declared if used before the function definition; functions used without prior declaration were presumed to return type codice_29.
For example:
long some_function();
/* int */ other_function();
/* int */ calling_function()
The codice_29 type specifiers which are commented out could be omitted in K&R C, but are required in later standards.
Since K&R function declarations did not include any information about function arguments, function parameter type checks were not performed, although some compilers would issue a warning message if a local function was called with the wrong number of arguments, or if multiple calls to an external function used different numbers or types of arguments. Separate tools such as Unix's lint utility were developed that (among other things) could check for consistency of function use across multiple source files.
In the years following the publication of K&R C, several features were added to the language, supported by compilers from AT&T (in particular PCC) and some other vendors. These included:
The large number of extensions and lack of agreement on a standard library, together with the language popularity and the fact that not even the Unix compilers precisely implemented the K&R specification, led to the necessity of standardization.
During the late 1970s and 1980s, versions of C were implemented for a wide variety of mainframe computers, minicomputers, and microcomputers, including the IBM PC, as its popularity began to increase significantly.
In 1983, the American National Standards Institute (ANSI) formed a committee, X3J11, to establish a standard specification of C. X3J11 based the C standard on the Unix implementation; however, the non-portable portion of the Unix C library was handed off to the IEEE working group 1003 to become the basis for the 1988 POSIX standard. In 1989, the C standard was ratified as ANSI X3.159-1989 "Programming Language C". This version of the language is often referred to as ANSI C, Standard C, or sometimes C89.
In 1990, the ANSI C standard (with formatting changes) was adopted by the International Organization for Standardization (ISO) as ISO/IEC 9899:1990, which is sometimes called C90. Therefore, the terms "C89" and "C90" refer to the same programming language.
ANSI, like other national standards bodies, no longer develops the C standard independently, but defers to the international C standard, maintained by the working group ISO/IEC JTC1/SC22/WG14. National adoption of an update to the international standard typically occurs within a year of ISO publication.
One of the aims of the C standardization process was to produce a superset of K&R C, incorporating many of the subsequently introduced unofficial features. The standards committee also included several additional features such as function prototypes (borrowed from C++), codice_14 pointers, support for international character sets and locales, and preprocessor enhancements. Although the syntax for parameter declarations was augmented to include the style used in C++, the K&R interface continued to be permitted, for compatibility with existing source code.
C89 is supported by current C compilers, and most modern C code is based on it. Any program written only in Standard C and without any hardware-dependent assumptions will run correctly on any platform with a conforming C implementation, within its resource limits. Without such precautions, programs may compile only on a certain platform or with a particular compiler, due, for example, to the use of non-standard libraries, such as GUI libraries, or to a reliance on compiler- or platform-specific attributes such as the exact size of data types and byte endianness.
In cases where code must be compilable by either standard-conforming or K&R C-based compilers, the codice_37 macro can be used to split the code into Standard and K&R sections to prevent the use on a K&R C-based compiler of features available only in Standard C.
After the ANSI/ISO standardization process, the C language specification remained relatively static for several years. In 1995, Normative Amendment 1 to the 1990 C standard (ISO/IEC 9899/AMD1:1995, known informally as C95) was published, to correct some details and to add more extensive support for international character sets.
The C standard was further revised in the late 1990s, leading to the publication of ISO/IEC 9899:1999 in 1999, which is commonly referred to as "C99". It has since been amended three times by Technical Corrigenda.
C99 introduced several new features, including inline functions, several new data types (including codice_38 and a codice_39 type to represent complex numbers), variable-length arrays and flexible array members, improved support for IEEE 754 floating point, support for variadic macros (macros of variable arity), and support for one-line comments beginning with codice_40, as in BCPL or C++. Many of these had already been implemented as extensions in several C compilers.
C99 is for the most part backward compatible with C90, but is stricter in some ways; in particular, a declaration that lacks a type specifier no longer has codice_29 implicitly assumed. A standard macro codice_42 is defined with value codice_43 to indicate that C99 support is available. GCC, Solaris Studio, and other C compilers now support many or all of the new features of C99. The C compiler in Microsoft Visual C++, however, implements the C89 standard and those parts of C99 that are required for compatibility with C++11.
In 2007, work began on another revision of the C standard, informally called "C1X" until its official publication on 2011-12-08. The C standards committee adopted guidelines to limit the adoption of new features that had not been tested by existing implementations.
The C11 standard adds numerous new features to C and the library, including type generic macros, anonymous structures, improved Unicode support, atomic operations, multi-threading, and bounds-checked functions. It also makes some portions of the existing C99 library optional, and improves compatibility with C++. The standard macro codice_42 is defined as codice_45 to indicate that C11 support is available.
Published in June 2018, C18 is the current standard for the C programming language. It introduces no new language features, only technical corrections, and clarifications to defects in C11. The standard macro codice_42 is defined as codice_47.
Historically, embedded C programming requires nonstandard extensions to the C language in order to support exotic features such as fixed-point arithmetic, multiple distinct memory banks, and basic I/O operations.
In 2008, the C Standards Committee published a technical report extending the C language to address these issues by providing a common standard for all implementations to adhere to. It includes a number of features not available in normal C, such as fixed-point arithmetic, named address spaces, and basic I/O hardware addressing.
C has a formal grammar specified by the C standard. Line endings are generally not significant in C; however, line boundaries do have significance during the preprocessing phase. Comments may appear either between the delimiters codice_48 and codice_49, or (since C99) following codice_40 until the end of the line. Comments delimited by codice_48 and codice_49 do not nest, and these sequences of characters are not interpreted as comment delimiters if they appear inside string or character literals.
C source files contain declarations and function definitions. Function definitions, in turn, contain declarations and statements. Declarations either define new types using keywords such as codice_11, codice_34, and codice_13, or assign types to and perhaps reserve storage for new variables, usually by writing the type followed by the variable name. Keywords such as codice_56 and codice_29 specify built-in types. Sections of code are enclosed in braces (codice_58 and codice_59, sometimes called "curly brackets") to limit the scope of declarations and to act as a single statement for control structures.
As an imperative language, C uses "statements" to specify actions. The most common statement is an "expression statement", consisting of an expression to be evaluated, followed by a semicolon; as a side effect of the evaluation, functions may be called and variables may be assigned new values. To modify the normal sequential execution of statements, C provides several control-flow statements identified by reserved keywords. Structured programming is supported by codice_60(-codice_61) conditional execution and by codice_62-codice_4, codice_4, and codice_2 iterative execution (looping). The codice_2 statement has separate initialization, testing, and reinitialization expressions, any or all of which can be omitted. codice_67 and codice_68 can be used to leave the innermost enclosing loop statement or skip to its reinitialization. There is also a non-structured codice_69 statement which branches directly to the designated label within the function. codice_5 selects a codice_71 to be executed based on the value of an integer expression.
Expressions can use a variety of built-in operators and may contain function calls. The order in which arguments to functions and operands to most operators are evaluated is unspecified. The evaluations may even be interleaved. However, all side effects (including storage to variables) will occur before the next "sequence point"; sequence points include the end of each expression statement, and the entry to and return from each function call. Sequence points also occur during evaluation of expressions containing certain operators (codice_72, codice_10, codice_74 and the comma operator). This permits a high degree of object code optimization by the compiler, but requires C programmers to take more care to obtain reliable results than is needed for other programming languages.
Kernighan and Ritchie say in the Introduction of "The C Programming Language": "C, like any other language, has its blemishes. Some of the operators have the wrong precedence; some parts of the syntax could be better." The C standard did not attempt to correct many of these blemishes, because of the impact of such changes on already existing software.
The basic C source character set includes the following characters:
Newline indicates the end of a text line; it need not correspond to an actual single character, although for convenience C treats it as one.
Additional multi-byte encoded characters may be used in string literals, but they are not entirely portable. The latest C standard (C11) allows multi-national Unicode characters to be embedded portably within C source text by using codice_82 or codice_83 encoding (where the codice_84 denotes a hexadecimal character), although this feature is not yet widely implemented.
The basic C execution character set contains the same characters, along with representations for alert, backspace, and carriage return. Run-time support for extended character sets has increased with each revision of the C standard.
C89 has 32 reserved words, also known as keywords, which are the words that cannot be used for any purposes other than those for which they are predefined:
C99 reserved five more words:
C11 reserved seven more words:
Most of the recently reserved words begin with an underscore followed by a capital letter, because identifiers of that form were previously reserved by the C standard for use only by implementations. Since existing program source code should not have been using these identifiers, it would not be affected when C implementations started supporting these extensions to the programming language. Some standard headers do define more convenient synonyms for underscored identifiers. The language previously included a reserved word called codice_129, but this was seldom implemented, and has now been removed as a reserved word.
C supports a rich set of operators, which are symbols used within an expression to specify the manipulations to be performed while evaluating that expression. C has operators for:
C uses the operator codice_135 (used in mathematics to express equality) to indicate assignment, following the precedent of Fortran and PL/I, but unlike ALGOL and its derivatives. C uses the operator codice_155 to test for equality. The similarity between these two operators (assignment and equality) may result in the accidental use of one in place of the other, and in many cases, the mistake does not produce an error message (although some compilers produce warnings). For example, the conditional expression codice_175 might mistakenly be written as codice_176, which will be evaluated as true if codice_75 is not zero after the assignment.
The C operator precedence is not always intuitive. For example, the operator codice_155 binds more tightly than (is executed prior to) the operators codice_9 (bitwise AND) and codice_148 (bitwise OR) in expressions such as codice_181, which must be written as codice_182 if that is the coder's intent.
The "hello, world" example, which appeared in the first edition of "K&R", has become the model for an introductory program in most programming textbooks. The program prints "hello, world" to the standard output, which is usually a terminal or screen display.
The original version was:
main()
A standard-conforming "hello, world" program is:
int main(void)
The first line of the program contains a preprocessing directive, indicated by codice_183. This causes the compiler to replace that line with the entire text of the codice_184 standard header, which contains declarations for standard input and output functions such as codice_185 and codice_186. The angle brackets surrounding codice_184 indicate that codice_184 is located using a search strategy that prefers headers provided with the compiler to other headers having the same name, as opposed to double quotes which typically include local or project-specific header files.
The next line indicates that a function named codice_189 is being defined. The codice_189 function serves a special purpose in C programs; the run-time environment calls the codice_189 function to begin program execution. The type specifier codice_29 indicates that the value that is returned to the invoker (in this case the run-time environment) as a result of evaluating the codice_189 function, is an integer. The keyword codice_14 as a parameter list indicates that this function takes no arguments.
The opening curly brace indicates the beginning of the definition of the codice_189 function.
The next line "calls" (diverts execution to) a function named codice_185, which in this case is supplied from a system library. In this call, the codice_185 function is "passed" (provided with) a single argument, the address of the first character in the string literal codice_198. The string literal is an unnamed array with elements of type codice_56, set up automatically by the compiler with a final 0-valued character to mark the end of the array (codice_185 needs to know this). The codice_201 is an "escape sequence" that C translates to a "newline" character, which on output signifies the end of the current line. The return value of the codice_185 function is of type codice_29, but it is silently discarded since it is not used. (A more careful program might test the return value to determine whether or not the codice_185 function succeeded.) The semicolon codice_205 terminates the statement.
The closing curly brace indicates the end of the code for the codice_189 function. According to the C99 specification and newer, the codice_189 function, unlike any other function, will implicitly return a value of codice_79 upon reaching the codice_59 that terminates the function. (Formerly an explicit codice_210 statement was required.) This is interpreted by the run-time system as an exit code indicating successful execution.
The type system in C is static and weakly typed, which makes it similar to the type system of ALGOL descendants such as Pascal. There are built-in types for integers of various sizes, both signed and unsigned, floating-point numbers, and enumerated types (codice_13). Integer type codice_56 is often used for single-byte characters. C99 added a boolean datatype. There are also derived types including arrays, pointers, records (codice_11), and unions (codice_34).
C is often used in low-level systems programming where escapes from the type system may be necessary. The compiler attempts to ensure type correctness of most expressions, but the programmer can override the checks in various ways, either by using a "type cast" to explicitly convert a value from one type to another, or by using pointers or unions to reinterpret the underlying bits of a data object in some other way.
Some find C's declaration syntax unintuitive, particularly for function pointers. (Ritchie's idea was to declare identifiers in contexts resembling their use: "declaration reflects use".)
C's "usual arithmetic conversions" allow for efficient code to be generated, but can sometimes produce unexpected results. For example, a comparison of signed and unsigned integers of equal width requires a conversion of the signed value to unsigned. This can generate unexpected results if the signed value is negative.
C supports the use of pointers, a type of reference that records the address or location of an object or function in memory. Pointers can be "dereferenced" to access data stored at the address pointed to, or to invoke a pointed-to function. Pointers can be manipulated using assignment or pointer arithmetic. The run-time representation of a pointer value is typically a raw memory address (perhaps augmented by an offset-within-word field), but since a pointer's type includes the type of the thing pointed to, expressions including pointers can be type-checked at compile time. Pointer arithmetic is automatically scaled by the size of the pointed-to data type. Pointers are used for many purposes in C. Text strings are commonly manipulated using pointers into arrays of characters. Dynamic memory allocation is performed using pointers. Many data types, such as trees, are commonly implemented as dynamically allocated codice_11 objects linked together using pointers. Pointers to functions are useful for passing functions as arguments to higher-order functions (such as qsort or bsearch) or as callbacks to be invoked by event handlers.
A "null pointer value" explicitly points to no valid location. Dereferencing a null pointer value is undefined, often resulting in a segmentation fault. Null pointer values are useful for indicating special cases such as no "next" pointer in the final node of a linked list, or as an error indication from functions returning pointers. In appropriate contexts in source code, such as for assigning to a pointer variable, a "null pointer constant" can be written as codice_79, with or without explicit casting to a pointer type, or as the codice_217 macro defined by several standard headers. In conditional contexts, null pointer values evaluate to false, while all other pointer values evaluate to true.
Void pointers (codice_218) point to objects of unspecified type, and can therefore be used as "generic" data pointers. Since the size and type of the pointed-to object is not known, void pointers cannot be dereferenced, nor is pointer arithmetic on them allowed, although they can easily be (and in many contexts implicitly are) converted to and from any other object pointer type.
Careless use of pointers is potentially dangerous. Because they are typically unchecked, a pointer variable can be made to point to any arbitrary location, which can cause undesirable effects. Although properly used pointers point to safe places, they can be made to point to unsafe places by using invalid pointer arithmetic; the objects they point to may continue to be used after deallocation (dangling pointers); they may be used without having been initialized (wild pointers); or they may be directly assigned an unsafe value using a cast, union, or through another corrupt pointer. In general, C is permissive in allowing manipulation of and conversion between pointer types, although compilers typically provide options for various levels of checking. Some other programming languages address these problems by using more restrictive reference types.
Array types in C are traditionally of a fixed, static size specified at compile time. (The more recent C99 standard also allows a form of variable-length arrays.) However, it is also possible to allocate a block of memory (of arbitrary size) at run-time, using the standard library's codice_219 function, and treat it as an array. C's unification of arrays and pointers means that declared arrays and these dynamically allocated simulated arrays are virtually interchangeable.
Since arrays are always accessed (in effect) via pointers, array accesses are typically "not" checked against the underlying array size, although some compilers may provide bounds checking as an option. Array bounds violations are therefore possible and rather common in carelessly written code, and can lead to various repercussions, including illegal memory accesses, corruption of data, buffer overruns, and run-time exceptions. If bounds checking is desired, it must be done manually.
C does not have a special provision for declaring multi-dimensional arrays, but rather relies on recursion within the type system to declare arrays of arrays, which effectively accomplishes the same thing. The index values of the resulting "multi-dimensional array" can be thought of as increasing in row-major order.
Multi-dimensional arrays are commonly used in numerical algorithms (mainly from applied linear algebra) to store matrices. The structure of the C array is well suited to this particular task. However, since arrays are passed merely as pointers, the bounds of the array must be known fixed values or else explicitly passed to any subroutine that requires them, and dynamically sized arrays of arrays cannot be accessed using double indexing. (A workaround for this is to allocate the array with an additional "row vector" of pointers to the columns.)
C99 introduced "variable-length arrays" which address some, but not all, of the issues with ordinary C arrays.
The subscript notation codice_220 (where codice_221 designates a pointer) is syntactic sugar for codice_222. Taking advantage of the compiler's knowledge of the pointer type, the address that codice_223 points to is not the base address (pointed to by codice_221) incremented by codice_26 bytes, but rather is defined to be the base address incremented by codice_26 multiplied by the size of an element that codice_221 points to. Thus, codice_220 designates the codice_229th element of the array.
Furthermore, in most expression contexts (a notable exception is as operand of codice_107), the name of an array is automatically converted to a pointer to the array's first element. This implies that an array is never copied as a whole when named as an argument to a function, but rather only the address of its first element is passed. Therefore, although function calls in C use pass-by-value semantics, arrays are in effect passed by reference.
The size of an element can be determined by applying the operator codice_107 to any dereferenced element of codice_221, as in codice_233 or codice_234, and the number of elements in a declared array codice_77 can be determined as codice_236. The latter only applies to array names: variables declared with subscripts (codice_237). Due to the semantics of C, it is not possible to determine the entire size of arrays through pointers to arrays or those created by dynamic allocation (codice_219); code such as codice_239 (where codice_240 designates a pointer) will not work since the compiler assumes the size of the pointer itself is being requested. Since array name arguments to codice_107 are not converted to pointers, they do not exhibit such ambiguity. However, arrays created by dynamic allocation are accessed by pointers rather than true array variables, so they suffer from the same codice_107 issues as array pointers.
Thus, despite this apparent equivalence between array and pointer variables, there is still a distinction to be made between them. Even though the name of an array is, in most expression contexts, converted into a pointer (to its first element), this pointer does not itself occupy any storage; the array name is not an l-value, and its address is a constant, unlike a pointer variable. Consequently, what an array "points to" cannot be changed, and it is impossible to assign a new address to an array name. Array contents may be copied, however, by using the codice_243 function, or by accessing the individual elements.
One of the most important functions of a programming language is to provide facilities for managing memory and the objects that are stored in memory. C provides three distinct ways to allocate memory for objects:
These three approaches are appropriate in different situations and have various trade-offs. For example, static memory allocation has little allocation overhead, automatic allocation may involve slightly more overhead, and dynamic memory allocation can potentially have a great deal of overhead for both allocation and deallocation. The persistent nature of static objects is useful for maintaining state information across function calls, automatic allocation is easy to use but stack space is typically much more limited and transient than either static memory or heap space, and dynamic memory allocation allows convenient allocation of objects whose size is known only at run-time. Most C programs make extensive use of all three.
Where possible, automatic or static allocation is usually simplest because the storage is managed by the compiler, freeing the programmer of the potentially error-prone chore of manually allocating and releasing storage. However, many data structures can change in size at runtime, and since static allocations (and automatic allocations before C99) must have a fixed size at compile-time, there are many situations in which dynamic allocation is necessary. Prior to the C99 standard, variable-sized arrays were a common example of this. (See the article on codice_219 for an example of dynamically allocated arrays.) Unlike automatic allocation, which can fail at run time with uncontrolled consequences, the dynamic allocation functions return an indication (in the form of a null pointer value) when the required storage cannot be allocated. (Static allocation that is too large is usually detected by the linker or loader, before the program can even begin execution.)
Unless otherwise specified, static objects contain zero or null pointer values upon program startup. Automatically and dynamically allocated objects are initialized only if an initial value is explicitly specified; otherwise they initially have indeterminate values (typically, whatever bit pattern happens to be present in the storage, which might not even represent a valid value for that type). If the program attempts to access an uninitialized value, the results are undefined. Many modern compilers try to detect and warn about this problem, but both false positives and false negatives can occur.
Another issue is that heap memory allocation has to be synchronized with its actual usage in any program in order for it to be reused as much as possible. For example, if the only pointer to a heap memory allocation goes out of scope or has its value overwritten before codice_248 is called, then that memory cannot be recovered for later reuse and is essentially lost to the program, a phenomenon known as a "memory leak." Conversely, it is possible for memory to be freed but continue to be referenced, leading to unpredictable results. Typically, the symptoms will appear in a portion of the program far removed from the actual error, making it difficult to track down the problem. (Such issues are ameliorated in languages with automatic garbage collection.)
The C programming language uses libraries as its primary method of extension. In C, a library is a set of functions contained within a single "archive" file. Each library typically has a header file, which contains the prototypes of the functions contained within the library that may be used by a program, and declarations of special data types and macro symbols used with these functions. In order for a program to use a library, it must include the library's header file, and the library must be linked with the program, which in many cases requires compiler flags (e.g., codice_249, shorthand for "link the math library").
The most common C library is the C standard library, which is specified by the ISO and ANSI C standards and comes with every C implementation (implementations which target limited environments such as embedded systems may provide only a subset of the standard library). This library supports stream input and output, memory allocation, mathematics, character strings, and time values. Several separate standard headers (for example, codice_184) specify the interfaces for these and other standard library facilities.
Another common set of C library functions are those used by applications specifically targeted for Unix and Unix-like systems, especially functions which provide an interface to the kernel. These functions are detailed in various standards such as POSIX and the Single UNIX Specification.
Since many programs have been written in C, there are a wide variety of other libraries available. Libraries are often written in C because C compilers generate efficient object code; programmers then create interfaces to the library so that the routines can be used from higher-level languages like Java, Perl, and Python.
File input and output (I/O) is not part of the C language itself but instead is handled by libraries (such as the C standard library) and their associated header files (e.g. codice_184). File handling is generally implemented through high-level I/O which works through streams. A stream is from this perspective a data flow that is independent of devices, while a file is a concrete device. The high level I/O is done through the association of a stream to a file. In the C standard library, a buffer (a memory area or queue) is temporarily used to store data before it's sent to the final destination. This reduces the time spent waiting for slower devices, for example a hard drive or solid state drive. Low-level I/O functions are not part of the standard C library but are generally part of "bare metal" programming (programming that's independent of any operating system such as most but not all embedded programming). With few exceptions, implementations include low-level I/O.
A number of tools have been developed to help C programmers find and fix statements with undefined behavior or possibly erroneous expressions, with greater rigor than that provided by the compiler. The tool lint was the first such, leading to many others.
Automated source code checking and auditing are beneficial in any language, and for C many such tools exist, such as Lint. A common practice is to use Lint to detect questionable code when a program is first written. Once a program passes Lint, it is then compiled using the C compiler. Also, many compilers can optionally warn about syntactically valid constructs that are likely to actually be errors. MISRA C is a proprietary set of guidelines to avoid such questionable code, developed for embedded systems.
There are also compilers, libraries, and operating system level mechanisms for performing actions that are not a standard part of C, such as bounds checking for arrays, detection of buffer overflow, serialization, dynamic memory tracking, and automatic garbage collection.
Tools such as Purify or Valgrind and linking with libraries containing special versions of the memory allocation functions can help uncover runtime errors in memory usage.
C is widely used for systems programming in implementing operating systems and embedded system applications, because C code, when written for portability, can be used for most purposes, yet when needed, system-specific code can be used to access specific hardware addresses and to perform type punning to match externally imposed interface requirements, with a low run-time demand on system resources.
C can also be used for website programming using CGI as a "gateway" for information between the Web application, the server, and the browser. C is often chosen over interpreted languages because of its speed, stability, and near-universal availability.
One consequence of C's wide availability and efficiency is that compilers, libraries and interpreters of other programming languages are often implemented in C. The reference implementations of Python, Perl and PHP, for example, are all written in C.
Because the layer of abstraction is thin and the overhead is low, C enables programmers to create efficient implementations of algorithms and data structures, useful for computationally intense programs. For example, the GNU Multiple Precision Arithmetic Library, the GNU Scientific Library, Mathematica, and MATLAB are completely or partially written in C.
C is sometimes used as an intermediate language by implementations of other languages. This approach may be used for portability or convenience; by using C as an intermediate language, additional machine-specific code generators are not necessary. C has some features, such as line-number preprocessor directives and optional superfluous commas at the end of initializer lists, that support compilation of generated code. However, some of C's shortcomings have prompted the development of other C-based languages specifically designed for use as intermediate languages, such as C--.
C has also been widely used to implement end-user applications. However, such applications can also be written in newer, higher-level languages.
C has both directly and indirectly influenced many later languages such as C#, D, Go, Java, JavaScript, Limbo, LPC, Perl, PHP, Python, and Unix's C shell. The most pervasive influence has been syntactical, all of the languages mentioned combine the statement and (more or less recognizably) expression syntax of C with type systems, data models and/or large-scale program structures that differ from those of C, sometimes radically.
Several C or near-C interpreters exist, including Ch and CINT, which can also be used for scripting.
When object-oriented languages became popular, C++ and Objective-C were two different extensions of C that provided object-oriented capabilities. Both languages were originally implemented as source-to-source compilers; source code was translated into C, and then compiled with a C compiler.
The C++ programming language was devised by Bjarne Stroustrup as an approach to providing object-oriented functionality with a C-like syntax. C++ adds greater typing strength, scoping, and other tools useful in object-oriented programming, and permits generic programming via templates. Nearly a superset of C, C++ now supports most of C, with a few exceptions.
Objective-C was originally a very "thin" layer on top of C, and remains a strict superset of C that permits object-oriented programming using a hybrid dynamic/static typing paradigm. Objective-C derives its syntax from both C and Smalltalk: syntax that involves preprocessing, expressions, function declarations, and function calls is inherited from C, while the syntax for object-oriented features was originally taken from Smalltalk.
In addition to C++ and Objective-C, Ch, Cilk and Unified Parallel C are nearly supersets of C. | https://en.wikipedia.org/wiki?curid=6021 |
Castle of the Winds
Castle of the Winds (also known as "Castle of the Winds: Vanquish the Dark Forces") is a tile-based roguelike video game for Microsoft Windows. It was developed by Rick Saada in 1989 and distributed by Epic MegaGames in 1993.
The game was given around 1998 into the public domain and provided as Freeware download by the author. Though it is secondary to its hack and slash gameplay, "Castle of the Winds" has a plot loosely based on Norse mythology, told with setting changes, unique items, and occasional passages of text. The game is composed of two parts: A Question of Vengeance, released as shareware, and Lifthransir's Bane, sold commercially. A combined license for both parts was also sold.
The game differs from most roguelikes in a number of ways. Its interface is mouse-dependent, but supports keyboard shortcuts (such as 'g' to get an item). "Castle of the Winds" also allows the player to restore saved games after dying.
The game favors the use of magic in combat, as spells are the only weapons that work from a distance. The player character automatically gains a spell with each experience level, and can permanently gain others using corresponding books, until all thirty spells available are learned. There are two opposing pairs of elements: cold vs. fire and lightning vs. acid/poison. Spells are divided into six categories: attack, defense, healing, movement, divination, and miscellaneous.
"Castle of the Winds" possesses an inventory system that limits a player's load based on weight and bulk, rather than by number of items. It allows the character to use different containers, including packs, belts, chests, and bags. Other items include weapons, armor, protective clothing, purses, and ornamental jewellery. Almost every item in the game can be normal, cursed, or enchanted, with curses and enchantments working in a manner similar to "NetHack". Although items do not break with use, they may already be broken or rusted when found. Most objects that the character currently carries can be renamed.
Wherever the player goes before entering the dungeon, there is always a town which offers the basic services of a temple for healing and curing curses, a junk store where anything can be sold for a few copper coins, a sage who can identify items and (from the second town onwards) a bank for storing the total capacity of coins to lighten the player's load. Other services that differ and vary in what they sell are outfitters, weaponsmiths, armoursmiths, magic shops and general stores.
The game tracks how much time has been spent playing the game. Although story events are not triggered by the passage of time, it does determine when merchants rotate their stock. Victorious players are listed as "Valhalla's Champions" in the order of time taken, from fastest to slowest. If the player dies, they are still put on the list, but are categorized as "Dead", with their experience point total listed as at the final killing blow. The amount of time spent also determines the difficulty of the last boss.
The player begins in a tiny hamlet, near which he/she used to live. His/her farm has been destroyed and godparents killed. After clearing out an abandoned mine, the player finds a scrap of parchment that reveals the death of the player's godparents was ordered by an unknown enemy. The player then returns to the hamlet to find it pillaged, and decides to travel to Bjarnarhaven.
Once in Bjarnarhaven, the player explores the levels beneath a nearby fortress, eventually facing Hrungnir, the Hill Giant Lord, responsible for ordering the player's godparents' death. Hrungnir carries the Enchanted Amulet of Kings. Upon activating the amulet, the player is informed of his/her past by his/her dead father, after which the player is transported to the town of Crossroads, and "Part I" ends. The game can be imported or started over in "Part II".
The town of Crossroads is run by a Jarl who at first does not admit the player, but later (on up to three occasions) provides advice and rewards. The player then enters the nearby ruined titular Castle of the Winds. There the player meets his/her deceased grandfather, who instructs him/her to venture into the dungeons below, defeat Surtur, and reclaim their birthright. Venturing deeper, the player encounters monsters run rampant, a desecrated crypt, a necromancer, and the installation of various special rooms for elementals. The player eventually meets and defeats the Wolf-Man leader, Bear-Man leader, the four Jotun kings, a Demon Lord, and finally Surtur. Upon defeating Surtur and escaping the dungeons, the player sits upon the throne, completing the game.
Inspired by his love of RPGs and while learning Windows programming in the 80s, Rick Saada designed and completed Castle of the Winds. The game sold 13,500 copies. By 1998, the game's author, Rick Saada, decided to distribute the entirety of "Castle of the Winds" free of charge.
The game is public domain per Rick Saada's words:
All terrain tiles, some landscape features, all monsters and objects, and some spell/effect graphics take the form of Windows 3.1 icons and were done by Paul Canniff. Multi-tile graphics, such as ball spells and town buildings, are bitmaps included in the executable file. No graphics use colors other than the Windows-standard 16-color palette, plus transparency. They exist in monochrome versions as well, meaning that the game will display well on monochrome monitors.
The map view is identical to the playing-field view, except for scaling to fit on one screen. A simplified map view is available to improve performance on slower computers. The latter functionality also presents a cleaner display, as the aforementioned scaling routine does not always work correctly.
Computer Gaming World rated the gameplay as good and the graphics simple but effective, while noticing the lack of audio, but regarded the game itself enjoyable. | https://en.wikipedia.org/wiki?curid=6023 |
Calvinism
Calvinism (also called the Reformed tradition, Reformed Christianity, Reformed Protestantism, or the Reformed faith) is a major branch of Protestantism that follows the theological tradition and forms of Christian practice set down by John Calvin and other Reformation-era theologians.
Calvinists broke from the Roman Catholic Church in the 16th century. Calvinists differ from Lutherans (another major branch of the Reformation) on the real presence of Christ in the Eucharist, theories of worship, and the use of God's law for believers, among other things. The term "Calvinism" can be misleading, because the religious tradition which it denotes has always been diverse, with a wide range of influences rather than a single founder; however almost all of them drew heavily from the writings of Augustine of Hippo a millennium prior. In the context of the Reformation, Huldrych Zwingli began the Reformed tradition in 1519 in the city of Zürich. His followers were instantly labeled "Zwinglians", consistent with the Catholic practice of naming heresy after its founder. Very soon, Zwingli was joined by Martin Bucer, Wolfgang Capito, William Farel, Johannes Oecolampadius and other early Reformed thinkers.
The namesake of the movement, French reformer John Calvin, renounced Roman Catholicism and embraced Protestant views in the late 1520s or early 1530s, as the earliest notions of later Reformed tradition were already espoused by Huldrych Zwingli. The movement was first called "Calvinism", referring to John Calvin, by Lutherans who opposed it. Many within the tradition find it either an indescriptive or an inappropriate term and would prefer the word "Reformed" to be used instead. The most important Reformed theologians include Calvin, Zwingli, Martin Bucer, William Farel, Heinrich Bullinger, Peter Martyr Vermigli, Theodore Beza, and John Knox. In the twentieth century, Abraham Kuyper, Herman Bavinck, B. B. Warfield, J. Gresham Machen, Karl Barth, Martyn Lloyd-Jones, Cornelius Van Til, Gordon Clark, and R. C. Sproul were influential. Contemporary Reformed theologians include J. I. Packer, John MacArthur, Timothy J. Keller, David Wells, and Michael Horton.
The Reformed tradition is largely represented by the Continental Reformed, Presbyterian, Evangelical Anglican, Congregationalist, and Reformed Baptist denominational families. Several forms of ecclesiastical polity are exercised by a group of Reformed churches, including presbyterian, congregationalist, and some episcopalian. The biggest Reformed association is the World Communion of Reformed Churches with more than 100 million members in 211 member denominations around the world. There are more conservative Reformed federations such as the World Reformed Fellowship and the International Conference of Reformed Churches, as well as independent churches.
Calvinism is named after John Calvin. It was first used by a Lutheran theologian in 1552. It was a common practice of the Roman Catholic Church to name what it viewed as heresy after its founder. Nevertheless, the term first came out of Lutheran circles. Calvin denounced the designation himself:
Despite its negative connotation, this designation became increasingly popular in order to distinguish Calvinists from Lutherans and from newer Protestant branches that emerged later. The vast majority of churches that trace their history back to Calvin (including Presbyterians, Congregationalists, and a row of other Calvinist churches) do not use it themselves, since the designation "Reformed" is more generally accepted and preferred, especially in the English-speaking world. Moreover, these churches claim to be—in accordance with John Calvin's own words—"renewed accordingly with the true order of gospel".
Since the Arminian controversy, the Reformed tradition—as a branch of Protestantism distinguished from Lutheranism—divided into two separate groups: Arminians and Calvinists. However, it is now rare to call Arminians a part of the Reformed tradition, with the majority of Arminians today being members of the Methodist Churches and General Baptist Churches. While the Reformed theological tradition addresses all of the traditional topics of Christian theology, the word "Calvinism" is sometimes used to refer to particular Calvinist views on soteriology and predestination, which are summarized in part by the Five Points of Calvinism. Some have also argued that Calvinism as a whole stresses the sovereignty or rule of God in all things including salvation.
First-generation Reformed theologians include Huldrych Zwingli (1484–1531), Martin Bucer (1491–1551), Wolfgang Capito (1478–1541), John Oecolampadius (1482–1531), and Guillaume Farel (1489–1565). These reformers came from diverse academic backgrounds, but later distinctions within Reformed theology can already be detected in their thought, especially the priority of scripture as a source of authority. Scripture was also viewed as a unified whole, which led to a covenantal theology of the sacraments of baptism and the Lord's Supper as visible signs of the covenant of grace. Another Reformed distinctive present in these theologians was their denial of the bodily presence of Christ in the Lord's supper. Each of these theologians also understood salvation to be by grace alone, and affirmed a doctrine of particular election (the teaching that some people are chosen by God for salvation). Martin Luther and his successor Philipp Melanchthon were undoubtedly significant influences on these theologians, and to a larger extent later Reformed theologians. The doctrine of justification by faith alone was a direct inheritance from Luther.
John Calvin (1509–64), Heinrich Bullinger (1504–75), Wolfgang Musculus (1497–1563), Peter Martyr Vermigli (1500–62), and Andreas Hyperius (1511–64) belong to the second generation of Reformed theologians. Calvin's "Institutes of the Christian Religion" (1536–59) was one of the most influential theologies of the era. Toward the middle of the 16th century, the Reformed began to commit their beliefs to confessions of faith, which would shape the future definition of the Reformed faith. The 1549 "Consensus Tigurinus" brought together those who followed Zwingli and Bullinger's memorialist theology of the Lord's supper, which taught that the supper simply serves as a reminder of Christ's death, and Calvin's view that the supper serves as a means of grace with Christ actually present, though spiritually rather than bodily. The document demonstrates the diversity as well as unity in early Reformed theology. The remainder of the 16th century saw an explosion of confessional activity. The stability and breadth of Reformed theology during this period stand in marked contrast to the bitter controversy experienced by Lutherans prior to the 1579 Formula of Concord.
Due to Calvin's missionary work in France, his programme of reform eventually reached the French-speaking provinces of the Netherlands. Calvinism was adopted in the Electorate of the Palatinate under Frederick III, which led to the formulation of the Heidelberg Catechism in 1563, and in Navarre by Jeanne d'Albret. This and the Belgic Confession were adopted as confessional standards in the first synod of the Dutch Reformed Church in 1571. Leading divines, either Calvinist or those sympathetic to Calvinism, settled in England (Martin Bucer, Peter Martyr, and Jan Łaski) and Scotland (John Knox). During the English Civil War, the Calvinistic Puritans produced the Westminster Confession, which became the confessional standard for Presbyterians in the English-speaking world. Having established itself in Europe, the movement continued to spread to other parts of the world, including North America, South Africa, and Korea.
Calvin did not live to see the foundation of his work grow into an international movement; but his death allowed his ideas to break out of their city of origin, to succeed far beyond their borders, and to establish their own distinct character.
Although much of Calvin's work was in Geneva, his publications spread his ideas of a "correctly" Reformed church to many parts of Europe. In Switzerland, some cantons are still Reformed, and some are Catholic. Calvinism became the theological system of the majority in Scotland (see John Knox), the Netherlands (see William Ames, T. J. Frelinghuysen and Wilhelmus à Brakel), some communities in Flanders, and parts of Germany (especially these adjacent to the Netherlands) in the Palatinate, Kassel and Lippe with the likes of Olevianus and his colleague Zacharias Ursinus. In Hungary and the then-independent Transylvania, Calvinism was a significant religion. In the 16th century, the Reformation gained many supporters in Eastern Hungary and Hungarian-populated regions in Transylvania. In these parts, the Reformed nobles protected the faith. Almost all Transylvanian dukes were Reformed. Today there are about 3.5 million Hungarian Reformed people worldwide. It was influential in France, Lithuania and Poland before being mostly erased due to the counter-reformational activities taken up by the monarch in each country. In Poland disconnected from Calvinism the Polish stream called Polish Brethren. This steam was started on January 22, 1556, when Piotr of Goniądz (Peter Gonesius), a Polish student, spoke out against the doctrine of the Trinity during the general synod of the Reformed (Calvinist) churches of Poland held in the village of Secemin. Calvinism gained some popularity in Scandinavia, especially Sweden, but was rejected in favor of Lutheranism after the Synod of Uppsala in 1593.
Most settlers in the American Mid-Atlantic and New England were Calvinists, including the English Puritans, the French Huguenots and Dutch settlers of New Amsterdam (New York), and the Scotch-Irish Presbyterians of the Appalachian back country. Nonconforming Protestants, Puritans, Separatists, Independents, English religious groups coming out of the English Civil War, and other English dissenters not satisfied with the degree to which the Church of England had been reformed, held overwhelmingly Reformed views. They are often cited among the primary founders of the United States of America. Dutch and French Huguenot Calvinist settlers were also the first European colonizers of South Africa, beginning in the 17th century, who became known as Boers or Afrikaners.
Sierra Leone was largely colonized by Calvinist settlers from Nova Scotia, who were largely Black Loyalists, blacks who had fought for the British during the American War of Independence. John Marrant had organized a congregation there under the auspices of the Huntingdon Connection. Some of the largest Calvinist communions were started by 19th- and 20th-century missionaries. Especially large are those in Indonesia, Korea and Nigeria. In South Korea there are 20,000 Presbyterian congregations with about 9–10 million church members, scattered in more than 100 Presbyterian denominations. In South Korea, Presbyterianism is the largest Christian denomination.
A 2011 report of the Pew Forum on Religious and Public Life estimated that members of Presbyterian or Reformed churches make up 7% of the estimated 801 million Protestants globally, or approximately 56 million people. Though the broadly defined Reformed faith is much larger, as it constitutes Congregationalist (0.5%), most of the United and uniting churches (unions of different denominations) (7.2%) and most likely some of the other Protestant denominations (38.2%). All three are distinct categories from Presbyterian or Reformed (7%) in this report.
The Reformed family of churches is one of the largest Christian denominations. According to adherents.com the Reformed/Presbyterian/Congregational/United churches represent 75 million believers worldwide.
The World Communion of Reformed Churches, which includes some United Churches (most of these are primarily Reformed; see "Uniting and united churches" for details), has 80 million believers. WCRC is the third largest Christian communion in the world, after the Roman Catholic Church and the Eastern Orthodox Churches.
Many conservative Reformed churches which are strongly Calvinistic formed the World Reformed Fellowship which has about 70 member denominations. Most are not part of the World Communion of Reformed Churches because of its ecumenical attire. The International Conference of Reformed Churches is another conservative association.
Church of Tuvalu is the only officially established state church in the Calvinist tradition in the world.
Reformed theologians believe that God communicates knowledge of himself to people through the Word of God. People are not able to know anything about God except through this self-revelation. Speculation about anything which God has not revealed through his Word is not warranted. The knowledge people have of God is different from that which they have of anything else because God is infinite, and finite people are incapable of comprehending an infinite being. While the knowledge revealed by God to people is never incorrect, it is also never comprehensive.
According to Reformed theologians, God's self-revelation is always through his son Jesus Christ, because Christ is the only mediator between God and people. Revelation of God through Christ comes through two basic channels. The first is creation and providence, which is God's creating and continuing to work in the world. This action of God gives everyone knowledge about God, but this knowledge is only sufficient to make people culpable for their sin; it does not include knowledge of the gospel. The second channel through which God reveals himself is redemption, which is the gospel of salvation from condemnation which is punishment for sin.
In Reformed theology, the Word of God takes several forms. Jesus Christ himself is the Word Incarnate. The prophecies about him said to be found in the Old Testament and the ministry of the apostles who saw him and communicated his message are also the Word of God. Further, the preaching of ministers about God is the very Word of God because God is considered to be speaking through them. God also speaks through human writers in the Bible, which is composed of texts set apart by God for self-revelation. Reformed theologians emphasize the Bible as a uniquely important means by which God communicates with people. People gain knowledge of God from the Bible which cannot be gained in any other way.
Reformed theologians affirm that the Bible is true, but differences emerge among them over the meaning and extent of its truthfulness. Conservative followers of the Princeton theologians take the view that the Bible is true and inerrant, or incapable of error or falsehood, in every place. This view is very similar to that of Catholic orthodoxy as well as modern Evangelicalism. Another view, influenced by the teaching of Karl Barth and neo-orthodoxy, is found in the Presbyterian Church (U.S.A.)'s Confession of 1967. Those who take this view believe the Bible to be the primary source of our knowledge of God, but also that some parts of the Bible may be false, not witnesses to Christ, and not normative for today's church. In this view, Christ is the revelation of God, and the scriptures witness to this revelation rather than being the revelation itself.
Reformed theologians use the concept of covenant to describe the way God enters fellowship with people in history. The concept of covenant is so prominent in Reformed theology that Reformed theology as a whole is sometimes called "covenant theology". However, sixteenth and seventeenth-century theologians developed a particular theological system called "covenant theology" or "federal theology" which many conservative Reformed churches continue to affirm today. This framework orders God's life with people primarily in two covenants: the covenant of works and the covenant of grace. The covenant of works is made with Adam and Eve in the Garden of Eden. The terms of the covenant are that God provides a blessed life in the garden on condition that Adam and Eve obey God's law perfectly. Because Adam and Eve broke the covenant by eating the forbidden fruit, they became subject to death and were banished from the garden. This sin was passed down to all mankind because all people are said to be in Adam as a covenantal or "federal" head. Federal theologians usually infer that Adam and Eve would have gained immortality had they obeyed perfectly.
A second covenant, called the covenant of grace, is said to have been made immediately following Adam and Eve's sin. In it, God graciously offers salvation from death on condition of faith in God. This covenant is administered in different ways throughout the Old and New Testaments, but retains the substance of being free of a requirement of perfect obedience.
Through the influence of Karl Barth, many contemporary Reformed theologians have discarded the covenant of works, along with other concepts of federal theology. Barth saw the covenant of works as disconnected from Christ and the gospel, and rejected the idea that God works with people in this way. Instead, Barth argued that God always interacts with people under the covenant of grace, and that the covenant of grace is free of all conditions whatsoever. Barth's theology and that which follows him has been called "monocovenantal" as opposed to the "bi-covenantal" scheme of classical federal theology. Conservative contemporary Reformed theologians, such as John Murray, have also rejected the idea of covenants based on law rather than grace. Michael Horton, however, has defended the covenant of works as combining principles of law and love.
For the most part, the Reformed tradition did not modify the medieval consensus on the doctrine of God. God's character is described primarily using three adjectives: eternal, infinite, and unchangeable. Reformed theologians such as Shirley Guthrie have proposed that rather than conceiving of God in terms of his attributes and freedom to do as he pleases, the doctrine of God is to be based on God's work in history and his freedom to live with and empower people.
Traditionally, Reformed theologians have also followed the medieval tradition going back to before the early church councils of Nicaea and Chalcedon on the doctrine of the Trinity. God is affirmed to be one God in three persons: Father, Son, and Holy Spirit. The Son (Christ) is held to be eternally begotten by the Father and the Holy Spirit eternally proceeding from the Father and Son. However, contemporary theologians have been critical of aspects of Western views here as well. Drawing on the Eastern tradition, these Reformed theologians have proposed a "social trinitarianism" where the persons of the Trinity only exist in their life together as persons-in-relationship. Contemporary Reformed confessions such as the Barmen Confession and Brief Statement of Faith of the Presbyterian Church (USA) have avoided language about the attributes of God and have emphasized his work of reconciliation and empowerment of people. Feminist theologian Letty Russell used the image of partnership for the persons of the Trinity. According to Russell, thinking this way encourages Christians to interact in terms of fellowship rather than reciprocity. Conservative Reformed theologian Michael Horton, however, has argued that social trinitarianism is untenable because it abandons the essential unity of God in favor of a community of separate beings.
Reformed theologians affirm the historic Christian belief that Christ is eternally one person with a divine and a human nature. Reformed Christians have especially emphasized that Christ truly became human so that people could be saved. Christ's human nature has been a point of contention between Reformed and Lutheran Christology. In accord with the belief that finite humans cannot comprehend infinite divinity, Reformed theologians hold that Christ's human body cannot be in multiple locations at the same time. Because Lutherans believe that Christ is bodily present in the Eucharist, they hold that Christ is bodily present in many locations simultaneously. For Reformed Christians, such a belief denies that Christ actually became human. Some contemporary Reformed theologians have moved away from the traditional language of one person in two natures, viewing it as unintelligible to contemporary people. Instead, theologians tend to emphasize Jesus' context and particularity as a first-century Jew.
John Calvin and many Reformed theologians who followed him describe Christ's work of redemption in terms of three offices: prophet, priest, and king. Christ is said to be a prophet in that he teaches perfect doctrine, a priest in that he intercedes to the Father on believers' behalf and offered himself as a sacrifice for sin, and a king in that he rules the church and fights on believers' behalf. The threefold office links the work of Christ to God's work in ancient Israel. Many, but not all, Reformed theologians continue to make use of the threefold office as a framework because of its emphasis on the connection of Christ's work to Israel. They have, however, often reinterpreted the meaning of each of the offices. For example, Karl Barth interpreted Christ's prophetic office in terms of political engagement on behalf of the poor.
Christians believe Jesus' death and resurrection makes it possible for believers to attain forgiveness for sin and reconciliation with God through the atonement. Reformed Protestants generally subscribe to a particular view of the atonement called penal substitutionary atonement, which explains Christ's death as a sacrificial payment for sin. Christ is believed to have died in place of the believer, who is accounted righteous as a result of this sacrificial payment.
In Christian theology, people are created good and in the image of God but have become corrupted by sin, which causes them to be imperfect and overly self-interested. Reformed Christians, following the tradition of Augustine of Hippo, believe that this corruption of human nature was brought on by Adam and Eve's first sin, a doctrine called original sin. Although earlier Christian authors taught the elements of physical death, moral weakness, and a sin propensity within original sin, Augustine was the first Christian to add the concept of inherited guilt ("reatus") from Adam whereby every infant is born eternally damned and humans lack any residual ability to respond to God. Reformed theologians emphasize that this sinfulness affects all of a person's nature, including their will. This view, that sin so dominates people that they are unable to avoid sin, has been called total depravity. In colloquial English, the term "total depravity" can be easily misunderstood to mean that people are absent of any goodness or unable to do any good. However the Reformed teaching is actually that while people continue to bear God's image and may do things that appear outwardly good, their sinful intentions affect all of their nature and actions so that they are not pleasing to God.
Some contemporary theologians in the Reformed tradition, such as those associated with the PC(USA)'s Confession of 1967, have emphasized the social character of human sinfulness. These theologians have sought to bring attention to issues of environmental, economic, and political justice as areas of human life that have been affected by sin.
Reformed theologians, along with other Protestants, believe salvation from punishment for sin is to be given to all those who have faith in Christ. Faith is not purely intellectual, but involves trust in God's promise to save. Protestants do not hold there to be any other requirement for salvation, but that faith alone is sufficient.
Justification is the part of salvation where God pardons the sin of those who believe in Christ. It is historically held by Protestants to be the most important article of Christian faith, though more recently it is sometimes given less importance out of ecumenical concerns. People are not on their own able even to fully repent of their sin or prepare themselves to repent because of their sinfulness. Therefore, justification is held to arise solely from God's free and gracious act.
Sanctification is the part of salvation in which God makes the believer holy, by enabling them to exercise greater love for God and for other people. The good works accomplished by believers as they are sanctified are considered to be the necessary outworking of the believer's salvation, though they do not cause the believer to be saved. Sanctification, like justification, is by faith, because doing good works is simply living as the son of God one has become.
Reformed theologians teach that sin so affects human nature that they are unable even to exercise faith in Christ by their own will. While people are said to retain will, in that they willfully sin, they are unable not to sin because of the corruption of their nature due to original sin. Reformed Christians believe that God predestined some people to be saved and others were predestined to eternal damnation. This choice by God to save some is held to be unconditional and not based on any characteristic or action on the part of the person chosen. This view is opposed to the Arminian view that God's choice of whom to save is conditional or based on his foreknowledge of who would respond positively to God.
Karl Barth reinterpreted the Reformed doctrine of predestination to apply only to Christ. Individual people are only said to be elected through their being in Christ. Reformed theologians who followed Barth, including Jürgen Moltmann, David Migliore, and Shirley Guthrie, have argued that the traditional Reformed concept of predestination is speculative and have proposed alternative models. These theologians claim that a properly trinitarian doctrine emphasizes God's freedom to love all people, rather than choosing some for salvation and others for damnation. God's justice towards and condemnation of sinful people is spoken of by these theologians as out of his love for them and a desire to reconcile them to himself.
Most objections to and attacks on Calvinism focus on the "five points of Calvinism", also called the doctrines of grace, and remembered by the mnemonic "TULIP". The five points are popularly said to summarize the Canons of Dort; however, there is no historical relationship between them, and some scholars argue that their language distorts the meaning of the Canons, Calvin's theology, and the theology of 17th-century Calvinistic orthodoxy, particularly in the language of total depravity and limited atonement. | https://en.wikipedia.org/wiki?curid=6024 |
Countable set
In mathematics, a countable set is a set with the same cardinality (number of elements) as some subset of the set of natural numbers. A countable set is either a finite set or a countably infinite set. Whether finite or infinite, the elements of a countable set can always be counted one at a time and, although the counting may never finish, every element of the set is associated with a unique natural number.
Some authors use countable set to mean "countably infinite" alone. To avoid this ambiguity, the term "at most countable" may be used when finite sets are included and "countably infinite", "enumerable", or "denumerable" otherwise.
Georg Cantor introduced the term "countable set", contrasting sets that are countable with those that are "uncountable" (i.e., "nonenumerable" or "nondenumerable"). Today, countable sets form the foundation of a branch of mathematics called "discrete mathematics".
A set is "countable" if there exists an injective function from to the natural numbers }.
If such an can be found that is also surjective (and therefore bijective), then is called "countably infinite."
In other words, a set is "countably infinite" if it has one-to-one correspondence with the natural number set, .
This terminology is not universal. Some authors use countable to mean what is here called "countably infinite," and do not include finite sets.
Alternative (equivalent) formulations of the definition in terms of a bijective function or a surjective function can also be given. See below.
In 1874, in his first set theory article, Cantor proved that the set of real numbers is uncountable, thus showing that not all infinite sets are countable. In 1878, he used one-to-one correspondences to define and compare cardinalities. In 1883, he extended the natural numbers with his infinite ordinals, and used sets of ordinals to produce an infinity of sets having different infinite cardinalities.
A "set" is a collection of "elements", and may be described in many ways. One way is simply to list all of its elements; for example, the set consisting of the integers 3, 4, and 5 may be denoted {3, 4, 5}. This is only effective for small sets, however; for larger sets, this would be time-consuming and error-prone. Instead of listing every single element, sometimes an ellipsis ("...") is used, if the writer believes that the reader can easily guess what is missing; for example, {1, 2, 3, ..., 100} presumably denotes the set of integers from 1 to 100. Even in this case, however, it is still "possible" to list all the elements, because the set is "finite".
Some sets are "infinite"; these sets have more than "n" elements for any integer "n". For example, the set of natural numbers, denotable by {0, 1, 2, 3, 4, 5, ...}, has infinitely many elements, and we cannot use any normal number to give its size. Nonetheless, it turns out that infinite sets do have a well-defined notion of size (or more properly, of "cardinality", which is the technical term for the number of elements in a set), and not all infinite sets have the same cardinality.
To understand what this means, we first examine what it "does not" mean. For example, there are infinitely many odd integers, infinitely many even integers, and (hence) infinitely many integers overall. However, it turns out that the number of even integers, which is the same as the number of odd integers, is also the same as the number of integers overall. This is because we can arrange things such that for every integer, there is a distinct even integer: ... −2→−4, −1→−2, 0→0, 1→2, 2→4, ...; or, more generally, "n"→2"n", see picture. What we have done here is arrange the integers and the even integers into a "one-to-one correspondence" (or "bijection"), which is a function that maps between two sets such that each element of each set corresponds to a single element in the other set.
However, not all infinite sets have the same cardinality. For example, Georg Cantor (who introduced this concept) demonstrated that the real numbers cannot be put into one-to-one correspondence with the natural numbers (non-negative integers), and therefore that the set of real numbers has a greater cardinality than the set of natural numbers.
A set is "countable" if: (1) it is finite, or (2) it has the same cardinality (size) as the set of natural numbers. Equivalently, a set is "countable" if it has the same cardinality as some subset of the set of natural numbers. Otherwise, it is "uncountable".
By definition a set "S" is "countable" if there exists an injective function "f" : "S" → N from "S" to the natural numbers N = {0, 1, 2, 3, ...}.
It might seem natural to divide the sets into different classes: put all the sets containing one element together; all the sets containing two elements together; ...; finally, put together all infinite sets and consider them as having the same size.
This view is not tenable, however, under the natural definition of size.
To elaborate this we need the concept of a bijection. Although a "bijection" seems a more advanced concept than a number, the usual development of mathematics in terms of set theory defines functions before numbers, as they are based on much simpler sets. This is where the concept of a bijection comes in: define the correspondence
Since every element of {"a", "b", "c"} is paired with "precisely one" element of {1, 2, 3}, "and" vice versa, this defines a bijection.
We now generalize this situation and "define" two sets as of the same size if (and only if) there is a bijection between them. For all finite sets this gives us the usual definition of "the same size". What does it tell us about the size of infinite sets?
Consider the sets "A" = {1, 2, 3, ... }, the set of positive integers and "B" = {2, 4, 6, ... }, the set of even positive integers. We claim that, under our definition, these sets have the same size, and that therefore "B" is countably infinite. Recall that to prove this we need to exhibit a bijection between them. But this is easy, using "n" ↔ 2"n", so that
As in the earlier example, every element of A has been paired off with precisely one element of B, and vice versa. Hence they have the same size. This is an example of a set of the same size as one of its proper subsets, which is impossible for finite sets.
Likewise, the set of all ordered pairs of natural numbers is countably infinite, as can be seen by following a path like the one in the picture: The resulting mapping is like this:
This mapping covers all such ordered pairs.
If each pair is treated as the numerator and denominator of a vulgar fraction, then for every positive fraction, we can come up with a distinct number corresponding to it. This representation includes also the natural numbers, since every natural number is also a fraction "N"/1. So we can conclude that there are exactly as many positive rational numbers as there are positive integers. This is true also for all rational numbers, as can be seen below.
Theorem: The Cartesian product of finitely many countable sets is countable.
This form of triangular mapping recursively generalizes to vectors of finitely many natural numbers by repeatedly mapping the first two elements to a natural number. For example, (0,2,3) maps to (5,3), which maps to 39.
Sometimes more than one mapping is useful: the set to be shown to be countably infinite is mapped onto another set, then this other set is mapped onto to the natural numbers. For example, the positive rational numbers can easily be mapped to (a subset of) the pairs of natural numbers because "p"/"q "maps to ("p", "q").
What about infinite subsets of countably infinite sets? Do these have fewer elements than N?
Theorem: Every subset of a countable set is countable. In particular, every infinite subset of a countably infinite set is countably infinite.
For example, the set of prime numbers is countable, by mapping the "n"-th prime number to "n":
What about sets being naturally "larger than" N? For instance, Z the set of all integers or Q, the set of all rational numbers, which intuitively may seem much bigger than N. But looks can be deceiving, for we assert:
Theorem: Z (the set of all integers) and Q (the set of all rational numbers) are countable.
In a similar manner, the set of algebraic numbers is countable.
These facts follow easily from a result that many individuals find non-intuitive.
Theorem: Any finite union of countable sets is countable.
With the foresight of knowing that there are uncountable sets, we can wonder whether or not this last result can be pushed any further. The answer is "yes" and "no", we can extend it, but we need to assume a new axiom to do so.
Theorem: (Assuming the axiom of countable choice) The union of countably many countable sets is countable.
For example, given countable sets a, b, c, ...
Using a variant of the triangular enumeration we saw above:
This only works if the sets a, b, c, ... are disjoint. If not, then the union is even smaller and is therefore also countable by a previous theorem.
We need the axiom of countable choice to index "all" the sets a, b, c, ... simultaneously.
Theorem: The set of all finite-length sequences of natural numbers is countable.
This set is the union of the length-1 sequences, the length-2 sequences, the length-3 sequences, each of which is a countable set (finite Cartesian product). So we are talking about a countable union of countable sets, which is countable by the previous theorem.
Theorem: The set of all finite subsets of the natural numbers is countable.
The elements of any finite subset can be ordered into a finite sequence. There are only countably many finite sequences, so also there are only countably many finite subsets.
The following theorem gives equivalent formulations in terms of a bijective function or a surjective function. A proof of this result can be found in Lang's text.
(Basic) Theorem: Let "S" be a set. The following statements are equivalent:
Corollary: Let "S" and "T" be sets.
Cantor's theorem asserts that if "A" is a set and "P"("A") is its power set, i.e. the set of all subsets of "A", then there is no surjective function from "A" to "P"("A"). A proof is given in the article Cantor's theorem. As an immediate consequence of this and the Basic Theorem above we have:
Proposition: The set "P"(N) is not countable; i.e. it is uncountable.
For an elaboration of this result see Cantor's diagonal argument.
The set of real numbers is uncountable (see Cantor's first uncountability proof), and so is the set of all infinite sequences of natural numbers.
The proofs of the statements in the above section rely upon the existence of functions with certain properties. This section presents functions commonly used in this role, but not the verifications that these functions have the required properties. The Basic Theorem and its Corollary are often used to simplify proofs. Observe that in that theorem can be replaced with any countably infinite set.
Proposition: Any finite set is countable.
Proof: Let be such a set. Two cases are to be considered: Either is empty or it isn't. 1.) The empty set is even itself a subset of the natural numbers, so it is countable. 2.) If is nonempty and finite, then by definition of finiteness there is a bijection between and the set {1, 2, ..., } for some positive natural number . This function is an injection from into .
Proposition: Any subset of a countable set is countable.
Proof: The restriction of an injective function to a subset of its domain is still injective.
Proposition: If is a countable set then } is countable.
Proof: If there is nothing to be shown. Otherwise let be an injection. Define by and
Proposition: If and are countable sets then is countable.
Proof: Let and be injections. Define a new injection by if is in and if is in but not in .
Proposition: The Cartesian product of two countable sets and is countable.
Proof: Observe that is countable as a consequence of the definition because the function given by is injective. It then follows from the Basic Theorem and the Corollary that the Cartesian product of any two countable sets is countable. This follows because if and are countable there are surjections and . So
is a surjection from the countable set to the set and the Corollary implies is countable. This result generalizes to the Cartesian product of any finite collection of countable sets and the proof follows by induction on the number of sets in the collection.
Proposition: The integers are countable and the rational numbers are countable.
Proof: The integers are countable because the function given by if is non-negative and if is negative, is an injective function. The rational numbers are countable because the function given by is a surjection from the countable set to the rationals .
Proposition: The algebraic numbers are countable.
Proof: Per definition, every algebraic number (including complex numbers) is a root of a polynomial with integer coefficients. Given an algebraic number formula_1, let formula_2 be a polynomial with integer coefficients such that formula_1 is the "k"th root of the polynomial, where the roots are sorted by absolute value from small to big, then sorted by argument from small to big. We can define an injection (i. e. one-to-one) function given by formula_4, while formula_5 is the "n"-th prime.
Proposition: If is a countable set for each in then the union of all is also countable.
Proof: This is a consequence of the fact that for each there is a surjective function and hence the function
given by is a surjection. Since is countable, the Corollary implies that the union is countable. We use the axiom of countable choice in this proof to pick for each in a surjection from the non-empty collection of surjections from to .
A topological proof for the uncountability of the real numbers is described at finite intersection property.
If there is a set that is a standard model (see inner model) of ZFC set theory, then there is a minimal standard model ("see" Constructible universe). The Löwenheim–Skolem theorem can be used to show that this minimal model is countable. The fact that the notion of "uncountability" makes sense even in this model, and in particular that this model "M" contains elements that are:
was seen as paradoxical in the early days of set theory, see Skolem's paradox.
The minimal standard model includes all the algebraic numbers and all effectively computable transcendental numbers, as well as many other kinds of numbers.
Countable sets can be totally ordered in various ways, e.g.:
In both examples of well orders here, any subset has a "least element"; and in both examples of non-well orders, "some" subsets do not have a "least element".
This is the key definition that determines whether a total order is also a well order. | https://en.wikipedia.org/wiki?curid=6026 |
Cahn–Ingold–Prelog priority rules
The Cahn–Ingold–Prelog (CIP) sequence rules, named for organic chemists Robert Sidney Cahn, Christopher Kelk Ingold, and Vladimir Prelog — alternatively termed the CIP priority rules, "system", or "conventions" — are a standard process used in organic chemistry to completely and unequivocally name a stereoisomer of a molecule. The purpose of the CIP system is to assign an to each stereocenter and an "E" or "Z" descriptor to each double bond so that the configuration of the entire molecule can be specified uniquely by including the descriptors in its systematic name. A molecule may contain any number of stereocenters and any number of double bonds, and each usually gives rise to two possible isomers. A molecule with an integer "n" describing the number of its stereogenic centers will usually have 2"n" stereoisomers, and 2"n"−1 diastereomers each having an associated pair of enantiomers. The CIP sequence rules contribute to the precise naming of every stereoisomer of every organic and organometallic molecule with all atoms of ligancy of fewer than 4 (but including ligancy of 6 as well, this term referring to the "number of neighboring atoms" bonded to a center).
The key article setting out the CIP sequence rules was published in 1966, and was followed by further refinements, before it was incorporated into the rules of the International Union of Pure and Applied Chemistry (IUPAC), the official body that defines organic nomenclature. The IUPAC presentation of the rules constitute the official, formal standard for their use, and it notes that "the method has been developed to cover all compounds with ligancy up to 4... and… [extended to the case of] ligancy 6… [as well as] for all configurations and conformations of such compounds." Nevertheless, though the IUPAC documentation presents a thorough introduction, it includes the caution that "it is essential to study the original papers, especially the 1966 paper, before using the sequence rule for other than fairly simple cases."
A recent paper argues for changes to the specific rules to accommodate for certain structures, after they found deficiencies in some of the CIP rules (sequence rules 1b and 2).
The steps for naming molecules using the CIP system are often presented as:
and "E"/"Z" descriptors are assigned by using a system for ranking priority of the groups attached to each stereocenter. This procedure, often known as "the sequence rules", is the heart of the CIP system.
If two groups differ only in isotopes, then the larger atomic mass is used to set the priority.
If an atom A is double-bonded to an atom B, A is treated as being singly bonded to two atoms: B and a "ghost atom" that is a duplicate of B (has the same atomic number) but is not attached to anything except A. When B is replaced with a list of attached atoms, A itself, but not its "ghost", is excluded in accordance with the general principle of not doubling back along a bond that has just been followed. A triple bond is handled the same way except that A and B are each connected to two ghost atoms of the other.
If two substituents on an atom are geometric isomers of each other, the "Z"-isomer has higher priority than the "E"-isomer.
To handle a molecule containing one or more cycles, one must first expand it into a tree (called a hierarchical digraph) by traversing bonds in all possible paths starting at the stereocenter. When the traversal encounters an atom through which the current path has already passed, a ghost atom is generated in order to keep the tree finite. A single atom of the original molecule may appear in many places (some as ghosts, some not) in the tree.
After the substituents of a stereocenter have been assigned their priorities, the molecule is oriented in space so that the group with the lowest priority is pointed away from the observer. If the substituents are numbered from 1 (highest priority) to 4 (lowest priority), then the sense of rotation of a curve passing through 1, 2 and 3 distinguishes the stereoisomers. A center with a clockwise sense of rotation is an "R" ("rectus") center and a center with a counterclockwise sense of rotation is an "S" ("sinister") center. The names are derived from the Latin for 'right' and 'left', respectively.
A practical method of determining whether an enantiomer is "R" or "S" is by using the right-hand rule: one wraps the molecule with the fingers in the direction . If the thumb points in the direction of the fourth substituent, the enantiomer is "R"; otherwise, it is "S".
It is possible in rare cases that two substituents on an atom differ only in their absolute configuration ("R" or "S"). If the relative priorities of these substituents need to be established, "R" takes priority over "S". When this happens, the descriptor of the stereocenter is a lowercase letter ("r" or "s") instead of the uppercase letter normally used.
For alkenes and similar double bonded molecules, the same prioritizing process is followed for the substituents. In this case, it is the placing of the two highest priority substituents with respect to the double bond which matters. If both high priority substituents are on the same side of the double bond, i.e. in the "cis" configuration, then the stereoisomer is assigned a "Z" ("zusammen") . If by contrast they are in a "trans" configuration, then the stereoisomer is assigned an "E" ("entgegen"). In this case the identifying letters are derived from German for 'together' and 'opposite', respectively.
The following are examples of application of the nomenclature.
If a compound has more than one stereocenter each center is denoted by either "R" or "S". For example, ephedrine exists with both (1"R",2"S") and (1"S",2"R") configuration, known as enantiomers. This compound also exists with a (1"R",2"R") and (1"S",2"S") configuration. The last two stereoisomers are not ephedrine, but pseudoephedrine. All isomers are 2-methylamino-1-phenyl-1-propanol in systematic nomenclature. Pseudoephedrine is chemically distinct from ephedrine with only the three-dimensional configuration in space, as notated by the Cahn–Ingold–Prelog rules. The two compounds, ephedrine and pseudoephedrine, are diastereomers, or stereoisomers that are not enantiomers. They have different names because, as diastereomers, they have different chemical properties.
In pairs of enantiomers, all descriptors are opposite: ("R","R") and ("S","S"), or ("R","S") and ("S","R"). Diastereomers have one descriptor in common: ("R","S") and ("R","R"), or ("S","R") and ("S","S"). This holds true for compounds with more than two stereocenters; if at least one descriptor is the same in both pairs, the compounds are diastereomers. If all the stereocenters are opposite, they are enantiomers.
The relative configuration of two stereoisomers may be denoted by the descriptors "R" and "S" with an asterisk (*). ("R"*,"R"*) means two centers having identical configurations, ("R","R") or ("S","S"); ("R"*,"S"*) means two centers having opposite configurations, ("R","S") or ("S","R"). To begin, the lowest-numbered (according to IUPAC systematic numbering) stereogenic center is given the "R"* descriptor.
To designate two anomers the relative stereodescriptors alpha (α) and beta (β) are used. In the α anomer the "anomeric carbon atom" and the "reference atom" do have opposite configurations ("R","S") or ("S","R"), whereas in the β anomer they are the same ("R","R") or ("S","S").
Stereochemistry also plays a role assigning "faces" to trigonal molecules such as ketones. A nucleophile in a nucleophilic addition can approach the carbonyl group from two opposite sides or faces. When an achiral nucleophile attacks acetone, both faces are identical and there is only one reaction product. When the nucleophile attacks butanone, the faces are not identical ("enantiotopic") and a racemic product results. When the nucleophile is a chiral molecule diastereoisomers are formed. When one face of a molecule is shielded by substituents or geometric constraints compared to the other face the faces are called diastereotopic. The same rules that determine the stereochemistry of a stereocenter ("R" or "S") also apply when assigning the face of a molecular group. The faces are then called the Re"-face and Si"-face. In the example displayed on the right, the compound acetophenone is viewed from the "Re"-face. Hydride addition as in a reduction process from this side will form the ("S")-enantiomer and attack from the opposite "Si"-face will give the ("R")-enantiomer. However, one should note that adding a chemical group to the prochiral center from the "Re"-face will not always lead to an ("S")-stereocenter, as the priority of the chemical group has to be taken into account. That is, the absolute stereochemistry of the product is determined on its own and not by considering which face it was attacked from. In the above-mentioned example, if chloride ("Z" = 17) were added to the prochiral center from the "Re"-face, this would result in an ("R")-enantiomer. | https://en.wikipedia.org/wiki?curid=6034 |
Celibacy
Celibacy (from Latin "cælibatus") is the state of voluntarily being unmarried, sexually abstinent, or both, usually for religious reasons. It is often in association with the role of a religious official or devotee. In its narrow sense, the term "celibacy" is applied only to those for whom the unmarried state is the result of a sacred vow, act of renunciation, or religious conviction. In a wider sense, it is commonly understood to only mean abstinence from sexual activity.
Celibacy has existed in one form or another throughout history, in virtually all the major religions of the world, and views on it have varied. Judaism strongly opposes celibacy. However, the priests of the Essenes, a Jewish sect during the Second Temple period, practised celibacy. The Romans viewed celibacy as an aberration and legislated fiscal penalties against it, with the sole exception granted to the Vestal Virgins. The Islamic attitudes toward celibacy have been complex as well. Several Hadiths indicate that Prophet Muhammad denounced celibacy, but some Sufi orders embrace it.
Classical Hindu culture encouraged asceticism and celibacy in the later stages of life, after one has met one's societal obligations. Jainism, on the other hand, preached complete celibacy even for young monks and considered celibacy to be an essential behavior to attain moksha. Buddhism is similar to Jainism in this respect. There were, however, significant cultural differences in the various areas where Buddhism spread, which affected the local attitudes toward celibacy. It was not well received in China, for example, where other religions movements such as Daoism were opposed to it. A somewhat similar situation existed in Japan, where the Shinto tradition also opposed celibacy. In most native African and American Indian religious traditions, celibacy has been viewed negatively as well, although there were exceptions like periodic celibacy practiced by some Mesoamerican warriors.
The English word "celibacy" derives from the Latin "caelibatus", "state of being unmarried", from Latin , meaning "unmarried". This word derives from two Proto-Indo-European stems, * "alone" and * "living".
The words "abstinence" and "celibacy" are often used interchangeably, but are not necessarily the same thing. Sexual abstinence, also known as "continence", is abstaining from some or all aspects of sexual activity, often for some limited period of time, while celibacy may be defined as a voluntary religious vow not to marry or engage in sexual activity. Asexuality is commonly conflated with celibacy and sexual abstinence, but it is considered distinct from the two, as celibacy and sexual abstinence are behavioral and those who use those terms for themselves are generally motivated by factors such as an individual's personal or religious beliefs.
A. W. Richard Sipe, while focusing on the topic of celibacy in Catholicism, states that "the most commonly assumed definition of "celibate" is simply an unmarried or single person, and celibacy is perceived as synonymous with sexual abstinence or restraint." Sipe adds that even in the relatively uniform milieu of Catholic priests in the United States "there is simply no clear operational definition of celibacy". Elizabeth Abbott commented on the terminology in her "A History of Celibacy" (2001): "I also drafted a definition that discarded the rigidly pedantic and unhelpful distinctions between celibacy, chastity and virginity".
The concept of "new celibacy" was introduced by Gabrielle Brown in her 1980 book "The New Celibacy". In a revised version (1989) of her book, she claims that "abstinence is a response on the outside to what's going on, and celibacy is a response from the inside". According to her definition, celibacy (even short-term celibacy that is pursued for non-religious reasons) is much more than not having sex. It is more intentional than abstinence, and its goal is personal growth and empowerment. This new perspective on celibacy is echoed by several authors including Elizabeth Abbott, Wendy Keller, and Wendy Shalit.
The rule of celibacy in the Buddhist religion, whether Mahayana or Theravada, has a long history. Celibacy was advocated as an ideal rule of life for all monks and nuns by Gautama Buddha, except for Japan where it is not strictly followed due to historical and political developments following the Meiji Restoration. In Japan, celibacy was an ideal among Buddhist clerics for hundreds of years. But violations of clerical celibacy were so common for so long that, finally, in 1872, state laws made marriage legal for Buddhist clerics. Subsequently, ninety percent of Buddhist monks/clerics married. An example is Higashifushimi Kunihide, a prominent Buddhist priest of Japanese royal ancestry who was married and a father whilst serving as a monk for most of his lifetime.
Gautama, later known as the Buddha, is known for his renunciation of his wife, Princess Yasodharā, and son, Rahula. In order to pursue an ascetic life, he needed to renounce aspects of the impermanent world, including his wife and son. Later on both his wife and son joined the ascetic community and are mentioned in the Buddhist texts to have become enlightened. In another sense, a buddhavacana recorded the zen patriarch Vimalakirti as being an advocate of marital continence instead of monastic renunciation, the sutra became somewhat popular due to its brash humour as well as integrating the role of women in laity as well as spiritual life.
In the religious movement of Brahma Kumaris, celibacy is also promoted for peace and to defeat power of lust and also to prepare for life in forthcoming Heaven on earth for 2,500 years when children will be created by the power of the mind even for householders to like holy brother and sister.
In this belief system, celibacy is given the utmost importance. It is said that, as per the direction of the Supreme God those lead a pure and celibate life will be successfully able to conquer the surging vices. The power of celibacy creates an unseen environment of divinity bringing peace, power, purity, prosperity and fortune. Those with the power of celibacy are eligible to claim a bright future of Golden Age of heaven / Paradise. Brahma Kumaris' concept of identifying the self as a soul, different from physical body, is deeply linked to the philosophy of celibacy. It is said that the craving for sex and impure thoughts are the reason for the whole trouble in the universe today. And celibacy is to lead the pure relationship in one's life.
When Jesus discusses marriage, he points out that there is some responsibility for a man marrying a woman (and vice versa). Not having assets of their own, women needed to be protected from the risk of their husbands' putting them on the street at whim. In those times marriage was an economic matter rather than one of love. A woman and her children could easily be rejected. Restriction of divorce was based on the necessity of protecting the woman and her position in society, not necessarily in a religious context, but an economic context. He also points out that there are those "which were made eunuchs of men: and there be eunuchs, which have made themselves eunuchs for the kingdom of heaven's sake", but in the original Greek, the word εὐνοῦχος means "castrated person". It was the custom at the time Jesus lived for priests of some ancient gods and goddesses to be castrated. In the pre-Christian period Vestals, who served the virgin goddess of the hearth, were obliged to forgo marriage, and so were some priests and servants of some ancient deities such as Isis.
There is no commandment in the New Testament that Jesus' disciples have to live in celibacy. The general view on sexuality among the early Jewish Christians was quite positive. Jesus himself does not speak in negative terms of the body in the New Testament. While the Jewish sect of essenes practiced celibacy the general practice of the Jewish community by that time prescribed marriage for everybody, and at an early age. Saint Peter, also known as Simon Peter, the Apostle was married; Jesus healed Simon Peter's mother-in-law (Matt. 8:14), and other apostles and church members among the early Jewish Christians were also married: Paul's personal friends, Priscilla and Aquila (), who were Paul's coworkers, Andronicus of Pannonia (), and Junia (), who were highly regarded among the apostles, Ananias and Sapphira (Ap 5:1), Apphia and Philemon (Phil 1: 1). According to Eusebius Church History ("Historia Ecclesiastica"), Paul the Apostle, also known as Saul of Tarsus, was also married. It was the custom in the Jewish community to marry early.
In his early writings, Paul the Apostle described marriage as a social obligation that has the potential of distracting from Christ. Sex, in turn, is not sinful but natural, and sex within marriage is both proper and necessary. In his later writings, Paul made parallels between the relations between spouses and God's relationship with the church. "Husbands love your wives even as Christ loved the church. Husbands should love their wives as their own bodies" (Ephesians 5:25–28). The early Christians lived in the belief that the End of the World would soon come upon them, and saw no point in planning new families and having children. This was why Paul encouraged both celibate and marital lifestyles among the members of the Corinthian congregation, regarding celibacy as the preferable of the two.
Paul the Apostle emphasized the importance of overcoming the desires of the flesh and saw the state of celibacy being superior to the marriage.
In the Catholic Church, a consecrated virgin, is a woman who has been consecrated by the church to a life of perpetual virginity in the service of God. According to most Christian thought, the first sacred virgin was Mary, the mother of Jesus, who was consecrated by the Holy Spirit during the Annunciation. Tradition also has it that the Apostle Matthew consecrated virgins. A number of early Christian martyrs were women or girls who had given themselves to Christ in perpetual virginity, such as Saint Agnes and Saint Lucy.
The Desert Fathers were Christian hermits, and ascetics who had a major influence on the development of Christianity and celibacy. Paul of Thebes is often credited with being the first hermit monk to go to the desert, but it was Anthony the Great who launched the movement that became the Desert Fathers. Sometime around AD 270, Anthony heard a Sunday sermon stating that perfection could be achieved by selling all of one's possessions, giving the proceeds to the poor, and following Christ.(Matt. 19.21) He followed the advice and made the further step of moving deep into the desert to seek complete solitude.
Over time, the model of Anthony and other hermits attracted many followers, who lived alone in the desert or in small groups. They chose a life of extreme asceticism, renouncing all the pleasures of the senses, rich food, baths, rest, and anything that made them comfortable. Thousands joined them in the desert, mostly men but also a handful of women. Religious seekers also began going to the desert seeking advice and counsel from the early Desert Fathers. By the time of Anthony's death, there were so many men and women living in the desert in celibacy that it was described as "a city" by Anthony's biographer. The first Conciliar document on celibacy of the Western Christian Church (Canon 33 of the Synod of Elvira, ) states that the discipline of celibacy is to refrain from the "use" of marriage, i.e. refrain from having carnal contact with one's spouse.
According to the later St. Jerome (420) celibacy is a moral virtue, consisting of living in the flesh, but outside the flesh, and so being not corrupted by it ("vivere in carne praeter carnem"). Celibacy excludes not only libidinous acts, but also sinful thoughts or desires of the flesh. Jerome referred to marriage prohibition for priests when he claimed in "Against Jovinianus" that Peter and the other apostles had been married before they were called, but subsequently gave up their marital relations.
Celibacy as a vocation may be independent from religious vows (as is the case with consecrated virgins, ascetics and hermits). In the Catholic, Orthodox and Oriental Orthodox traditions, bishops are required to be celibate. In the Eastern Christian traditions, priests and deacons are allowed to be married, yet have to remain celibate if they are unmarried at the time of ordination.
In the early Church higher clerics lived in marriages. Augustine of Hippo was one of the first to develop a theory that sexual feelings were sinful and negative. Augustine taught that the original sin of Adam and Eve was either an act of "foolishness" ("insipientia") followed by "pride" and "disobedience" to God, or else inspired by pride. The first couple disobeyed God, who had told them not to eat of the Tree of the knowledge of good and evil (Gen 2:17). The tree was a symbol of the order of creation. Self-centeredness made Adam and Eve eat of it, thus failing to acknowledge and respect the world as it was created by God, with its hierarchy of beings and values. They would not have fallen into pride and lack of wisdom, if Satan had not sown into their senses "the root of evil" ("radix Mali"). Their nature was wounded by concupiscence or libido, which affected human intelligence and will, as well as affections and desires, including sexual desire.
The sin of Adam is inherited by all human beings. Already in his pre-Pelagian writings, Augustine taught that Original Sin was transmitted by concupiscence, which he regarded as the passion of both soul and body, making humanity a "massa damnata" (mass of perdition, condemned crowd) and much enfeebling, though not destroying, the freedom of the will.
In the early 3rd century, the Canons of the Apostolic Constitutions decreed that only lower clerics might still marry after their ordination, but marriage of bishops, priests, and deacons were not allowed. Augustine's view of sexual feelings as sinful affected his view of women. For example, he considered a man's erection to be sinful, though involuntary, because it did not take place under his conscious control. His solution was to place controls on women to limit their ability to influence men. He equated flesh with woman and spirit with man.
He believed that the serpent approached Eve because she was less rational and lacked self-control, while Adam's choice to eat was viewed as an act of kindness so that Eve would not be left alone. Augustine believed sin entered the world because man (the spirit) did not exercise control over woman (the flesh). Augustine's views on women were not all negative, however. In his "Tractates on the Gospel of John", Augustine, commenting on the Samaritan woman from John 4:1–42, uses the woman as a figure of the church.
According to Raming, the authority of the "Decretum Gratiani", a collection of Roman Catholic canon law which prohibits women from leading, teaching, or being a witness, rests largely on the views of the early church fathers, especially St. Augustine. The laws and traditions founded upon St. Augustine's views of sexuality and women continue to exercise considerable influence over church doctrinal positions regarding the role of women in the church.
One explanation for the origin of obligatory celibacy is that it is based on the writings of Saint Paul, who wrote of the advantages celibacy allowed a man in serving the Lord. Celibacy was popularised by the early Christian theologians like Saint Augustine of Hippo and Origen. Another possible explanation for the origins of obligatory celibacy revolves around more practical reason, "the need to avoid claims on church property by priests' offspring". It remains a matter of Canon Law (and often a criterion for certain religious orders, especially Franciscans) that priests may not own land and therefore cannot pass it on to legitimate or illegitimate children. The land belongs to the Church through the local diocese as administered by the Local Ordinary (usually a bishop), who is often an "ex officio" corporation sole. Celibacy is viewed differently by the Catholic Church and the various Protestant communities. It includes clerical celibacy, celibacy of the consecrated life, voluntary lay celibacy, and celibacy outside of marriage.
The Protestant Reformation rejected celibate life and sexual continence for preachers. Protestant celibate communities have emerged, especially from Anglican and Lutheran backgrounds. A few minor Christian sects advocate celibacy as a better way of life. These groups included the Shakers, the Harmony Society and the Ephrata Cloister. Celibacy not only for religious and monastics (brothers/monks and sisters/nuns) but also for bishops is upheld by the Catholic Church traditions.
Many evangelicals prefer the term "abstinence" to "celibacy." Assuming everyone will marry, they focus their discussion on refraining from premarital sex and focusing on the joys of a future marriage. But some evangelicals, particularly older singles, desire a positive message of celibacy that moves beyond the "wait until marriage" message of abstinence campaigns. They seek a new understanding of celibacy that is focused on God rather than a future marriage or a lifelong vow to the Church.
There are also many Pentecostal churches which practice celibate ministry. For instance, The full-time ministers of the Pentecostal Mission are celibate. Most of them are single, married couples can become celibate.
During the first three or four centuries, no law was promulgated prohibiting clerical marriage. Celibacy was a matter of choice for bishops, priests, and deacons.
Statutes forbidding clergy from having wives were written beginning with the Council of Elvira (306) but these early statutes were not universal and were often defied by clerics and then retracted by hierarchy. The Synod of Gangra (345) condemned a false asceticism whereby worshipers boycotted celebrations presided over by married clergy." The Apostolic Constitutions () excommunicated a priest or bishop who left his wife 'under the pretense of piety"’ (Mansi, 1:51).
"A famous letter of Synesius of Cyrene () is evidence both for the respecting of personal decision in the matter and for contemporary appreciation of celibacy. For priests and deacons clerical marriage continued to be in vogue".
"The Second Lateran Council (1139) seems to have enacted the first written law making sacred orders a diriment impediment to marriage for the universal Church." Celibacy was first required of some clerics in 1123 at the First Lateran Council. Because clerics resisted it, the celibacy mandate was restated at the Second Lateran Council (1139) and the Council of Trent (1545–64). In places, coercion and enslavement of clerical wives and children was apparently involved in the enforcement of the law. "The earliest decree in which the children [of clerics] were declared to be slaves and never to be enfranchised [freed] seems to have been a canon of the Synod of Pavia in 1018. Similar penalties were promulgated against wives and concubines (see the Synod of Melfi, 1189 can. Xii), who by the very fact of their unlawful connexion with a subdeacon or clerk of higher rank became liable to be seized by the over-lord". Celibacy for priests continues to be a contested issue even today.
In the Roman Catholic Church, the Twelve Apostles are considered to have been the first priests and bishops of the Church. Some say the call to be eunuchs for the sake of Heaven in Matthew 19 was a call to be sexually continent and that this developed into celibacy for priests as the successors of the apostles. Others see the call to be sexually continent in Matthew 19 to be a caution for men who were too readily divorcing and remarrying.
The view of the Church is that celibacy is a reflection of life in Heaven, a source of detachment from the material world which aids in one's relationship with God. Celibacy is designed to "consecrate themselves with undivided heart to the Lord and to "the affairs of the Lord, they give themselves entirely to God and to men. It is a sign of this new life to the service of which the Church's minister is consecrated; accepted with a joyous heart celibacy radiantly proclaims the Reign of God." In contrast, Saint Peter, whom the Church considers its first Pope, was married given that he had a mother-in-law whom Christ healed (Matthew 8).
Usually, only celibate men are ordained as priests in the Latin Rite. Married clergy who have converted from other Christian denominations can be ordained Roman Catholic priests without becoming celibate. Priestly celibacy is not "doctrine" of the Church (such as the belief in the Assumption of Mary) but a matter of discipline, like the use of the vernacular (local) language in Mass or Lenten fasting and abstinence. As such, it can theoretically change at any time though it still must be obeyed by Catholics until the change were to take place. The Eastern Catholic Churches ordain both celibate and married men. However, in both the East and the West, bishops are chosen from among those who are celibate. In Ireland, several priests have fathered children, the two most prominent being Bishop Eamonn Casey and Father Michael Cleary.
The classical heritage flourished throughout the Middle Ages in both the Byzantine Greek East and the Latin West. Will Durant has made a case that certain prominent features of Plato's ideal community were discernible in the organization, dogma and effectiveness of "the" Medieval Church in Europe:
"The clergy, like Plato's guardians, were placed in authority... by their talent as shown in ecclesiastical studies and administration, by their disposition to a life of meditation and simplicity, and ... by the influence of their relatives with the powers of state and church. In the latter half of the period in which they ruled [AD 800 onwards], the clergy were as free from family cares as even Plato could desire [for such guardians]... [Clerical] Celibacy was part of the psychological structure of the power of the clergy; for on the one hand they were unimpeded by the narrowing egoism of the family, and on the other their apparent superiority to the call of the flesh added to the awe in which lay sinners held them..." "In the latter half of the period in which they ruled, the clergy were as free from family cares as even Plato could desire".
"Greater understanding of human psychology has led to questions regarding the impact of celibacy on the human development of the clergy. The realization that many non-European countries view celibacy negatively has prompted questions concerning the value of retaining celibacy as an absolute and universal requirement for ordained ministry in the Roman Catholic Church"
"The declining number of priests in active ministry, the exemption from the requirement of celibacy for married clergy who enter the Catholic Church after having been ordained in the Episcopal Church, and reported incidences of de facto nonobservance of the requirement by clergy in various parts of the world, especially in Africa and Latin America, suggests that the discussion [of celibacy] will continue."
The reintroduction of a permanent diaconate has permitted the Church to allow married men to become deacons but they may not go on to become priests.
Some homosexual Christians choose to be celibate following their denomination's teachings on homosexuality.
In 2014, the American Association of Christian Counselors amended its code of ethics to eliminate the promotion of conversion therapy for homosexuals and encouraged them to be celibate instead.
In Hinduism, celibacy is usually associated with the "sadhus" ("holy men"), ascetics who withdraw from society and renounce all worldly ties. Celibacy, termed "brahmacharya" in Vedic scripture, is the fourth of the "yamas" and the word literally translated means "dedicated to the Divinity of Life". The word is often used in yogic practice to refer to celibacy or denying pleasure, but this is only a small part of what "brahmacharya" represents. The purpose of practicing "brahmacharya" is to keep a person focused on the purpose in life, the things that instill a feeling of peace and contentment. It is also used to cultivate occult powers and many supernatural feats, called siddhi.
Islamic attitudes toward celibacy have been complex, Muhammad denounced it, however some Sufi orders embrace it. Islam does not promote celibacy; rather it condemns premarital sex and extramarital sex. In fact, according to Islam, marriage enables one to attain the highest form of righteousness within this sacred spiritual bond and is as such to be sought after and desired. It disagrees with the concept that marriage acts as a form of distraction in attaining nearness to God. The Qur'an (57:27) states, "But the Monasticism which they invented for themselves, We did not prescribe for them but only to please God therewith, but that they did not observe it with the right observance."
The following sayings about Muhammad also address celibacy:
"There have been people who have come to the prophet and explained how they love to be engaged in prayer and fasting for the sake of God. The Prophet Mohammed told them that, despite this being good, it is also a blessing to raise a family, to remain moderate and not to concentrate too much on one aspect as not only can this be unhealthy for an individual as well as upon society, it may also take one away from God."
Celibacy appears as a peculiarity among some Sufis.
Celibacy was practiced by women saints in Sufism. Celibacy was debated along with women's roles in Sufism in medieval times.
Celibacy, poverty, meditation, and mysticism within an ascetic context along with worship centered around Saint's tombs were promoted by the Qadiri Sufi order among Hui Muslims in China. In China, unlike other Muslim sects, the leaders (Shaikhs) of the Qadiriyya Sufi order are celibate. Unlike other Sufi orders in China, the leadership within the order is not a hereditary position, rather, one of the disciples of the celibate Shaikh is chosen by the Shaikh to succeed him . The 92-year-old celibate Shaikh Yang Shijun was the leader of the Qadiriya order in China as of 1998.
Celibacy is practiced by Haydariya Sufi dervishes.
The spiritual teacher Meher Baba stated that "[F]or the [spiritual] aspirant a life of strict celibacy is preferable to married life, if restraint comes to him easily without undue sense of self-repression. Such restraint is difficult for most persons and sometimes impossible, and for them married life is decidedly more helpful than a life of celibacy. For ordinary persons, married life is undoubtedly advisable unless they have a special aptitude for celibacy". Baba also asserted that "The value of celibacy lies in the habit of restraint and the sense of detachment and independence which it gives" and that "The aspirant must choose one of the two courses which are open to him. He must take to the life of celibacy or to the married life, and he must avoid at all costs a cheap compromise between the two. Promiscuity in sex gratification is bound to land the aspirant in a most pitiful and dangerous chaos of ungovernable lust."
In Sparta and many other Greek cities, failure to marry was grounds for loss of citizenship, and could be prosecuted as a crime. Both Cicero and Dionysius of Halicarnassus stated that Roman law forbade celibacy. There are no records of such a prosecution, nor is the Roman punishment for refusing to marry known.
Pythagoreanism was the system of esoteric and metaphysical beliefs held by Pythagoras and his followers. Pythagorean thinking was dominated by a profoundly mystical view of the world. The Pythagorean code further restricted his members from eating meat, fish, and beans which they practised for religious, ethical and ascetic reasons, in particular the idea of metempsychosis – the transmigration of souls into the bodies of other animals.
"Pythagoras himself established a small community that set a premium on study, vegetarianism, and sexual restraint or abstinence. Later philosophers believed that celibacy would be conducive to the detachment and equilibrium required by the philosopher's calling."
The tradition of sworn virgins developed out of the "Kanuni i Lekë Dukagjinit" (, or simply the "Kanun"). The "Kanun" is not a religious document – many groups follow it, including Roman Catholics, the Albanian Orthodox, and Muslims.
Women who become sworn virgins make a vow of celibacy, and are allowed to take on the social role of men: inheriting land, wearing male clothing, etc. | https://en.wikipedia.org/wiki?curid=6035 |
Coalition government
A coalition government is a form of government in which political parties cooperate, reducing the dominance of any one party within that "coalition". The usual reason for this arrangement is that no party on its own can achieve a majority in the election. A coalition government might also be created in a time of national difficulty or crisis (for example, during wartime or economic crisis) to give a government the high degree of perceived political legitimacy or collective identity, it can also play a role in diminishing internal political strife. In such times, parties have formed all-party coalitions (national unity governments, grand coalitions). If a coalition collapses, a confidence vote is held or a motion of no confidence is taken.
When a general election does not produce a clear majority for a single party, parties either form coalition cabinets, supported by a parliamentary majority, or minority cabinets which may consist of one or more parties. Cabinets based on a group of parties that command a majority in parliament tend to be more stable and long-lived than minority cabinets. While the former are prone to internal struggles, they have less reason to fear votes of no confidence. Majority governments based on a single party are typically even more stable, as long as their majority can be maintained.
Countries which often operate with coalition cabinets include: the Nordic countries, the Benelux countries, Australia, Austria, Cyprus, France, Germany, Greece, India, Indonesia, Ireland, Israel, Italy, Japan, Kenya, Kosovo, Lithuania, Latvia, Lebanon, Nepal, New Zealand, Pakistan, Thailand, Trinidad and Tobago, Turkey and Ukraine. Switzerland has been ruled by a coalition of the four strongest parties in parliament from 1959 to 2008, called the "Magic Formula". Between 2010 and 2015, the United Kingdom also operated a formal coalition between the Conservative and the Liberal Democrat parties, but this was unusual: the UK usually has a single-party majority government.
In the United Kingdom, coalition governments (sometimes known as "national governments") usually have only been formed at times of national crisis. The most prominent was the National Government of 1931 to 1940. There were multi-party coalitions during both world wars. Apart from this, when no party has had a majority, minority governments normally have been formed with one or more opposition parties agreeing to vote in favour of the legislation which governments need to function: for instance the Labour government of James Callaghan formed a pact with the Liberals from March 1977 until July 1978, after a series of by-election defeats had eroded Labour's majority of three seats which had been gained at the October 1974 election. However, in the run-up to the 1997 general election, Labour opposition leader Tony Blair was in talks with Liberal Democrat leader Paddy Ashdown about forming a coalition government if Labour failed to win a majority at the election; but there proved to be no need for a coalition as Labour won the election by a landslide. The 2010 general election resulted in a hung parliament (Britain's first for 36 years), and the Conservatives, led by David Cameron, which had won the largest number of seats, formed a coalition with the Liberal Democrats in order to gain a parliamentary majority, ending 13 years of Labour government. This was the first time that the Conservatives and Lib Dems had made a power-sharing deal at Westminster. It was also the first full coalition in Britain since 1945, having been formed 70 years virtually to the day after the establishment of Winston Churchill's wartime coalition,
Labour and the Liberal Democrats have entered into a coalition twice in the Scottish Parliament, as well as twice in the Welsh Assembly.
In Germany, for instance, coalition government is the norm, as it is rare for either the Christian Democratic Union of Germany together with their partners the Christian Social Union in Bavaria (CDU/CSU), or the Social Democratic Party of Germany (SPD), to win an unqualified majority in a national election. Thus, at the federal level, governments are formed with at least two parties. For example, Helmut Kohl's CDU governed for years in coalition with the Free Democratic Party (FDP); from 1998 to 2005 Gerhard Schröder's SPD was in power with the Greens; and from 2009 Angela Merkel, CDU/CSU was in power with the FDP.
"Grand coalitions" of the two large parties also occur, but these are relatively rare, as large parties usually prefer to associate with small ones. However, if none of the larger parties can receive enough votes to form their preferred coalition, a grand coalition might be their only choice for forming a government. This was the situation in Germany in 2005 when Angela Merkel became Chancellor: in early elections, the CDU/CSU did not garner enough votes to form a majority coalition with the FDP; similarly the SPD and Greens did not have enough votes to continue with their formerly ruling coalition. A grand coalition government was subsequently forged between the CDU/CSU and the SPD. Partnerships like these typically involve carefully structured cabinets. The CDU/CSU ended up holding the Chancellery while the SPD took the majority of cabinet posts. Parties frequently make statements ahead of elections which coalitions they categorically reject, similar to election promises or shadow cabinets in other countries.
In Germany, coalitions rarely consist of more than two parties (CDU and CSU, two allies which always form a single caucus, are in this regard considered a single party). However, in the 2010s coalitions on the state level increasingly included three different parties, often FDP, Greens and one of the major parties or "red red green" coalitions of SPD, Linkspartei and Greens. By 2016, the Greens have joined governments on the state level in eleven coalitions in seven various constellations.
Armenia became an independent state in 1991, following the collapse of the Soviet Union. Since then, many political parties were formed in it, who mainly work with each other to form coalition governments. Currently the country is governed by the My Step Alliance coalition after successfully gaining a majority in the National Assembly of Armenia following the 2018 Armenian parliamentary election.
In federal Australian politics, the conservative Liberal, National, Country Liberal and Liberal National parties are united in a coalition, known simply as the Coalition. The Coalition has become so stable, at least at the federal level, that in practice the lower house of Parliament has become a two-party house, with the Coalition and the Labor Party being the major parties. This coalition is also found in the states of New South Wales and Victoria. In South Australia and Western Australia the Liberal and National parties compete separately, while in the Northern Territory and Queensland the two parties have merged, forming the Country Liberal Party, in 1978, and the Liberal National Party, in 2008, respectively.
The other federal coalition has been:
In Belgium, where there are separate Dutch-speaking and French-speaking parties for each political grouping, coalition cabinets of up to six parties are common.
In Canada, the Great Coalition was formed in 1864 by the Clear Grits, Parti bleu, and Liberal-Conservative Party. During the First World War, Prime Minister Robert Borden attempted to form a coalition with the opposition Liberals to broaden support for controversial conscription legislation. The Liberal Party refused the offer but some of their members did cross the floor and join the government. Although sometimes referred to as a coalition government, according to the definition above, it was not. It was disbanded after the end of the war.
As a result of the 1919 Ontario election, the United Farmers of Ontario and the Labour Party, together with three independent MLAs, formed a coalition that governed Ontario until 1923.
In British Columbia, the governing Liberals formed a coalition with the opposition Conservatives in order to prevent the surging, left-wing Cooperative Commonwealth Federation from taking power in the 1941 British Columbia general election. Liberal premier Duff Pattullo refused to form a coalition with the third-place Conservatives, so his party removed him. The Liberal–Conservative coalition introduced a winner-take-all preferential voting system (the "Alternative Vote") in the hopes that their supporters would rank the other party as their second preference; however, this strategy did not take CCF second preferences into account. In the 1952 British Columbia general election, to the surprise of many, the right-wing populist BC Social Credit Party won a minority. They were able to win a majority in the subsequent election as Liberal and Conservative supporters shifted their anti-CCF vote to Social Credit.
Manitoba has had more formal coalition governments than any other province. Following gains by the United Farmer's/Progressive movement elsewhere in the country, the United Farmers of Manitoba unexpectedly won the 1921 election. Like their counterparts in Ontario, they had not expected to win and did not have a leader. They asked John Bracken, a professor in animal husbandry, to become leader and premier. Bracken changed the party's name to the Progressive Party of Manitoba. During the Great Depression, Bracken survived at a time when other premiers were being defeated by forming a coalition government with the Manitoba Liberals (eventually, the two parties would merge into the , and decades later, the party would change its name to the Manitoba Liberal Party). In 1940, Bracken formed a wartime coalition government with almost every party in the Manitoba Legislature (the Conservatives, CCF, and Social Credit; however, the CCF broke with the coalition after a few years over policy differences). The only party not included was the small, communist Labor-Progressive Party, which had a handful of seats.
In Saskatchewan, NDP premier Roy Romanow formed a formal coalition with the Saskatchewan Liberals in 1999 after being reduced to a minority. After two years, the newly elected Liberal leader David Karwacki ordered the coalition be disbanded, the Liberal caucus disagreed with him and left the Liberals to run as New Democrats in the upcoming election. The Saskatchewan NDP was re-elected with a majority under its new leader Lorne Calvert, while the Saskatchewan Liberals lost their remaining seats and have not been competitive in the province since.
According to historian Christopher Moore, coalition governments in Canada became much less possible in 1919, when the leaders of parties were no longer chosen by elected MPs but instead began to be chosen by party members. Such a manner of leadership election had never been tried in any parliamentary system before. According to Moore, as long as that kind of leadership selection process remains in place and concentrates power in the hands of the leader, as opposed to backbenchers, then coalition governments will be very difficult to form. Moore shows that the diffusion of power within a party tends to also lead to a diffusion of power in the parliament in which that party operates, thereby making coalitions more likely.
During the 2008–09 Canadian parliamentary dispute, two of Canada's opposition parties signed an agreement to form what would become the country's second coalition government since Confederation if the minority Conservative government was defeated on a vote of non-confidence, unseating Stephen Harper as Prime Minister. The agreement outlined a formal coalition consisting of two opposition parties, the Liberal Party and the New Democratic Party. The Bloc Québécois agreed to support the proposed coalition on confidence matters for 18 months. In the end, parliament was prorogued by the Governor General, and the coalition dispersed before parliament was reconvened.
From the creation of the Folketing in 1849 through the introduction of proportional representation in 1918, there were only single-party governments in Denmark. Thorvald Stauning formed his second government and Denmark's first coalition government in 1929. With the exception of a string of one-party governments during the 1970s, the norm since 1929 has been coalition governments. Every government from 1982 until the 2015 elections were coalitions. The most recent coalition was Løkke's third government, which was replaced by the one-party Frederiksen government in 2019.
When the Social Democrats under Stauning won 46% of the votes in the 1935 election, this was the closest any party has gotten to winning an outright majority in parliament. One party has thus never held a majority alone, and even one-party governments since 1918 have needed the support of at least one other party to govern. For example, the current government consists only of the Social Democrats, but also relies on the support of the Social Liberal Party, the Socialist People's Party, and the Red–Green Alliance.
In Finland, no party has had an absolute majority in the parliament since independence, and multi-party coalitions have been the norm. Finland experienced its most stable government (Lipponen I and II) since independence with a five-party governing coalition, a so-called "rainbow government". The Lipponen cabinets set the stability record and were unusual in the respect that both the centre-left (SDP) and radical left-wing (Left Alliance) parties sat in the government with the major centre-right party (National Coalition). The Katainen cabinet was also a rainbow coalition of a total of five parties.
Since India's Independence on 15 August 1947, Indian National Congress, the major political party instrumental in Indian independence movement, ruled the nation. The first Prime Minister Jawaharlal Nehru, second PM Lal Bahadur Shastri and the third PM Indira Gandhi, all were from the Congress party. However, Raj Narain, who had unsuccessfully contested election against Indira from the constituency of Rae Bareilly in 1971, lodged a case, alleging electoral malpractices. In June 1975, Indira was found guilty and barred by High Court from holding public office for six years. In response, an ungracious Emergency was declared under the pretext of national security. The next election's result was that India's first-ever coalition government was formed at the national level under the Prime Ministership of Morarji Desai, which was also the first non-Congress national government, which existed from 24 March 1977 to 15 July 1979, headed by the Janata Party, an amalgam of political parties opposed to Emergency imposed between 1975 and 1977. As the popularity of Janata Party dwindled, Morarji Desai had to resign and Charan Singh, a rival of Desai became the fifth PM. However, due to lack of support, this coalition government did not complete its five-year term.
Congress returned to the power in 1980 under Indira Gandhi, and later under Rajiv Gandhi as the 6th PM. However, the next general election of 1989 once again brought a coalition government under National Front, which lasted until 1991, with two Prime Ministers, the second one being supported by Congress. The 1991 election resulted in a Congress led stable minority government for five years. The next 11th parliament produced three Prime Ministers in two years and forced the country back to the polls in 1998. The first successful coalition government in India which completed the whole 5-year term was the Bharatiya Janata Party (BJP) led National Democratic Alliance with Atal Bihari Vajpayee as PM from 1999 to 2004. Then another coalition, Congress led United Progressive Alliance, consisting of 13 separate parties ruled India for two terms from 2004 to 2014 with Manmohan Singh as PM. However, in the 16th general election in May 2014, BJP secured majority on its own (first party to do so since 1984 election) and National Democratic Alliance came into power, with Narendra Modi as Prime Minister. In 2019, Narendra Modi got re-elected as Prime Minister for the second time as National Democratic Alliance again secured majority in the 17th general election.
As a result of the toppling of Suharto, political freedom is significantly increased. Compared to only three parties allowed to exist in the New Order era, a total of 48 political parties participated in the 1999 election, a total of 24 parties in the 2004 election, 38 parties in the 2009 election, and 15 parties in the 2014 election. There are no majority winner of those elections and coalition governments are inevitable. The current government is a coalition of seven parties led by the PDIP and Golkar.
In Ireland, coalition governments are common; not since 1977 has a single party formed a majority government. Coalition governments to date have been led by either Fianna Fáil or Fine Gael. They have been joined in government by one or more smaller parties or independent members of parliament (TDs).
Ireland's first coalition government was formed after the 1948 general election, with five parties and independents represented at cabinet. Before 1989, Fianna Fáil had opposed participation in coalition governments, preferring single-party minority government instead. It formed a coalition government with the Progressive Democrats in that year.
The Labour Party has been in government on eight occasions. On all but one of those occasions, it was as a junior coalition party to Fine Gael. The exception was a government with Fianna Fáil from 1993 to 1994. The Government of the 31st Dáil (2011–16), though a traditional Fine Gael–Labour coalition, was a grand coalition of the two largest parties, as Fianna Fáil had fallen to third place in the Dáil.
The current government is minority Fine Gael government with Independents at cabinet, supported by a confidence and supply arrangement with Fianna Fáil.
A similar situation exists in Israel, which typically has at least 10 parties holding representation in the Knesset. The only faction to ever gain the majority of Knesset seats was Alignment, an alliance of the Labor Party and Mapam that held an absolute majority for a brief period from 1968 to 1969. Historically, control of the Israeli government has alternated between periods of rule by the right-wing Likud in coalition with several right-wing and religious parties and periods of rule by the center-left Labor in coalition with several left-wing parties. Ariel Sharon's formation of the centrist Kadima party in 2006 drew support from former Labor and Likud members, and Kadima ruled in coalition with several other parties.
Israel also formed a national unity government from 1984–1988. The premiership and foreign ministry portfolio were held by the head of each party for two years, and they switched roles in 1986.
Post-World War II Japan has historically been dominated by the Liberal Democratic Party, but there was a brief coalition government formed after the 1993 election following LDP's first loss of its overall House of Representatives majority since 1955, winning only 223 out of 511 seats. The LDP government was replaced by an eight-party coalition government, which consisted of all of the previous opposition parties excluding the Japanese Communist Party, who together controlled 243 seats. Every Japanese government since then has been a coalition government in one way or another.
MMP was introduced in New Zealand in the 1996 election.
In order to get into power, parties need to get a total of 50% of the 121 seats in parliament – 61. Since no parties have ever gotten a full majority, they must form coalitions with other parties. For example, during the 2017 general election, Labour got 46 seats and New Zealand First got nine. The two formed a Coalition Government with confidence and supply from the Green Party which got eight seats.
Since 2015, there are many more coalition governments than previously in municipalities, autonomous regions and, since 2020 (coming from the November 2019 Spanish general election), in the Spanish Government. There are two ways of conforming them: all of them based on a program and its institutional architecture, one consists on distributing the different areas of government between the parties conforming the coalition and the other one is, like in the Valencian Community, where the ministries are structured with members of all the political parties being represented, so that conflicts that may occur are regarding competences and not fights between parties.
Coalition governments in Spain had already existed during the 2nd Republic, and have been common in some specific Autonomous Communities since the 80's. Nonetheless, the prevalence of two big parties overall has been eroded and the need for coalitions appears to be the new normal since around 2015.
Since the 1989 election, there have been 4 coalition governments, all including at least both the conservative National Party and the liberal Colorado Party. The first one was after the election of the blanco Luis Alberto Lacalle and lasted until 1992 due to policy disagreements, the longest lasting coalition was the Colorado-led coalition under the second government of Julio María Sanguinetti, in which the national leader Alberto Volonté was frequently described as a "Prime Minister", the next coalition (under president Jorge Batlle) was also Colorado-led, but it lasted only until after the 2002 Uruguay banking crisis, when the blancos abandoned the government. After the 2019 Uruguayan general election, the blanco Luis Lacalle Pou formed the coalición multicolor, composed of his own National Party, the liberal Colorado Party, the right wing populist Open Cabildo and the center left Independent Party.
Advocates of proportional representation suggest that a coalition government leads to more consensus-based politics, as a government comprising differing parties (often based on different ideologies) need to compromise about governmental policy. Another stated advantage is that a coalition government better reflects the popular opinion of the electorate within a country.
Those who disapprove of coalition governments believe that such governments have a tendency to be fractious and prone to disharmony, as their component parties hold differing beliefs and thus may not always agree on policy. Sometimes the results of an election mean that the coalitions which are mathematically most probable are ideologically infeasible, for example in Flanders or Northern Ireland. A second difficulty might be the ability of minor parties to play "kingmaker" and, particularly in close elections, gain far more power in exchange for their support than the size of their vote would otherwise justify.
Coalition governments have also been criticized for sustaining a consensus on issues when disagreement and the consequent discussion would be more fruitful. To forge a consensus, the leaders of ruling coalition parties can agree to silence their disagreements on an issue to unify the coalition against the opposition. The coalition partners, if they control the parliamentary majority, can collude to make the parliamentary discussion on the issue irrelevant by consistently disregarding the arguments of the opposition and voting against the opposition's proposals — even if there is disagreement within the ruling parties about the issue.
Powerful parties can also act in an oligocratic way to form an alliance to stifle the growth of emerging parties. Of course, such an event is rare in coalition governments when compared to two-party systems, which typically exist because of stifling of the growth of emerging parties, often through discriminatory nomination rules regulations and plurality voting systems, and so on.
A single, more powerful party can shape the policies of the coalition disproportionately. Smaller or less powerful parties can be intimidated to not openly disagree. In order to maintain the coalition, they would have to vote against their own party's platform in the parliament. If they do not, the party has to leave the government and loses executive power. However, this is contradicted by the "kingmaker" factor mentioned above. | https://en.wikipedia.org/wiki?curid=6036 |
Chemical engineering
Chemical engineering is a branch of engineering that uses principles of chemistry, physics, mathematics, biology, and economics to efficiently use, produce, design, transport and transform energy and materials. The work of chemical engineers can range from the utilisation of nano-technology and nano-materials in the laboratory to large-scale industrial processes that convert chemicals, raw materials, living cells, microorganisms, and energy into useful forms and products.
Chemical engineers are involved in many aspects of plant design and operation, including safety and hazard assessments, process design and analysis, modeling, control engineering, chemical reaction engineering, nuclear engineering, biological engineering, construction specification, and operating instructions.
Chemical engineers typically hold a degree in Chemical Engineering or Process Engineering. Practising engineers may have professional certification and be accredited members of a professional body. Such bodies include the Institution of Chemical Engineers (IChemE) or the American Institute of Chemical Engineers (AIChE).
A degree in chemical engineering is directly linked with all of the other engineering disciplines, to various extents.
A 1996 "British Journal for the History of Science" article cites James F. Donnelly for mentioning an 1839 reference to chemical engineering in relation to the production of sulfuric acid. In the same paper however, George E. Davis, an English consultant, was credited for having coined the term. Davis also tried to found a "Society of Chemical Engineering", but instead it was named the Society of Chemical Industry (1881), with Davis as its first Secretary. The "History of Science in United States: An Encyclopedia" puts the use of the term around 1890. "Chemical engineering", describing the use of mechanical equipment in the chemical industry, became common vocabulary in England after 1850. By 1910, the profession, "chemical engineer," was already in common use in Britain and the United States.
Chemical engineering emerged upon the development of unit operations, a fundamental concept of the discipline of chemical engineering. Most authors agree that Davis invented the concept of unit operations if not substantially developed it. He gave a series of lectures on unit operations at the Manchester Technical School (later part of University of Manchester Institute of Science and Technology and now the University of Manchester) in 1887, considered to be one of the earliest such about chemical engineering. Three years before Davis' lectures, Henry Edward Armstrong taught a degree course in chemical engineering at the City and Guilds of London Institute. Armstrong's course failed simply because its graduates were not especially attractive to employers. Employers of the time would have rather hired chemists and mechanical engineers. Courses in chemical engineering offered by Massachusetts Institute of Technology (MIT) in the United States, Owens College in Manchester, England, and University College London suffered under similar circumstances.
Starting from 1888, Lewis M. Norton taught at MIT the first chemical engineering course in the United States. Norton's course was contemporaneous and essentially similar to Armstrong's course. Both courses, however, simply merged chemistry and engineering subjects along with product design. "Its practitioners had difficulty convincing engineers that they were engineers and chemists that they were not simply chemists." Unit operations was introduced into the course by William Hultz Walker in 1905. By the early 1920s, unit operations became an important aspect of chemical engineering at MIT and other US universities, as well as at Imperial College London. The American Institute of Chemical Engineers (AIChE), established in 1908, played a key role in making chemical engineering considered an independent science, and unit operations central to chemical engineering. For instance, it defined chemical engineering to be a "science of itself, the basis of which is ... unit operations" in a 1922 report; and with which principle, it had published a list of academic institutions which offered "satisfactory" chemical engineering courses. Meanwhile, promoting chemical engineering as a distinct science in Britain led to the establishment of the Institution of Chemical Engineers (IChemE) in 1922. IChemE likewise helped make unit operations considered essential to the discipline.
In 1940s, it became clear that unit operations alone were insufficient in developing chemical reactors. While the predominance of unit operations in chemical engineering courses in Britain and the United States continued until the 1960s, transport phenomena started to experience greater focus. Along with other novel concepts, such as process systems engineering (PSE), a "second paradigm" was defined. Transport phenomena gave an analytical approach to chemical engineering while PSE focused on its synthetic elements, such as control system and process design. Developments in chemical engineering before and after World War II were mainly incited by the petrochemical industry, however, advances in other fields were made as well. Advancements in biochemical engineering in the 1940s, for example, found application in the pharmaceutical industry, and allowed for the mass production of various antibiotics, including penicillin and streptomycin. Meanwhile, progress in polymer science in the 1950s paved way for the "age of plastics".
Concerns regarding the safety and environmental impact of large-scale chemical manufacturing facilities were also raised during this period. "Silent Spring", published in 1962, alerted its readers to the harmful effects of DDT, a potent insecticide. The 1974 Flixborough disaster in the United Kingdom resulted in 28 deaths, as well as damage to a chemical plant and three nearby villages. The 1984 Bhopal disaster in India resulted in almost 4,000 deaths. These incidents, along with other incidents, affected the reputation of the trade as industrial safety and environmental protection were given more focus. In response, the IChemE required safety to be part of every degree course that it accredited after 1982. By the 1970s, legislation and monitoring agencies were instituted in various countries, such as France, Germany, and the United States.
Advancements in computer science found applications designing and managing plants, simplifying calculations and drawings that previously had to be done manually. The completion of the Human Genome Project is also seen as a major development, not only advancing chemical engineering but genetic engineering and genomics as well. Chemical engineering principles were used to produce DNA sequences in large quantities.
Chemical engineering involves the application of several principles. Key concepts are presented below.
Chemical engineering design concerns the creation of plans, specification, and economic analyses for pilot plants, new plants or plant modifications. Design engineers often work in a consulting role, designing plants to meet clients' needs. Design is limited by a number of factors, including funding, government regulations and safety standards. These constraints dictate a plant's choice of process, materials and equipment.
Plant construction is coordinated by project engineers and project managers depending on the size of the investment. A chemical engineer may do the job of project engineer full-time or part of the time, which requires additional training and job skills or act as a consultant to the project group. In the USA the education of chemical engineering graduates from the Baccalaureate programs accredited by ABET do not usually stress project engineering education, which can be obtained by specialized training, as electives, or from graduate programs. Project engineering jobs are some of the largest employers for chemical engineers.
A unit operation is a physical step in an individual chemical engineering process. Unit operations (such as crystallization, filtration, drying and evaporation) are used to prepare reactants, purifying and separating its products, recycling unspent reactants, and controlling energy transfer in reactors. On the other hand, a unit process is the chemical equivalent of a unit operation. Along with unit operations, unit processes constitute a process operation. Unit processes (such as nitration and oxidation) involve the conversion of material by biochemical, thermochemical and other means. Chemical engineers responsible for these are called process engineers.
Process design requires the definition of equipment types and sizes as well as how they are connected together and the materials of construction. Details are often printed on a Process Flow Diagram which is used to control the capacity and reliability of a new or modified chemical factory.
Education for chemical engineers in the first college degree 3 or 4 years of study stresses the principles and practices of process design. The same skills are used in existing chemical plants to evaluate the efficiency and make recommendations for improvements.
Modeling and analysis of transport phenomena is essential for many industrial applications. Transport phenomena involve fluid dynamics, heat transfer and mass transfer, which are governed mainly by momentum transfer, energy transfer and transport of chemical species, respectively. Models often involve separate considerations for macroscopic, microscopic and molecular level phenomena. Modeling of transport phenomena, therefore, requires an understanding of applied mathematics.
Chemical engineers "develop economic ways of using materials and energy". Chemical engineers use chemistry and engineering to turn raw materials into usable products, such as medicine, petrochemicals and plastics on a large-scale, industrial setting. They are also involved in waste management and research. Both applied and research facets could make extensive use of computers.
Chemical engineers may be involved in industry or university research where they are tasked with designing and performing experiments to create better and safer methods for production, pollution control, and resource conservation. They may be involved in designing and constructing plants as a project engineer. Chemical engineers serving as project engineers use their knowledge in selecting optimal production methods and plant equipment to minimize costs and maximize safety and profitability. After plant construction, chemical engineering project managers may be involved in equipment upgrades, process changes, troubleshooting, and daily operations in either full-time or consulting roles. | https://en.wikipedia.org/wiki?curid=6038 |
Compact space
In mathematics, more specifically in general topology, compactness is a property that generalizes the notion of a subset of Euclidean space being closed (i.e., containing all its limit points) and bounded (i.e., having all its points lie within some fixed distance of each other). Examples include a closed interval, a rectangle, or a finite set of points. This notion is defined for more general topological spaces than Euclidean space in various ways.
One such generalization is that a topological space is "sequentially" compact if every infinite sequence of points sampled from the space has an infinite subsequence that converges to some point of the space. The Bolzano–Weierstrass theorem states that a subset of Euclidean space is compact in this sequential sense if and only if it is closed and bounded. Thus, if one chooses an infinite number of points in the "closed" unit interval , some of those points will get arbitrarily close to some real number in that space. For instance, some of the numbers in the sequence accumulate to 0 (while others accumulate to 1). The same set of points would not accumulate to any point of the "open" unit interval ; so the open unit interval is not compact. Euclidean space itself is not compact since it is not bounded. In particular, the sequence of points , which is not bounded, has no subsequence that converges to any real number.
Apart from closed and bounded subsets of Euclidean space, typical examples of compact spaces are encountered in mathematical analysis, where the property of compactness of some topological spaces arises in the hypotheses or in the conclusions of many fundamental theorems, such as the Bolzano–Weierstrass theorem, the extreme value theorem, the Arzelà–Ascoli theorem, and the Peano existence theorem. Another example is the definition of distributions, which uses the space of smooth functions that are zero outside of some (unspecified) compact space.
Various equivalent notions of compactness, including sequential compactness and limit point compactness, can be developed in general metric spaces. In general topological spaces, however, different notions of compactness are not necessarily equivalent. The most useful notion, which is the standard definition of the unqualified term "compactness", is phrased in terms of the existence of finite families of open sets that "cover" the space in the sense that each point of the space lies in some set contained in the family. This more subtle notion, introduced by Pavel Alexandrov and Pavel Urysohn in 1929, exhibits compact spaces as generalizations of finite sets. In spaces that are compact in this sense, it is often possible to patch together information that holds locally—that is, in a neighborhood of each point—into corresponding statements that hold throughout the space, and many theorems are of this character.
The term compact set is sometimes used as a synonym for compact space, but often refers to a compact subspace of a topological space as well.
In the 19th century, several disparate mathematical properties were understood that would later be seen as consequences of compactness. On the one hand, Bernard Bolzano (1817) had been aware that any bounded sequence of points (in the line or plane, for instance) has a subsequence that must eventually get arbitrarily close to some other point, called a limit point. Bolzano's proof relied on the method of bisection: the sequence was placed into an interval that was then divided into two equal parts, and a part containing infinitely many terms of the sequence was selected. The process could then be repeated by dividing the resulting smaller interval into smaller and smaller parts—until it closes down on the desired limit point. The full significance of Bolzano's theorem, and its method of proof, would not emerge until almost 50 years later when it was rediscovered by Karl Weierstrass.
In the 1880s, it became clear that results similar to the Bolzano–Weierstrass theorem could be formulated for spaces of functions rather than just numbers or geometrical points. The idea of regarding functions as themselves points of a generalized space dates back to the investigations of Giulio Ascoli and Cesare Arzelà. The culmination of their investigations, the Arzelà–Ascoli theorem, was a generalization of the Bolzano–Weierstrass theorem to families of continuous functions, the precise conclusion of which was that it was possible to extract a uniformly convergent sequence of functions from a suitable family of functions. The uniform limit of this sequence then played precisely the same role as Bolzano's "limit point". Towards the beginning of the twentieth century, results similar to that of Arzelà and Ascoli began to accumulate in the area of integral equations, as investigated by David Hilbert and Erhard Schmidt. For a certain class of Green's functions coming from solutions of integral equations, Schmidt had shown that a property analogous to the Arzelà–Ascoli theorem held in the sense of mean convergence—or convergence in what would later be dubbed a Hilbert space. This ultimately led to the notion of a compact operator as an offshoot of the general notion of a compact space. It was Maurice Fréchet who, in 1906, had distilled the essence of the Bolzano–Weierstrass property and coined the term "compactness" to refer to this general phenomenon (he used the term already in his 1904 paper which led to the famous 1906 thesis).
However, a different notion of compactness altogether had also slowly emerged at the end of the 19th century from the study of the continuum, which was seen as fundamental for the rigorous formulation of analysis. In 1870, Eduard Heine showed that a continuous function defined on a closed and bounded interval was in fact uniformly continuous. In the course of the proof, he made use of a lemma that from any countable cover of the interval by smaller open intervals, it was possible to select a finite number of these that also covered it. The significance of this lemma was recognized by Émile Borel (1895), and it was generalized to arbitrary collections of intervals by Pierre Cousin (1895) and Henri Lebesgue (1904). The Heine–Borel theorem, as the result is now known, is another special property possessed by closed and bounded sets of real numbers.
This property was significant because it allowed for the passage from local information about a set (such as the continuity of a function) to global information about the set (such as the uniform continuity of a function). This sentiment was expressed by , who also exploited it in the development of the integral now bearing his name. Ultimately, the Russian school of point-set topology, under the direction of Pavel Alexandrov and Pavel Urysohn, formulated Heine–Borel compactness in a way that could be applied to the modern notion of a topological space. showed that the earlier version of compactness due to Fréchet, now called (relative) sequential compactness, under appropriate conditions followed from the version of compactness that was formulated in terms of the existence of finite subcovers. It was this notion of compactness that became the dominant one, because it was not only a stronger property, but it could be formulated in a more general setting with a minimum of additional technical machinery, as it relied only on the structure of the open sets in a space.
Any finite space is trivially compact. A non-trivial example of a compact space is the (closed) unit interval of real numbers. If one chooses an infinite number of distinct points in the unit interval, then there must be some accumulation point in that interval. For instance, the odd-numbered terms of the sequence get arbitrarily close to 0, while the even-numbered ones get arbitrarily close to 1. The given example sequence shows the importance of including the boundary points of the interval, since the limit points must be in the space itself — an open (or half-open) interval of the real numbers is not compact. It is also crucial that the interval be bounded, since in the interval , one could choose the sequence of points , of which no sub-sequence ultimately gets arbitrarily close to any given real number.
In two dimensions, closed disks are compact since for any infinite number of points sampled from a disk, some subset of those points must get arbitrarily close either to a point within the disc, or to a point on the boundary. However, an open disk is not compact, because a sequence of points can tend to the boundary—without getting arbitrarily close to any point in the interior. Likewise, spheres are compact, but a sphere missing a point is not since a sequence of points can still tend to the missing point, thereby not getting arbitrarily close to any point "within" the space. Lines and planes are not compact, since one can take a set of equally-spaced points in any given direction without approaching any point.
Various definitions of compactness may apply, depending on the level of generality. A subset of Euclidean space in particular is called compact if it is closed and bounded. This implies, by the Bolzano–Weierstrass theorem, that any infinite sequence from the set has a subsequence that converges to a point in the set. Various equivalent notions of compactness, such as sequential compactness and limit point compactness, can be developed in general metric spaces.
In contrast, the different notions of compactness are not equivalent in general topological spaces, and the most useful notion of compactness—originally called "bicompactness"—is defined using covers consisting of open sets (see "Open cover definition" below). That this form of compactness holds for closed and bounded subsets of Euclidean space is known as the Heine–Borel theorem. Compactness, when defined in this manner, often allows one to take information that is known locally—in a neighbourhood of each point of the space—and to extend it to information that holds globally throughout the space. An example of this phenomenon is Dirichlet's theorem, to which it was originally applied by Heine, that a continuous function on a compact interval is uniformly continuous; here, continuity is a local property of the function, and uniform continuity the corresponding global property.
Formally, a topological space is called "compact" if each of its open covers has a finite subcover. That is, is compact if for every collection of open subsets of such that
there is a finite subset of such that
Some branches of mathematics such as algebraic geometry, typically influenced by the French school of Bourbaki, use the term "quasi-compact" for the general notion, and reserve the term "compact" for topological spaces that are both Hausdorff and "quasi-compact". A compact set is sometimes referred to as a "compactum", plural "compacta".
A subset of a topological space is said to be compact if it is compact as a subspace (in the subspace topology). That is, is compact if for every arbitrary collection of open subsets of such that
there is a finite subset of such that
Compactness is a "topological" property. That is, if formula_5, with subset equipped with the subspace topology, then is compact in if and only if is compact in .
The following are equivalent:
For any subset "A" of Euclidean space R"n", "A" is compact if and only if it is closed and bounded; this is the Heine–Borel theorem.
As a Euclidean space is a metric space, the conditions in the next subsection also apply to all of its subsets. Of all of the equivalent conditions, it is in practice easiest to verify that a subset is closed and bounded, for example, for a closed interval or closed "n"-ball.
For any metric space ("X", "d"), the following are equivalent:
A compact metric space ("X", "d") also satisfies the following properties:
Let "X" be a topological space and C("X") the ring of real continuous functions on "X". For each "p"∈"X", the evaluation map formula_6
given by ev"p"("f")="f"("p") is a ring homomorphism. The kernel of ev"p" is a maximal ideal, since the residue field is the field of real numbers, by the first isomorphism theorem. A topological space "X" is pseudocompact if and only if every maximal ideal in C("X") has residue field the real numbers. For completely regular spaces, this is equivalent to every maximal ideal being the kernel of an evaluation homomorphism. There are pseudocompact spaces that are not compact, though.
In general, for non-pseudocompact spaces there are always maximal ideals "m" in C("X") such that the residue field C("X")/"m" is a (non-Archimedean) hyperreal field. The framework of non-standard analysis allows for the following alternative characterization of compactness: a topological space "X" is compact if and only if every point "x" of the natural extension "*X" is infinitely close to a point "x"0 of "X" (more precisely, "x" is contained in the monad of "x"0).
A space "X" is compact if its hyperreal extension "*X" (constructed, for example, by the ultrapower construction) has the property that every point of "*X" is infinitely close to some point of "X"⊂"*X". For example, an open real interval is not compact because its hyperreal extension *(0,1) contains infinitesimals, which are infinitely close to 0, which is not a point of "X".
A continuous image of a compact space is compact.
This implies the extreme value theorem: a continuous real-valued function on a nonempty compact space is bounded above and attains its supremum. (Slightly more generally, this is true for an upper semicontinuous function.) As a sort of converse to the above statements, the pre-image of a compact space under a proper map is compact.
A closed subset of a compact space is compact, and a finite union of compact sets is compact.
The product of any collection of compact spaces is compact. (This is Tychonoff's theorem, which is equivalent to the axiom of choice.)
Every topological space "X" is an open dense subspace of a compact space having at most one point more than "X", by the Alexandroff one-point compactification. By the same construction, every locally compact Hausdorff space "X" is an open dense subspace of a compact Hausdorff space having at most one point more than "X".
A nonempty compact subset of the real numbers has a greatest element and a least element.
Let "X" be a simply ordered set endowed with the order topology. Then "X" is compact if and only if "X" is a complete lattice (i.e. all subsets have suprema and infima). | https://en.wikipedia.org/wiki?curid=6042 |
Cicero
Marcus Tullius Cicero ( ; ; 3 January 106 BC – 7 December 43 BC), most commonly known as simply Cicero, was a Roman statesman, lawyer and Academic Skeptic philosopher who played an important role in the politics of the late Republic and vainly tried to uphold republican principles during the crises that led to the establishment of the Roman Empire. His extensive writings include treatises on rhetoric, philosophy and politics, and he is considered one of Rome's greatest orators and prose stylists. He came from a wealthy municipal family of the Roman equestrian order, and served as consul in the year 63 BC.
His influence on the Latin language was immense: it has been said that subsequent prose was either a reaction against or a return to his style, not only in Latin but in European languages up to the 19th century. Cicero introduced into Latin the arguments of the chief schools of Hellenistic philosophy and created a Latin philosophical vocabulary with neologisms such as "evidentia", "humanitas", "qualitas", "quantitas", and "essentia", distinguishing himself as a translator and philosopher.
Though he was an accomplished orator and successful lawyer, Cicero believed his political career was his most important achievement. It was during his consulship that the second Catilinarian conspiracy attempted to overthrow the government through an attack on the city by outside forces, and Cicero suppressed the revolt by summarily and controversially executing five conspirators. During the chaotic latter half of the 1st century BC marked by civil wars and the dictatorship of Gaius Julius Caesar, Cicero championed a return to the traditional republican government. Following Julius Caesar's death, Cicero became an enemy of Mark Antony in the ensuing power struggle, attacking him in a series of speeches. He was proscribed as an enemy of the state by the Second Triumvirate and consequently executed by soldiers operating on their behalf in 43 BC after having been intercepted during an attempted flight from the Italian peninsula. His severed hands and head were then, as a final revenge of Mark Antony, displayed on the Rostra.
Petrarch's rediscovery of Cicero's letters is often credited for initiating the 14th-century Renaissance in public affairs, humanism, and classical Roman culture. According to Polish historian Tadeusz Zieliński, "the Renaissance was above all things a revival of Cicero, and only after him and through him of the rest of Classical antiquity." The peak of Cicero's authority and prestige came during the 18th-century Enlightenment, and his impact on leading Enlightenment thinkers and political theorists such as John Locke, David Hume, Montesquieu and Edmund Burke was substantial.
His works rank among the most influential in European culture, and today still constitute one of the most important bodies of primary material for the writing and revision of Roman history, especially the last days of the Roman Republic.
Marcus Tullius Cicero () was born on 3 January 106 BC in Arpinum, a hill town southeast of Rome. He belonged to the "tribus" Cornelia. His father was a well-to-do member of the equestrian order and possessed good connections in Rome. However, being a semi-invalid, he could not enter public life and studied extensively to compensate. Although little is known about Cicero's mother, Helvia, it was common for the wives of important Roman citizens to be responsible for the management of the household. Cicero's brother Quintus wrote in a letter that she was a thrifty housewife.
Cicero's cognomen, or personal surname, comes from the Latin for chickpea, "cicer". Plutarch explains that the name was originally given to one of Cicero's ancestors who had a cleft in the tip of his nose resembling a chickpea. However, it is more likely that Cicero's ancestors prospered through the cultivation and sale of chickpeas. Romans often chose down-to-earth personal surnames. The famous family names of Fabius, Lentulus, and Piso come from the Latin names of beans, lentils, and peas, respectively. Plutarch writes that Cicero was urged to change this deprecatory name when he entered politics, but refused, saying that he would make "Cicero" more glorious than "Scaurus" ("Swollen-ankled") and "Catulus" ("Puppy").
During this period in Roman history, "cultured" meant being able to speak both Latin and Greek. Cicero was therefore educated in the teachings of the ancient Greek philosophers, poets and historians; as he obtained much of his understanding of the theory and practice of rhetoric from the Greek poet Archias and from the Greek rhetorician Apollonius. Cicero used his knowledge of Greek to translate many of the theoretical concepts of Greek philosophy into Latin, thus translating Greek philosophical works for a larger audience. It was precisely his broad education that tied him to the traditional Roman elite.
Cicero's interest in philosophy figured heavily in his later career and led to him providing a comprehensive account of Greek philosophy for a Roman audience, including creating a philosophical vocabulary in Latin. In 87 BC, Philo of Larissa, the head of the Academy that was founded by Plato in Athens about 300 years earlier, arrived in Rome. Cicero, "inspired by an extraordinary zeal for philosophy", sat enthusiastically at his feet and absorbed Plato's philosophy.
Cicero said of Plato's Dialogues, that if Zeus were to speak, he would use their language.
According to Plutarch, Cicero was an extremely talented student, whose learning attracted attention from all over Rome, affording him the opportunity to study Roman law under Quintus Mucius Scaevola. Cicero's fellow students were Gaius Marius Minor, Servius Sulpicius Rufus (who became a famous lawyer, one of the few whom Cicero considered superior to himself in legal matters), and Titus Pomponius. The latter two became Cicero's friends for life, and Pomponius (who later received the nickname "Atticus", and whose sister married Cicero's brother) would become, in Cicero's own words, "as a second brother", with both maintaining a lifelong correspondence.
In 79 BC, Cicero left for Greece, Asia Minor and Rhodes. This was perhaps to avoid the potential wrath of Sulla, as Plutarch claims, though Cicero himself says it was to hone his skills and improve his physical fitness. In Athens he studied philosophy with Antiochus of Ascalon, the 'Old Academic' and initiator of Middle Platonism. In Asia Minor, he met the leading orators of the region and continued to study with them. Cicero then journeyed to Rhodes to meet his former teacher, Apollonius Molon, who had previously taught him in Rome. Molon helped Cicero hone the excesses in his style, as well as train his body and lungs for the demands of public speaking. Charting a middle path between the competing Attic and Asiatic styles, Cicero would ultimately become considered second only to Demosthenes among history's orators.
Cicero married Terentia probably at the age of 27, in 79 BC. According to the upper class mores of the day it was a marriage of convenience, but lasted harmoniously for nearly 30 years. Terentia's family was wealthy, probably the plebeian noble house of Terenti Varrones, thus meeting the needs of Cicero's political ambitions in both economic and social terms. She had a half-sister named Fabia, who as a child had become a Vestal Virgin, a great honour. Terentia was a strong willed woman and (citing Plutarch) "she took more interest in her husband's political career than she allowed him to take in household affairs."
In the 50s BC, Cicero's letters to Terentia became shorter and colder. He complained to his friends that Terentia had betrayed him but did not specify in which sense. Perhaps the marriage simply could not outlast the strain of the political upheaval in Rome, Cicero's involvement in it, and various other disputes between the two. The divorce appears to have taken place in 51 BC or shortly before. In 46 or 45 BC, Cicero married a young girl, Publilia, who had been his ward. It is thought that Cicero needed her money, particularly after having to repay the dowry of Terentia, who came from a wealthy family. This marriage did not last long.
Although his marriage to Terentia was one of convenience, it is commonly known that Cicero held great love for his daughter Tullia. When she suddenly became ill in February 45 BC and died after having seemingly recovered from giving birth to a son in January, Cicero was stunned. "I have lost the one thing that bound me to life" he wrote to Atticus. Atticus told him to come for a visit during the first weeks of his bereavement, so that he could comfort him when his pain was at its greatest. In Atticus's large library, Cicero read everything that the Greek philosophers had written about overcoming grief, "but my sorrow defeats all consolation." Caesar and Brutus as well as Servius Sulpicius Rufus sent him letters of condolence.
Cicero hoped that his son Marcus would become a philosopher like him, but Marcus himself wished for a military career. He joined the army of Pompey in 49 BC and after Pompey's defeat at Pharsalus 48 BC, he was pardoned by Caesar. Cicero sent him to Athens to study as a disciple of the peripatetic philosopher Kratippos in 48 BC, but he used this absence from "his father's vigilant eye" to "eat, drink and be merry." After Cicero's murder he joined the army of the "Liberatores" but was later pardoned by Augustus. Augustus's bad conscience for not having objected to Cicero's being put on the proscription list during the Second Triumvirate led him to aid considerably Marcus Minor's career. He became an augur, and was nominated consul in 30 BC together with Augustus. As such, he was responsible for revoking the honors of Mark Antony, who was responsible for the proscription, and could in this way take revenge. Later he was appointed proconsul of Syria and the province of Asia.
Cicero wanted to pursue a public career in politics along the steps of the Cursus honorum. In 90–88 BC, he served both Pompeius Strabo and Lucius Cornelius Sulla as they campaigned in the Social War, though he had no taste for military life, being an intellectual first and foremost.
Cicero started his career as a lawyer around 83–81 BC. The first extant speech is a private case from 81 BC (the "pro Quinctio"), delivered when Cicero was aged 26, though he refers throughout to previous defenses he had already undertaken. His first major public case, of which a written record is still extant, was his 80 BC defense of Sextus Roscius on the charge of patricide. Taking this case was a courageous move for Cicero; patricide was considered an appalling crime, and the people whom Cicero accused of the murder, the most notorious being Chrysogonus, were favorites of Sulla. At this time it would have been easy for Sulla to have the unknown Cicero murdered. Cicero's defense was an indirect challenge to the dictator Sulla, and on the strength of his case, Roscius was acquitted. Soon after, Cicero again challenged Sulla, by criticising his disenfranchisement of Italian towns in a lost speech on behalf of a woman from Arretium.
Cicero's case in the "Pro Roscio Amerino" was divided into three parts. The first part detailed exactly the charge brought by Ericius. Cicero explained how a rustic son of a farmer, who lives off the pleasures of his own land, would not have gained anything from committing patricide because he would have eventually inherited his father's land anyway. The second part concerned the boldness and greed of two of the accusers, Magnus and Capito. Cicero told the jury that they were the more likely perpetrators of murder because the two were greedy, both for conspiring together against a fellow kinsman and, in particular, Magnus, for his boldness and for being unashamed to appear in court to support the false charges. The third part explained that Chrysogonus had immense political power, and the accusation was successfully made due to that power. Even though Chrysogonus may not have been what Cicero said he was, through rhetoric Cicero successfully made him appear to be a foreign freed man who prospered by devious means in the aftermath of the civil war. Cicero surmised that it showed what kind of a person he was and that something like murder was not beneath him.
His first office was as one of the twenty annual quaestors, a training post for serious public administration in a diversity of areas, but with a traditional emphasis on administration and rigorous accounting of public monies under the guidance of a senior magistrate or provincial commander. Cicero served as quaestor in western Sicily in 75 BC and demonstrated honesty and integrity in his dealings with the inhabitants. As a result, the grateful Sicilians asked Cicero to prosecute Gaius Verres, a governor of Sicily, who had badly plundered the province. His prosecution of Gaius Verres was a great forensic success for Cicero. Governor Gaius Verres hired the prominent lawyer of a noble family Quintus Hortensius Hortalus. After a lengthy period in Sicily collecting testimonials and evidence and persuading witnesses to come forward, Cicero returned to Rome and won the case in a series of dramatic court battles. His unique style of oratory set him apart from the flamboyant Hortensius. On the conclusion of this case, Cicero came to be considered the greatest orator in Rome. The view that Cicero may have taken the case for reasons of his own is viable. Hortensius was, at this point, known as the best lawyer in Rome; to beat him would guarantee much success and the prestige that Cicero needed to start his career. Cicero's oratorical skill is shown in his character assassination of Verres and various other techniques of persuasion used on the jury. One such example is found in the speech "Against Verres I", where he states "with you on this bench, gentlemen, with Marcus Acilius Glabrio as your president, I do not understand what Verres can hope to achieve". Oratory was considered a great art in ancient Rome and an important tool for disseminating knowledge and promoting oneself in elections, in part because there were no regular newspapers or mass media. Cicero was neither a patrician nor a plebeian noble; his rise to political office despite his relatively humble origins has traditionally been attributed to his brilliance as an orator.
Cicero grew up in a time of civil unrest and war. Sulla's victory in the first of a series of civil wars led to a new constitutional framework that undermined "libertas" (liberty), the fundamental value of the Roman Republic. Nonetheless, Sulla's reforms strengthened the position of the equestrian class, contributing to that class's growing political power. Cicero was both an Italian "eques" and a "novus homo", but more importantly he was a Roman constitutionalist. His social class and loyalty to the Republic ensured that he would "command the support and confidence of the people as well as the Italian middle classes". The "optimates" faction never truly accepted Cicero; and this undermined his efforts to reform the Republic while preserving the constitution. Nevertheless, he successfully ascended the "cursus honorum", holding each magistracy at or near the youngest possible age: quaestor in 75 BC (age 30), aedile in 69 BC (age 36), and praetor in 66 BC (age 39), when he served as president of the "Reclamation" (or extortion) Court. He was then elected consul at age 42.
Cicero, seizing the opportunity offered by optimate fear of reform, was elected consul for the year 63 BC; he was elected with the support of every unit of the centuriate assembly, rival members of the post-Sullan establishment, and the leaders of municipalities throughout post-Social War Italy. His co-consul for the year, Gaius Antonius Hybrida, played a minor role.
He began his consular year by opposing a land bill proposed by a plebeian tribune which would have appointed commissioners with semi-permanent authority over land reform. Cicero was also active in the courts, defending Gaius Rabirius from accusations of participating in the unlawful killing of plebeian tribune Lucius Appuleius Saturninus in 100 BC. The prosecution occurred before the "comita centuriata" and threatened to reopen conflict between the Marian and Sullan factions at Rome. Cicero defended the use of force as being authorised by a "senatus consultum ultimum", which would prove similar to his own use of force under such conditions.
Most famously—in part because of his own publicity—he thwarted a conspiracy led by Lucius Sergius Catilina to overthrow the Roman Republic with the help of foreign armed forces. Cicero procured a "senatus consultum ultimum" (a recommendation from the senate attempting to legitimise the use of force) and drove Catiline from the city with four vehement speeches (the Catiline Orations), which to this day remain outstanding examples of his rhetorical style. The Orations listed Catiline and his followers' debaucheries, and denounced Catiline's senatorial sympathizers as roguish and dissolute debtors clinging to Catiline as a final and desperate hope. Cicero demanded that Catiline and his followers leave the city. At the conclusion of his first speech, Catiline hurriedly left the Senate, (which was being held in the Temple of Jupiter Stator). In his following speeches, Cicero did not directly address Catiline. He delivered the second and third orations before the people, and the last one again before the Senate. By these speeches, Cicero wanted to prepare the Senate for the worst possible case; he also delivered more evidence, against Catiline.
Catiline fled and left behind his followers to start the revolution from within while he himself assaulted the city with an army of "moral bankrupts and honest fanatics". It is alleged that Catiline had attempted to involve the Allobroges, a tribe of Transalpine Gaul, in their plot, but Cicero, working with the Gauls, was able to seize letters that incriminated the five conspirators and forced them to confess in front of the senate. The senate then deliberated upon the conspirators' punishment. As it was the dominant advisory body to the various legislative assemblies rather than a judicial body, there were limits to its power; however, martial law was in effect, and it was feared that simple house arrest or exile – the standard options – would not remove the threat to the state. At first Decimus Junius Silanus spoke for the "extreme penalty"; many were swayed by Julius Caesar, who decried the precedent it would set and argued in favor of life imprisonment in various Italian towns. Cato the Younger rose in defense of the death penalty and the entire Senate finally agreed on the matter. Cicero had the conspirators taken to the Tullianum, the notorious Roman prison, where they were strangled. Cicero himself accompanied the former consul Publius Cornelius Lentulus Sura, one of the conspirators, to the Tullianum.
Cicero received the honorific ""pater patriae"" for his efforts to suppress the conspiracy, but lived thereafter in fear of trial or exile for having put Roman citizens to death without trial. While the "senatus consultum ultimum" gave some legitimacy to the use of force against the conspirators, Cicero also argued that Catiline's conspiracy, by virtue of its treason, made the conspirators enemies of the state and forfeited the protections intrinsically possessed by Roman citizens. The consuls moved decisively. Antonius Hybrida was dispatched to defeat Catiline in battle that year, preventing Crassus or Pompey from exploiting the situation for their own political aims.
After the suppression of the conspiracy, Cicero was proud of his accomplishment. Some of his political enemies argued that though the act gained Cicero popularity, he exaggerated the extent of his success. He overestimated his popularity again several years later after being exiled from Italy and then allowed back from exile. At this time, he claimed that the republic would be restored along with him. Many Romans at the time, led by Populares politicians Gaius Julius Caesar and patrician turned plebeian Publius Clodius Pulcher believed that Cicero's evidence against Catiline was fabricated and the witnesses were bribed. Cicero, who had been elected consul with the support of the Optimates, promoted their position as advocates of the status quo resisting social changes, especially more privileges for the average inhabitants of Rome.
Shortly after completing his consulship, in late 62 BC, Cicero arranged the purchase of a large townhouse on the Palatine Hill previously owned by Rome's richest citizen, Marcus Licinius Crassus. It cost an exorbitant sum, 3.5 million sesterces, which required Cicero to arrange for a loan from his co-consul Gaius Antonius Hybrida based on the expected profits from Antonius's proconsulship in Macedonia. At the beginning of his consulship, Cicero had made an arrangement with Hybrida to grant Hybrida the profitable province of Macedonia that had been granted to Cicero by the Senate in exchange for Hybrida staying out of Cicero's way for the year and a quarter of the profits from the province. In return Cicero gained a lavish house which he proudly boasted was ""in conspectu prope totius urbis"" (in sight of nearly the whole city), only a short walk away from the Roman Forum.
In 60 BC, Julius Caesar invited Cicero to be the fourth member of his existing partnership with Pompey and Marcus Licinius Crassus, an assembly that would eventually be called the First Triumvirate. Cicero refused the invitation because he suspected it would undermine the Republic.
During Caesar's consulship of 59 BC, the triumvirate had achieved many of their goals of land reform, publicani debt forgiveness, ratification of Pompeian conquests, etc. With Caesar leaving for his provinces, they wished to maintain their stranglehold on politics. They engineered the adoption of patrician Publius Clodius Pulcher into a plebeian family and had him elected as of the ten tribunes of the plebs for 58 BC. Clodius used the triumvirate's backing to push through legislation that benefited them all. He introduced several laws (the "leges Clodiae") that made him very popular with the people, strengthening his power base, then he turned on Cicero by threatening exile to anyone who executed a Roman citizen without a trial. Cicero, having executed members of the Catiline conspiracy four years previously without formal trial, was clearly the intended target. Furthermore, many believed that Clodius acted in concert with the triumvirate who feared that Cicero would seek to abolish many of Caesar's accomplishments while consul the year before. Cicero argued that the "senatus consultum ultimum" indemnified him from punishment, and he attempted to gain the support of the senators and consuls, especially of Pompey.
Cicero grew out his hair, dressed in mourning and toured the streets. Clodius' gangs dogged him, hurling abuse, stones and even excrement. Hortensius, trying to rally to his old rival's support, was almost lynched. The Senate and the consuls were cowed. Caesar, who was still encamped near Rome, was apologetic but said he could do nothing when Cicero brought himself to grovel in the proconsul's tent. Everyone seemed to have abandoned Cicero.
After Clodius passed a law to deny to Cicero fire and water (i.e. shelter) within four hundred miles of Rome, Cicero went into exile. He arrived at Thessalonica, on 23 May 58 BC. In his absence, Clodius, who lived next door to Cicero on the Palatine, arranged for Cicero's house to be confiscated by the state, and was even able to purchase a part of the property in order to extend his own house. After demolishing Cicero's house, Clodius had the land consecrated and symbolically erected a temple of Liberty ("aedes Libertatis") on the vacant spot.
Cicero's exile caused him to fall into depression. He wrote to Atticus: "Your pleas have prevented me from committing suicide. But what is there to live for? Don't blame me for complaining. My afflictions surpass any you ever heard of earlier". After the intervention of recently elected tribune Titus Annius Milo, acting on the behalf of Pompey who wanted Cicero as a client, the senate voted in favor of recalling Cicero from exile. Clodius cast the single vote against the decree. Cicero returned to Italy on 5 August 57 BC, landing at Brundisium. He was greeted by a cheering crowd, and, to his delight, his beloved daughter Tullia. In his "Oratio De Domo Sua Ad Pontifices", Cicero convinced the College of Pontiffs to rule that the consecration of his land was invalid, thereby allowing him to regain his property and rebuild his house on the Palatine.
Cicero tried to re-enter politics as an independent operator, but his attempts to attack portions of Caesar's legislation were unsuccessful and encouraged Caesar to re-solidify his political alliance with Pompey and Crassus. The conference at Luca in 56 BC left the three-man alliance in domination of the republic's politics; this forced Cicero to recant and support the triumvirate out of fear from being entirely excluded from public life. After the conference Cicero lavishly praised Caesar's achievements, got the Senate to vote a thanksgiving for Caesar's victories and grant money to pay his troops. He also delivered a speech 'On the consular provinces' () which checked an attempt by Caesar's enemies to strip him of his provinces in Gaul. After this, a cowed Cicero concentrated on his literary works. It is uncertain whether he was directly involved in politics for the following few years.
In 51 BC he reluctantly accepted a promagistracy (as proconsul) in Cilicia for the year; there were few other former consuls eligible as a result of a legislative requirement enacted by Pompey in 52 BC specifying an interval of five years between a consulship or praetorship and a provincial command. He served as proconsul of Cilicia from May 51, arriving in the provinces three months later around August. He was given instructions to keep nearby Cappadocia loyal to King Ariobarzanes III, which he achieved 'satisfactorily without war'. In 53 BC Marcus Licinius Crassus had been defeated by the Parthians at the Battle of Carrhae. This opened up the Roman East for a Parthian invasion causing much unrest in Syria and Cilicia. Cicero restored calm by his mild system of government. He discovered that much of public property had been embezzled by corrupt previous governors and their staffs, and did his utmost to restore it. Thus he greatly improved the condition of the cities. He retained the civil rights of, and exempted from penalties, the men who gave the property back. Besides this, he was extremely frugal in his outlays for staff and private expenses during his governorship, and this made him highly popular among the natives. Previous governors had extorted enormous sums from the provincials in order to supply their households and bodyguard.
Besides his activity in ameliorating the hard pecuniary situation of the province, Cicero was also creditably active in the military sphere. Early in his governorship he received information that prince Pacorus, son of Orodes II the king of the Parthians, had crossed the Euphrates, and was ravaging the Syrian countryside and had even besieged Cassius (the interim Roman commander in Syria) in Antioch. Cicero eventually marched with two understrength legions and a large contingent of auxiliary cavalry to Cassius's relief. Pacorus and his army had already given up on besieging Antioch and were heading south through Syria, ravaging the countryside again, Cassius and his legions followed them, harrying them wherever they went, eventually ambushing and defeating them near Antigonea. Another large troop of Parthian horsemen was defeated by Cicero's cavalry who happened to run into them while scouting ahead of the main army. Cicero next defeated some robbers who were based on Mount Amanus and was hailed as imperator by his troops. Afterwards he led his army against the independent Cilician mountain tribes, besieging their fortress of Pindenissum. It took him 47 days to reduce the place, which fell in December. Then Cicero left the province on 30 July to his brother Quintus, who had accompanied him on his governorship as his legate. On his way back to Rome he stopped over in Rhodes and then went to Athens, where he caught up with his old friend Titus Pomponius Atticus and met men of great learning.
Cicero arrived in Rome on 4 January 49 BC. He stayed outside the pomerium, to retain his promagisterial powers: either in expectation of a triumph or to retain his independent command authority in the coming civil war. The struggle between Pompey and Julius Caesar grew more intense in 50 BC. Cicero favored Pompey, seeing him as a defender of the senate and Republican tradition, but at that time avoided openly alienating Caesar. When Caesar invaded Italy in 49 BC, Cicero fled Rome. Caesar, seeking an endorsement by a senior senator, courted Cicero's favor, but even so Cicero slipped out of Italy and traveled to Dyrrachium (Epidamnos), Illyria, where Pompey's staff was situated. Cicero traveled with the Pompeian forces to Pharsalus in 48 BC, though he was quickly losing faith in the competence and righteousness of the Pompeian side. Eventually, he provoked the hostility of his fellow senator Cato, who told him that he would have been of more use to the cause of the "optimates" if he had stayed in Rome. After Caesar's victory at the Battle of Pharsalus on 9 August, Cicero refused to take command of the Pompeian forces and continue the war. He returned to Rome, still as a promagistrate with his lictors, in 47 BC, and dismissed them upon his crossing the pomerium and renouncing his command. Caesar pardoned him and Cicero tried to adjust to the situation and maintain his political work, hoping that Caesar might revive the Republic and its institutions.
In a letter to Varro on c. 20 April 46 BC, Cicero outlined his strategy under Caesar's dictatorship. Cicero, however, was taken completely by surprise when the "Liberatores" assassinated Caesar on the ides of March, 44 BC. Cicero was not included in the conspiracy, even though the conspirators were sure of his sympathy. Marcus Junius Brutus called out Cicero's name, asking him to restore the republic when he lifted his bloodstained dagger after the assassination. A letter Cicero wrote in February 43 BC to Trebonius, one of the conspirators, began, "How I could wish that you had invited me to that most glorious banquet on the Ides of March"! Cicero became a popular leader during the period of instability following the assassination. He had no respect for Mark Antony, who was scheming to take revenge upon Caesar's murderers. In exchange for amnesty for the assassins, he arranged for the Senate to agree not to declare Caesar to have been a tyrant, which allowed the Caesarians to have lawful support and kept Caesar's reforms and policies intact.
Cicero and Antony now became the two leading men in Rome: Cicero as spokesman for the Senate; Antony as consul, leader of the Caesarian faction, and unofficial executor of Caesar's public will. Relations between the two, never friendly, worsened after Cicero claimed that Antony was taking liberties in interpreting Caesar's wishes and intentions. Octavian was Caesar's adopted son and heir. After he returned to Italy, Cicero began to play him against Antony. He praised Octavian, declaring he would not make the same mistakes as his father. He attacked Antony in a series of speeches he called the Philippics, after Demosthenes's denunciations of Philip II of Macedon. At the time Cicero's popularity as a public figure was unrivalled.
Cicero supported Decimus Junius Brutus Albinus as governor of Cisalpine Gaul ("Gallia Cisalpina") and urged the Senate to name Antony an enemy of the state. The speech of Lucius Piso, Caesar's father-in-law, delayed proceedings against Antony. Antony was later declared an enemy of the state when he refused to lift the siege of Mutina, which was in the hands of Decimus Brutus. Cicero's plan to drive out Antony failed. Antony and Octavian reconciled and allied with Lepidus to form the Second Triumvirate after the successive battles of Forum Gallorum and Mutina. The Triumvirate began proscribing their enemies and potential rivals immediately after legislating the alliance into official existence for a term of five years with consular "imperium". Cicero and all of his contacts and supporters were numbered among the enemies of the state, even though Octavian argued for two days against Cicero being added to the list.
Cicero was one of the most viciously and doggedly hunted among the proscribed. He was viewed with sympathy by a large segment of the public and many people refused to report that they had seen him. He was caught 7 December 43 BC leaving his villa in Formiae in a litter going to the seaside where he hoped to embark on a ship destined for Macedonia. When his killers – Herennius (a centurion) and Popilius (a tribune) – arrived, Cicero's own slaves said they had not seen him, but he was given away by Philologus, a freedman of his brother Quintus Cicero.
Cicero's last words are said by Cassius Dio to have been "There is nothing proper about what you are doing, soldier, but do try to kill me properly." He bowed to his captors, leaning his head out of the litter in a gladiatorial gesture to ease the task. By baring his neck and throat to the soldiers, he was indicating that he would not resist. According to Plutarch, Herennius first slew him, then cut off his head. On Antony's instructions his hands, which had penned the Philippics against Antony, were cut off as well; these were nailed along with his head on the Rostra in the Forum Romanum according to the tradition of Marius and Sulla, both of whom had displayed the heads of their enemies in the Forum. Cicero was the only victim of the proscriptions who was displayed in that manner. According to Cassius Dio (in a story often mistakenly attributed to Plutarch), Antony's wife Fulvia took Cicero's head, pulled out his tongue, and jabbed it repeatedly with her hairpin in final revenge against Cicero's power of speech.
Cicero's son, Marcus Tullius Cicero Minor, during his year as a consul in 30 BC, avenged his father's death, to a certain extent, when he announced to the Senate Mark Antony's naval defeat at Actium in 31 BC by Octavian and his capable commander-in-chief, Agrippa.
Octavian is reported to have praised Cicero as a patriot and a scholar of meaning in later times, within the circle of his family. However, it was Octavian's acquiescence that had allowed Cicero to be killed, as Cicero was proscribed by the new triumvirate.
Cicero's career as a statesman was marked by inconsistencies and a tendency to shift his position in response to changes in the political climate. His indecision may be attributed to his sensitive and impressionable personality; he was prone to overreaction in the face of political and private change.
"Would that he had been able to endure prosperity with greater self-control, and adversity with more fortitude!" wrote C. Asinius Pollio, a contemporary Roman statesman and historian.
Cicero has been traditionally considered the master of Latin prose, with Quintilian declaring that Cicero was "not the name of a man, but of eloquence itself." The English words "Ciceronian" (meaning "eloquent") and "cicerone" (meaning "local guide") derive from his name. He is credited with transforming Latin from a modest utilitarian language into a versatile literary medium capable of expressing abstract and complicated thoughts with clarity. Julius Caesar praised Cicero's achievement by saying "it is more important to have greatly extended the frontiers of the Roman spirit ("ingenium") than the frontiers of the Roman empire". According to John William Mackail, "Cicero's unique and imperishable glory is that he created the language of the civilized world, and used that language to create a style which nineteen centuries have not replaced, and in some respects have hardly altered."
Cicero was also an energetic writer with an interest in a wide variety of subjects, in keeping with the Hellenistic philosophical and rhetorical traditions in which he was trained. The quality and ready accessibility of Ciceronian texts favored very wide distribution and inclusion in teaching curricula, as suggested by a graffito at Pompeii, admonishing: "You will like Cicero, or you will be whipped".
Cicero was greatly admired by influential Church Fathers such as Augustine of Hippo, who credited Cicero's lost "Hortensius" for his eventual conversion to Christianity, and St. Jerome, who had a feverish vision in which he was accused of being "follower of Cicero and not of Christ" before the judgment seat.
This influence further increased after the Early Middle Ages in Europe, which more of his writings survived than any other Latin author. Medieval philosophers were influenced by Cicero's writings on natural law and innate rights.
Petrarch's rediscovery of Cicero's letters provided the impetus for searches for ancient Greek and Latin writings scattered throughout European monasteries, and the subsequent rediscovery of classical antiquity led to the Renaissance. Subsequently, Cicero became synonymous with classical Latin to such an extent that a number of humanist scholars began to assert that no Latin word or phrase should be used unless it appeared in Cicero's works, a stance criticised by Erasmus.
His voluminous correspondence, much of it addressed to his friend Atticus, has been especially influential, introducing the art of refined letter writing to European culture. Cornelius Nepos, the 1st century BC biographer of Atticus, remarked that Cicero's letters contained such a wealth of detail "concerning the inclinations of leading men, the faults of the generals, and the revolutions in the government" that their reader had little need for a history of the period.
Among Cicero's admirers were Desiderius Erasmus, Martin Luther, and John Locke. Following the invention of Johannes Gutenberg's printing press, "De Officiis" was the second book printed in Europe, after the Gutenberg Bible. Scholars note Cicero's influence on the rebirth of religious toleration in the 17th century.
Cicero was especially popular with the Philosophes of the 18th century, including Edward Gibbon, Diderot, David Hume, Montesquieu, and Voltaire. Gibbon wrote of his first experience reading the author's collective works thus: "I tasted the beauty of the language; I breathed the spirit of freedom; and I imbibed from his precepts and examples the public and private sense of a man...after finishing the great author, a library of eloquence and reason, I formed a more extensive plan of reviewing the Latin classics..." Voltaire called Cicero "the greatest as well as the most elegant of Roman philosophers" and even staged a play based on Cicero's role in the Catilinarian conspiracy, called "Rome Sauvée, ou Catilina", to "make young people who go to the theatre acquainted with Cicero." Voltaire was spurred to pen the drama as a rebuff to his rival Claude Prosper Jolyot de Crébillon's own play "Catilina", which had portrayed Cicero as a coward and villain who hypocritically married his own daughter to Catiline. Montesquieu produced his "Discourse on Cicero" in 1717, in which he heaped praise on the author because he rescued "philosophy from the hands of scholars, and freed it from the confusion of a foreign language". Montesquieu went on to declare that Cicero was "of all the ancients, the one who had the most personal merit, and whom I would prefer to resemble."
Internationally, Cicero the republican inspired the Founding Fathers of the United States and the revolutionaries of the French Revolution. John Adams said, "As all the ages of the world have not produced a greater statesman and philosopher united than Cicero, his authority should have great weight." Jefferson names Cicero as one of a handful of major figures who contributed to a tradition "of public right" that informed his draft of the Declaration of Independence and shaped American understandings of "the common sense" basis for the right of revolution. Camille Desmoulins said of the French republicans in 1789 that they were "mostly young people who, nourished by the reading of Cicero at school, had become passionate enthusiasts for liberty".
Jim Powell starts his book on the history of liberty with the sentence: "Marcus Tullius Cicero expressed principles that became the bedrock of liberty in the modern world."
Likewise, no other ancient personality has inspired as much venomous dislike as Cicero, especially in more modern times. His commitment to the values of the Republic accommodated a hatred of the poor and persistent opposition to the advocates and mechanisms of popular representation. Friedrich Engels referred to him as "the most contemptible scoundrel in history" for upholding republican "democracy" while at the same time denouncing land and class reforms. Cicero has faced criticism for exaggerating the democratic qualities of republican Rome, and for defending the Roman oligarchy against the popular reforms of Caesar. Michael Parenti admits Cicero's abilities as an orator, but finds him a vain, pompous and hypocritical personality who, when it suited him, could show public support for popular causes that he privately despised. Parenti presents Cicero's prosecution of the Catiline conspiracy as legally flawed at least, and possibly unlawful.
Cicero also had an influence on modern astronomy. Nicolaus Copernicus, searching for ancient views on earth motion, said that he "first ... found in Cicero that Hicetas supposed the earth to move."
Cicero was declared a righteous pagan by the Early Church, and therefore many of his works were deemed worthy of preservation. The Bogomils considered him a rare exception of a pagan saint. Subsequent Roman and medieval Christian writers quoted liberally from his works "De Re Publica" ("On the Commonwealth") and "De Legibus" ("On the Laws"), and much of his work has been recreated from these surviving fragments. Cicero also articulated an early, abstract conceptualization of rights, based on ancient law and custom. Of Cicero's books, six on rhetoric have survived, as well as parts of eight on philosophy. Of his speeches, 88 were recorded, but only 58 survive.
Cicero's great repute in Italy has led to numerous ruins being identified as having belonged to him, though none have been substantiated with absolute certainty. In Formia, two Roman-era ruins are popularly believed to be Cicero's mausoleum, the "Tomba di Cicerone", and the villa where he was assassinated in 43 BC. The latter building is centered around a central hall with Doric columns and a coffered vault, with a separate nymphaeum, on five acres of land near Formia. A modern villa was built on the site after the Rubino family purchased the land from Ferdinand II of the Two Sicilies in 1868. Cicero's supposed tomb is a 24-meter (79 feet) tall tower on an "opus quadratum" base on the ancient Via Appia outside of Formia. Some suggest that it is not in fact Cicero's tomb, but a monument built on the spot where Cicero was intercepted and assassinated while trying to reach the sea.
In Pompeii, a large villa excavated in the mid 18th century just outside the Herculaneum Gate was widely believed to have been Cicero's, who was known to have owned a holiday villa in Pompeii he called his "Pompeianum". The villa was stripped of its fine frescoes and mosaics and then re-buried after 1763 – it has yet to be re-excavated. However, contemporaneous descriptions of the building from the excavators combined with Cicero's own references to his "Pompeianum" differ, making it unlikely that it is Cicero's villa.
In Rome, the location of Cicero's house has been roughly identified from excavations of the Republican-era stratum on the northwestern slope of the Palatine Hill. Cicero's "domus" has long been known to have stood in the area, according to his own descriptions and those of later authors, but there is some debate about whether it stood near the base of the hill, very close to the Roman Forum, or nearer to the summit. During his life the area was the most desirable in Rome, densely occupied with Patrician houses including the "Domus Publica" of Julius Caesar and the home of Cicero's mortal enemy Clodius. In the early Imperial era these properties fell into the possession of the Julio-Claudian dynasty and the substructures of the "Domus Tiberiana" were built over the Republican-era buildings.
Ben Jonson dramatised the conspiracy of Catiline in his play "Catiline His Conspiracy", featuring Cicero as a character. Cicero also appears as a minor character in William Shakespeare's play "Julius Caesar".
Cicero was portrayed on the motion picture screen by British actor Alan Napier in the 1953 film "Julius Caesar", based on Shakespeare's play. He has also been played by such noted actors as Michael Hordern (in "Cleopatra"), and André Morell (in the 1970 "Julius Caesar"). Most recently, Cicero was portrayed by David Bamber in the HBO series "Rome" (2005–2007) and appeared in both seasons.
In the historical novel series "Masters of Rome", Colleen McCullough presents a not-so-flattering depiction of Cicero's career, showing him struggling with an inferiority complex and vanity, morally flexible and fatally indiscreet, while his rival Julius Caesar is shown in a more approving light. Cicero is portrayed as a hero in the novel "A Pillar of Iron" by Taylor Caldwell (1965). Robert Harris' novels "Imperium", "Lustrum" (published under the name "Conspirata" in the United States) and "Dictator" comprise a three-part series based on the life of Cicero. In these novels Cicero's character is depicted in a more favorable way than in those of McCullough, with his positive traits equaling or outweighing his weaknesses (while conversely Caesar is depicted as more sinister than in McCullough). Cicero is a major recurring character in the "Roma Sub Rosa" series of mystery novels by Steven Saylor. He also appears several times as a peripheral character in John Maddox Roberts' "SPQR" series.
Samuel Barnett portrays Cicero in a 2017 audio drama series pilot produced by Big Finish Productions. A full series was released the following year. All Episodes are written by David Llewellyn and directed and produced by Scott Handcock. Llewellyn, Handcock and Barnett re-teamed in the Doctor Who audio-drama Tartarus (also produced by Big Finish) starring Peter Davison as the 5th Doctor. It is not intended to be apart the Cicero series, in Vortex (Big Finish's official free online magazine) Llewellyn revealed that he was “ worried that if we had Cicero meeting aliens people might go back to the Cicero series and see it through a sci-fi lens. Then I remembered that Simon Callow still performs as Charles Dickens, and that he played Dickens before reprising him in the Doctor Who TV episode, "The Unquiet Dead" – so I got over myself!". it was released in September 2019. | https://en.wikipedia.org/wiki?curid=6046 |
Consul
Consul (abbrev. "cos."; Latin plural "consules") was the title of one of the two chief magistrates of the Roman Republic, and subsequently also an important title under the Roman Empire. The title was used in other European city states through antiquity and the Middle Ages, then revived in modern states, notably in the First French Republic. The related adjective is consular, from the Latin "consularis".
This usage contrasts with modern terminology, where a consul is a type of diplomat.
A consul held the highest elected political office of the Roman Republic (509 to 27 BC), and ancient Romans considered the consulship the highest level of the cursus honorum (an ascending sequence of public offices to which politicians aspired). Consuls were elected to office and held power for one year. There were always two consuls in power at any time.
It was not uncommon for an organization under Roman private law to copy the terminology of state and city institutions for its own statutory agents. The founding statute, or contract, of such an organisation was called "lex", 'law'. The people elected each year were patricians, members of the upper class.
While many cities, including the Gallic states and the Carthaginian Republic, had a double-headed chief magistracy, another title was often used, such as the Punic "sufet", "Duumvir", or native styles like "Meddix".
Throughout most of southern France, a consul ( or "") was an office equivalent to the of the north and roughly similar with English aldermen. The most prominent were those of Bordeaux and Toulouse, which came to be known as jurats and capitouls, respectively. The capitouls of Toulouse were granted transmittable nobility. In many other smaller towns the first consul was the equivalent of a mayor today, assisted by a variable number of secondary consuls and jurats. His main task was to levy and collect tax.
The Dukes of Gaeta often used also the title of "consul" in its Greek form "Hypatos" (see List of Hypati and Dukes of Gaeta).
The city-state of Genoa, unlike ancient Rome, bestowed the title of "consul" on various state officials, not necessarily restricted to the highest. Among these were Genoese officials stationed in various Mediterranean ports, whose role included helping Genoese merchants and sailors in difficulties with the local authorities. This institution, with its name, was later emulated by other powers and is reflected in the modern usage of the word (see Consul (representative)).
After Napoleon Bonaparte staged a coup against the Directory government in November 1799, the French Republic adopted a constitution which conferred executive powers upon three consuls, elected for a period of ten years. In reality, the first consul, Bonaparte, dominated his two colleagues and held supreme power, soon making himself consul for life (1802) and eventually, in 1804, emperor.
The office was held by:
The short-lived Bolognese Republic, proclaimed in 1796 as a French client republic in the Central Italian city of Bologna, had a government consisting of nine consuls and its head of state was the "Presidente del Magistrato", i.e., chief magistrate, a presiding office held for four months by one of the consuls. Bologna already had consuls at some parts of its Medieval history.
The French-sponsored Roman Republic (15 February 1798 – 23 June 1800) was headed by multiple consuls:
Consular rule was interrupted by the Neapolitan occupation (27 November – 12 December 1798), which installed a Provisional Government:
Rome was occupied by France (11 July – 28 September 1799) and again by Naples (30 September 1799 – 23 June 1800), bringing an end to the Roman Republic.
Among the many petty local republics that were formed during the first year of the Greek Revolution, prior to the creation of a unified Provisional Government at the First National Assembly at Epidaurus, were:
"Note: in Greek, the term for "consul" is "hypatos" (ὕπατος), which translates as "supreme one", and hence does not necessarily imply a joint office."
In between a series of juntas and various other short-lived regimes, the young republic was governed by "consuls of the republic", with two consuls alternating in power every 4 months:
After a few presidents of the Provisional Junta, there were again consuls of the republic, 14 March 1841 – 13 March 1844 (ruling jointly, but occasionally styled "first consul", "second consul"): Carlos Antonio López Ynsfrán (b. 1792 – d. 1862) + Mariano Roque Alonzo Romero (d. 1853) (the lasts of the aforementioned juntistas, Commandant-General of the Army)
Thereafter all republican rulers were styled "president".
In modern terminology, a consul is a type of diplomat. The "American Heritage Dictionary" defines consul as "an official appointed by a government to reside in a foreign country and represent its interests there."
In most governments, the consul is the head of the consular section of an embassy, and is responsible for all consular services such as immigrant and non-immigrant visas, passports, and citizen services for expatriates living or traveling in the host country.
A less common modern usage is when the consul of one country takes a governing role in the host country.
Differently named, but same function
Modern UN System
Specific | https://en.wikipedia.org/wiki?curid=6047 |
List of equations in classical mechanics
Classical mechanics is the branch of physics used to describe the motion of macroscopic objects. It is the most familiar of the theories of physics. The concepts it covers, such as mass, acceleration, and force, are commonly used and known. The subject is based upon a three-dimensional Euclidean space with fixed axes, called a frame of reference. The point of concurrency of the three axes is known as the origin of the particular space.
Classical mechanics utilises many equations—as well as other mathematical concepts—which relate various physical quantities to one another. These include differential equations, manifolds, Lie groups, and ergodic theory. This page gives a summary of the most important of these.
This article lists equations from Newtonian mechanics, see analytical mechanics for the more general formulation of classical mechanics (which includes Lagrangian and Hamiltonian mechanics).
Every conservative force has a potential energy. By following two principles one can consistently assign a non-relative value to "U":
In the following rotational definitions, the angle can be any angle about the specified axis of rotation. It is customary to use "θ", but this does not have to be the polar angle used in polar coordinate systems. The unit axial vector
defines the axis of rotation, formula_2 = unit vector in direction of r, formula_3 = unit vector tangential to the angle.
!
! Translation
! Rotation
!Velocity
Instantaneous:
Instantaneous:
Instantaneous: | https://en.wikipedia.org/wiki?curid=6050 |
Cursus honorum
The cursus honorum (; , or more colloquially 'ladder of offices') was the sequential order of public offices held by aspiring politicians in the Roman Republic and the early Roman Empire. It was designed for men of senatorial rank. The "cursus honorum" comprised a mixture of military and political administration posts. Each office had a minimum age for election. There were minimum intervals between holding successive offices and laws forbade repeating an office.
These rules were altered and flagrantly ignored in the course of the last century of the Republic. For example, Gaius Marius held consulships for five years in a row between 104 BC and 100 BC. He was consul seven times in all, also serving in 107 and 86. Officially presented as opportunities for public service, the offices often became mere opportunities for self-aggrandizement. The reforms of Sulla required a ten-year interval before holding the same office again for another term.
To have held each office at the youngest possible age ("suo anno", "in his own year") was considered a great political success. For instance, to miss out on a praetorship at 39 meant that one could not become consul at 42. Cicero expressed extreme pride not only in being a "novus homo" ("new man"; comparable to a "self-made man") who became consul even though none of his ancestors had ever served as a consul, but also in having become consul "in his year".
The "cursus honorum" began with ten years of military duty in the Roman cavalry (the "equites") or in the staff of a general who was a relative or a friend of the family. The ten years of service were intended to be mandatory in order to qualify for political office, but in practice, the rule was not always rigidly applied.
A more prestigious position was that of a military tribune. In the early Roman Republic, 24 men at the age of around 20 were elected by the Tribal Assembly to serve as commanders on a rotating basis; each commander retained six tribunes on his staff. Tribunes could also be appointed by the consuls or by military commanders in the field as necessary. After the reforms of Gaius Marius in 107 BC, the six tribunes acted as staff officers for the legionary "legatus" and were appointed tasks and command of units of troops whenever the need arose.
The subsequent steps of the "cursus honorum" were achieved by direct election every year.
The first official post was that of "quaestor". Candidates had to be at least 30 years old. However, men of patrician rank could subtract two years from this and other minimum age requirements.
Twenty quaestors served in the financial administration at Rome or as second-in-command to a governor in the provinces. They could also serve as the paymaster for a legion. A young man who obtained this job was expected to become a very important official. An additional task of all quaestors was the supervision of public games. As a quaestor, an official was allowed to wear the toga praetexta, but was not escorted by lictors, nor did he possess imperium.
At 36 years of age, proquaestor could stand for election to one of the "aedile" (pronounced , from "aedes", "temple edifice") positions. Of these aediles, two were plebeian and two were patrician, with the patrician aediles called Curule Aediles. The plebeian aediles were elected by the Plebeian Council and the curule aediles were either elected by the Tribal Assembly or appointed by the reigning consul. The aediles had administrative responsibilities in Rome. They had to take care of the temples (whence their title, from the Latin "aedes", "temple"), organize games, and be responsible for the maintenance of the public buildings in Rome. Moreover, they took charge of Rome's water and food supplies; in their capacity as market superintendents, they served sometimes as judges in mercantile affairs.
The Aedile was the supervisor of public works; the words "edifice" and "edification" stem from the same root. He oversaw the public works, temples and markets. Therefore, the Aediles would have been in some cooperation with the current Censors, who had similar or related duties. Also they oversaw the organization of festivals and games ("ludi"), which made this a very sought-after office for a career minded politician of the late republic, as it was a good means of gaining popularity by staging spectacles.
Curule Aediles were added at a later date in the 4th century BC, and their duties do not differ substantially from plebeian aediles. However, unlike plebeian aediles, curule aediles were allowed certain symbols of rank—the "sella curulis" or 'curule chair,' for example—and only patricians could stand for election to curule aedile. This later changed, and both Plebeians and Patricians could stand for Curule Aedileship.
The elections for Curule Aedile were at first alternated between Patricians and Plebeians, until late in the 2nd century BC, when the practice was abandoned and both classes became free to run during all years.
While part of the "cursus honorum", this step was optional and not required to hold future offices. Though the office was usually held after the quaestorship and before the praetorship, there are some cases with former praetors serving as aediles.
After serving either as quaestor or as aedile, a man of 39 years could run for "praetor". The number of praetors elected varied through history, generally increasing with time. During the republic, six or eight were generally elected each year to serve judicial functions throughout Rome and other governmental responsibilities. In the absence of the consuls, a praetor would be given command of the garrison in Rome or in Italy. Also, a praetor could exercise the functions of the consuls throughout Rome, but their main function was that of a judge. They would preside over trials involving criminal acts, grant court orders and validate "illegal" acts as acts of administering justice. A praetor was escorted by six lictors, and wielded "imperium". After a term as praetor, the magistrate would serve as a provincial governor with the title of propraetor, wielding propraetor imperium, commanding the province's legions, and possessing ultimate authority within his province(s).
Two of the praetors were more prestigious than the others. The first was the Praetor Peregrinus, who was the chief judge in trials involving one or more foreigners. The other was the Praetor Urbanus, the chief judicial office in Rome. He had the power to overturn any verdict by any other courts, and served as judge in cases involving criminal charges against provincial governors. The Praetor Urbanus was not allowed to leave the city for more than ten days. If one of these two Praetors was absent from Rome, the other would perform the duties of both.
The office of "consul" was the most prestigious of all of the offices on the cursus honorum, and represented the summit of a successful career. The minimum age was 42. Years were identified by the names of the two consuls elected for a particular year; for instance, "M. Messalla et M. Pisone consulibus", "in the consulship of Messalla and Piso," dates an event to 61 BC. Consuls were responsible for the city's political agenda, commanded large-scale armies and controlled important provinces. The consuls served for only a year (a restriction intended to limit the amassing of power by individuals) and could only rule when they agreed, because each consul could veto the other's decision.
The consuls would alternate monthly as the chairman of the Senate. They also were the supreme commanders in the Roman army, with each being granted two legions during their consular year. Consuls also exercised the highest juridical power in the Republic, being the only office with the power to override the decisions of the Praetor Urbanus. Only laws and the decrees of the Senate or the People's assembly limited their powers, and only the veto of a fellow consul or a tribune of the plebs could supersede their decisions.
A consul was escorted by twelve lictors, held imperium and wore the toga praetexta. Because the consul was the highest executive office within the Republic, they had the power to veto any action or proposal by any other magistrate, save that of the Tribune of the Plebs. After a consulship, a consul was assigned one of the more important provinces and acted as the governor in the same way that a Propraetor did, only owning Proconsular imperium. A second consulship could only be attempted after an interval of 10 years to prevent one man holding too much power.
Although not part of the Cursus Honorum, upon completing a term as either Praetor or Consul, an officer was required to serve a term as Propraetor and Proconsul, respectively, in one of Rome's many provinces. These Propraetors and Proconsuls held near autocratic authority within their selected province or provinces. Because each governor held equal imperium to the equivalent magistrate, they were escorted by the same number of lictors (12) and could only be vetoed by a reigning Consul or Praetor. Their abilities to govern were only limited by the decrees of the Senate or the people's assemblies, and the Tribune of the Plebs was unable to veto their acts as long as the governor remained at least a mile outside of Rome.
After a term as consul, the final step in the Cursus Honorum was the office of "censor". This was the only office in the Roman Republic whose term was a period of eighteen months instead of the usual twelve. Censors were elected every five years and although the office held no military imperium, it was considered a great honour. The censors took a regular census of the people and then apportioned the citizens into voting classes on the basis of income and tribal affiliation. The censors enrolled new citizens in tribes and voting classes as well. The censors were also in charge of the membership roll of the Senate, every five years adding new senators who had been elected to the requisite offices. Censors could also remove unworthy members from the Senate. This ability was lost during the dictatorship of Sulla. Censors were also responsible for construction of public buildings and the moral status of the city.
Censors also had financial duties, in that they had to put out to tender projects that were to be financed by the state. Also, the censors were in charge of the leasing out of conquered land for public use and auction. Though this office owned no imperium, meaning no lictors for protection, they were allowed to wear the "toga praetexta".
The office of Tribune of the Plebs was an important step in the political career of plebeians. Patricians could not hold the office. The Tribune was an office first created to protect the right of the common man in Roman politics and served as the head of the Plebeian Council. In the mid-to-late Republic, however, plebeians were often just as, and sometimes more, wealthy and powerful than patricians. Those who held the office were granted sacrosanctity (the right to be legally protected from any physical harm), the power to rescue any plebeian from the hands of a patrician magistrate, and the right to veto any act or proposal of any magistrate, including another tribune of the people and the consuls. The tribune also had the power to exercise capital punishment against any person who interfered in the performance of his duties. The tribunes could even convene a Senate meeting and lay legislation before it and arrest magistrates. Their houses had to remain open for visitors even during the night, and they were not allowed to be more than a day's journey from Rome. Due to their unique power of sacrosanctity, the Tribune had no need for lictors for protection and owned no imperium, nor could they wear the toga praetexta. For a period after Sulla's reforms, a person who had held the office of Tribune of the Plebs could no longer qualify for any other office, and the powers of the tribunes were more limited, but these restrictions were subsequently lifted.
Another office not officially a step in the "cursus honorum" was the "princeps senatus", an extremely prestigious office for a patrician. The "princeps senatus" served as the leader of the Senate and was chosen to serve a five-year term by each pair of Censors every five years. Censors could, however, confirm a "princeps senatus" for a period of another five years. The "princeps senatus" was chosen from all Patricians who had served as a Consul, with former Censors usually holding the office. The office originally granted the holder the ability to speak first at session on the topic presented by the presiding magistrate, but eventually gained the power to open and close the senate sessions, decide the agenda, decide where the session should take place, impose order and other rules of the session, meet in the name of the senate with embassies of foreign countries, and write in the name of the senate letters and dispatches. This office, like the Tribune, did not own "imperium", was not escorted by lictors, and could not wear the "toga praetexta".
Of all the offices within the Roman Republic, none granted as much power and authority as the position of "dictator", known as the Master of the People. In times of emergency, the Senate would declare that a dictator was required, and the current consuls would appoint a dictator. This was the only decision that could not be vetoed by the Tribune of the Plebs. The dictator was the sole exception to the Roman legal principles of having multiple magistrates in the same office and being legally able to be held to answer for actions in office. Essentially by definition, only one dictator could serve at a time, and no dictator could ever be held legally responsible for any action during his time in office for any reason.
The dictator was the highest magistrate in degree of imperium and was attended by twenty-four lictors (as were the former Kings of Rome). Although his term lasted only six months instead of twelve (except for the Dictatorships of Sulla and Caesar), all other magistrates reported to the dictator (except for the tribunes of the plebs - although they could not veto any of the dictator's acts), granting the dictator absolute authority in both civil and military matters throughout the Republic. The Dictator was free from the control of the Senate in all that he did, could execute anyone without a trial for any reason, and could ignore any law in the performance of his duties. The Dictator was the sole magistrate under the Republic that was truly independent in discharging his duties. All of the other offices were extensions of the Senate's executive authority and thus answerable to the Senate. Since the Dictator exercised his own authority, he did not suffer this limitation, which was the cornerstone of the office's power.
When a Dictator entered office, he appointed to serve as his second-in-command a "magister equitum", the Master of the Horse, whose office ceased to exist once the Dictator left office. The magister equitum held Praetorian imperium, was attended by six lictors, and was charged with assisting the Dictator in managing the State. When the Dictator was away from Rome, the magister equitum usually remained behind to administer the city. The magister equitum, like the Dictator, had unchallengeable authority in all civil and military affairs, with his decisions only being overturned by the Dictator himself.
The Dictatorship was definitively abolished in 44 BC after the assassination of Gaius Julius Caesar (Lex Antonia). | https://en.wikipedia.org/wiki?curid=6051 |
Continental drift
Continental drift is the theory that the Earth's continents have moved over geologic time relative to each other, thus appearing to have "drifted" across the ocean bed. The speculation that continents might have 'drifted' was first put forward by Abraham Ortelius in 1596. The concept was independently and more fully developed by Alfred Wegener in 1912, but his theory was rejected by many for lack of any motive mechanism. Arthur Holmes later proposed mantle convection for that mechanism. The idea of continental drift has since been subsumed by the theory of plate tectonics, which explains that the continents move by riding on plates of the Earth's lithosphere.
Abraham Ortelius , Theodor Christoph Lilienthal (1756), Alexander von Humboldt (1801 and 1845), Antonio Snider-Pellegrini , and others had noted earlier that the shapes of continents on opposite sides of the Atlantic Ocean (most notably, Africa and South America) seem to fit together. W. J. Kious described Ortelius' thoughts in this way:
In 1889, Alfred Russel Wallace remarked, "It was formerly a very general belief, even amongst geologists, that the great features of the earth's surface, no less than the smaller ones, were subject to continual mutations, and that during the course of known geological time the continents and great oceans had, again and again, changed places with each other." He quotes Charles Lyell as saying, "Continents, therefore, although permanent for whole geological epochs, shift their positions entirely in the course of ages." and claims that the first to throw doubt on this was James Dwight Dana in 1849.
In his "Manual of Geology" (1863), Dana wrote, "The continents and oceans had their general outline or form defined in earliest time. This has been proved with regard to North America from the position and distribution of the first beds of the Lower Silurian, – those of the Potsdam epoch. The facts indicate that the continent of North America had its surface near tide-level, part above and part below it (p.196); and this will probably be proved to be the condition in Primordial time of the other continents also. And, if the outlines of the continents were marked out, it follows that the outlines of the oceans were no less so". Dana was enormously influential in America – his "Manual of Mineralogy" is still in print in revised form – and the theory became known as "Permanence theory".
This appeared to be confirmed by the exploration of the deep sea beds conducted by the Challenger expedition, 1872–1876, which showed that contrary to expectation, land debris brought down by rivers to the ocean is deposited comparatively close to the shore on what is now known as the continental shelf. This suggested that the oceans were a permanent feature of the Earth's surface, and did not change places with the continents.
Apart from the earlier speculations mentioned in the previous section, the idea that the American continents had once formed a single landmass together with Europe and Asia before assuming their present shapes and positions was postulated by several scientists before Alfred Wegener's 1912 paper.
Although Wegener's theory was formed independently and was more complete than those of his predecessors, Wegener later credited a number of past authors with similar ideas:
Franklin Coxworthy (between 1848 and 1890), Roberto Mantovani (between 1889 and 1909), William Henry Pickering (1907) and Frank Bursley Taylor (1908). In addition, Eduard Suess had proposed a supercontinent Gondwana in 1885 and the Tethys Ocean in 1893, assuming a land-bridge between the present continents submerged in the form of a geosyncline, and John Perry had written an 1895 paper proposing that the earth's interior was fluid, and disagreeing with Lord Kelvin on the age of the earth.
For example: the similarity of southern continent geological formations had led Roberto Mantovani to conjecture in 1889 and 1909 that all the continents had once been joined into a supercontinent; Wegener noted the similarity of Mantovani's and his own maps of the former positions of the southern continents. In Mantovani's conjecture, this continent broke due to volcanic activity caused by thermal expansion, and the new continents drifted away from each other because of further expansion of the rip-zones, where the oceans now lie. This led Mantovani to propose an Expanding Earth theory which has since been shown to be incorrect.
Continental drift without expansion was proposed by Frank Bursley Taylor, who suggested in 1908 (published in 1910) that the continents were moved into their present positions by a process of "continental creep". In a later paper he proposed that this occurred by their being dragged towards the equator by tidal forces during the hypothesized capture of the moon in the Cretaceous, resulting in "general crustal creep" toward the equator. Although his proposed mechanism was wrong, he was the first to realize the insight that one of the effects of continental motion would be the formation of mountains, and attributed the formation of the Himalayas to the collision between the Indian subcontinent with Asia. Wegener said that of all those theories, Taylor's, although not fully developed, had the most similarities to his own. In the mid-20th century, the theory of continental drift was referred to as the "Taylor-Wegener hypothesis", although this terminology eventually fell out of common use.
Alfred Wegener first presented his hypothesis to the German Geological Society on 6 January 1912. His hypothesis was that the continents had once formed a single landmass, called Pangaea, before breaking apart and drifting to their present locations.
Wegener was the first to use the phrase "continental drift" (1912, 1915) (in German "die Verschiebung der Kontinente" – translated into English in 1922) and formally publish the hypothesis that the continents had somehow "drifted" apart. Although he presented much evidence for continental drift, he was unable to provide a convincing explanation for the physical processes which might have caused this drift. His suggestion that the continents had been pulled apart by the centrifugal pseudoforce ("Polflucht") of the Earth's rotation or by a small component of astronomical precession was rejected, as calculations showed that the force was not sufficient. The Polflucht hypothesis was also studied by Paul Sophus Epstein in 1920 and found to be implausible.
The theory of continental drift was not accepted for many years. One problem was that a plausible driving force was missing. A second problem was that Wegener's estimate of the speed of continental motion, 250 cm/year, was implausibly high. (The currently accepted rate for the separation of the Americas from Europe and Africa is about 2.5 cm/year). It also did not help that Wegener was not a geologist. Other geologists also believed that the evidence that Wegener had provided was not sufficient. It is now accepted that the plates carrying the continents do move across the Earth's surface, although not as fast as Wegener believed; ironically one of the chief outstanding questions is the one Wegener failed to resolve: what is the nature of the forces propelling the plates?
The British geologist Arthur Holmes championed the theory of continental drift at a time when it was deeply unfashionable. He proposed in 1931 that the Earth's mantle contained convection cells which dissipated heat produced by radioactive decay and moved the crust at the surface. His "Principles of Physical Geology", ending with a chapter on continental drift, was published in 1944.
Geological maps of the time showed huge land bridges spanning the Atlantic and Indian oceans to account for the similarities of fauna and flora and the divisions of the Asian continent in the Permian period but failing to account for glaciation in India, Australia and South Africa.
Geophysicist Jack Oliver is credited with providing seismologic evidence supporting plate tectonics which encompassed and superseded continental drift with the article "Seismology and the New Global Tectonics", published in 1968, using data collected from seismologic stations, including those he set up in the South Pacific.
It is now known that there are two kinds of crust: continental crust and oceanic crust. Continental crust is inherently lighter and its composition is different from oceanic crust, but both kinds reside above a much deeper "plastic" mantle. Oceanic crust is created at spreading centers, and this, along with subduction, drives the system of plates in a chaotic manner, resulting in continuous orogeny and areas of isostatic imbalance. The theory of plate tectonics explains all this, including the movement of the continents, better than Wegener's theory.
Hans Stille and Leopold Kober opposed the idea of continental drift and worked on a "fixist" geosyncline model with Earth contraction playing a key role in the formation of orogens. Other geologists who opposed continental drift were Bailey Willis, Charles Schuchert, Rollin Chamberlin and Walther Bucher. In 1939 an international geological conference was held in Frankfurt. This conference came to be dominated by the fixists, especially as those geologists specializing in tectonics were all fixists except Willem van der Gracht. Criticism of continental drift and mobilism was abundant at the conference not only from tectonicists but also from sedimentological (Nölke), paleontological (Nölke), mechanical (Lehmann) and oceanographic (Troll, Wüst) perspectives. Hans Cloos, the organizer of the conference, was also a fixist who together with Troll held the view that excepting the Pacific Ocean continents were not radically different from oceans in their behaviour. The mobilist theory of Émile Argand for the Alpine orogeny was criticized by Kurt Leuchs. The few drifters and mobilists at the conference appealed to biogeography (Kirsch, Wittmann), paleoclimatology (Wegener, K), paleontology (Gerth) and geodetic measurements (Wegener, K). F. Bernauer correctly equated Reykjanes in south-west Iceland with the Mid-Atlantic Ridge, arguing with this that the floor of the Atlantic Ocean was undergoing extension just like Reykjanes. Bernauer thought this extension had drifted the continents only 100–200 km apart, the approximate width of the volcanic zone in Iceland.
David Attenborough, who attended university in the second half of the 1940s, recounted an incident illustrating its lack of acceptance then: "I once asked one of my lecturers why he was not talking to us about continental drift and I was told, sneeringly, that if I could prove there was a force that could move continents, then he might think about it. The idea was moonshine, I was informed."
As late as 1953 – just five years before Carey introduced the theory of plate tectonics – the theory of continental drift was rejected by the physicist Scheidegger on the following grounds.
From the 1930s to the late 1950s, works by Vening-Meinesz, Holmes, Umbgrove, and numerous others outlined concepts that were close or nearly identical to modern plate tectonics theory. In particular, the English geologist Arthur Holmes proposed in 1920 that plate junctions might lie beneath the sea, and in 1928 that convection currents within the mantle might be the driving force. Holmes' views were particularly influential: in his bestselling textbook, "Principles of Physical Geology," he included a chapter on continental drift, proposing that Earth's mantle contained convection cells which dissipated radioactive heat and moved the crust at the surface. Holmes' proposal resolved the phase disequilibrium objection (the underlying fluid was kept from solidifying by radioactive heating from the core). However, scientific communication in the '30 and '40s was inhibited by the war, and the theory still required work to avoid foundering on the orogeny and isostasy objections. Worse, the most viable forms of the theory predicted the existence of convection cell boundaries reaching deep into the earth that had yet to be observed.
In 1947, a team of scientists led by Maurice Ewing confirmed the existence of a rise in the central Atlantic Ocean, and found that the floor of the seabed beneath the sediments was chemically and physically different from continental crust. As oceanographers continued to bathymeter the ocean basins, a system of mid-oceanic ridges was detected. An important conclusion was that along this system, new ocean floor was being created, which led to the concept of the "Great Global Rift".
Meanwhile, scientists began recognizing odd magnetic variations across the ocean floor using devices developed during World War II to detect submarines. Over the next decade, it became increasingly clear that the magnetization patterns were not anomalies, as had been originally supposed. In a series of papers in 1959–1963, Heezen, Dietz, Hess, Mason, Vine, Matthews, and Morley collectively realized that the magnetization of the ocean floor formed extensive, zebra-like patterns: one stripe would exhibit normal polarity and the adjoining stripes reversed polarity. The best explanation was the "conveyor belt" or Vine–Matthews–Morley hypothesis. New magma from deep within the Earth rises easily through these weak zones and eventually erupts along the crest of the ridges to create new oceanic crust. The new crust is magnetized by the earth's magnetic field, which undergoes occasional reversals. Formation of new crust then displaces the magnetized crust apart, akin to a conveyor belt – hence the name.
Without workable alternatives to explain the stripes, geophysicists were forced to conclude that Holmes had been right: ocean rifts were sites of perpetual orogeny at the boundaries of convection cells. By 1967, barely two decades after discovery of the mid-oceanic rifts, and a decade after discovery of the striping, plate tectonics had become axiomatic to modern geophysics.
In addition, Marie Tharp, in collaboration with Bruce Heezen, who initially ridiculed Tharp's observations that her maps confirmed continental drift theory, provided essential corroboration, using her skills in cartography and seismographic data, to confirm the theory.
Evidence for the movement of continents on tectonic plates is now extensive. Similar plant and animal fossils are found around the shores of different continents, suggesting that they were once joined. The fossils of "Mesosaurus", a freshwater reptile rather like a small crocodile, found both in Brazil and South Africa, are one example; another is the discovery of fossils of the land reptile "Lystrosaurus" in rocks of the same age at locations in Africa, India, and Antarctica. There is also living evidence, with the same animals being found on two continents. Some earthworm families (such as Ocnerodrilidae, Acanthodrilidae, Octochaetidae) are found in South America and Africa.
The complementary arrangement of the facing sides of South America and Africa is obvious but a temporary coincidence. In millions of years, slab pull, ridge-push, and other forces of tectonophysics will further separate and rotate those two continents. It was that temporary feature that inspired Wegener to study what he defined as continental drift although he did not live to see his hypothesis generally accepted.
The widespread distribution of Permo-Carboniferous glacial sediments in South America, Africa, Madagascar, Arabia, India, Antarctica and Australia was one of the major pieces of evidence for the theory of continental drift. The continuity of glaciers, inferred from oriented glacial striations and deposits called tillites, suggested the existence of the supercontinent of Gondwana, which became a central element of the concept of continental drift. Striations indicated glacial flow away from the equator and toward the poles, based on continents' current positions and orientations, and supported the idea that the southern continents had previously been in dramatically different locations that were contiguous with one another. | https://en.wikipedia.org/wiki?curid=6056 |
Commodores
Commodores is an American funk and soul band, which was at its peak in the late 1970s through the mid 1980s. The members of the group met as mostly freshmen at Tuskegee Institute (now Tuskegee University) in 1968, and signed with Motown in November 1972, having first caught the public eye opening for the Jackson 5 while on tour.
The group's most successful period was in the late 1970s and early 1980s when Lionel Richie was the co-lead singer. The band's biggest hit singles are ballads such as "Easy", "Three Times a Lady", and "Nightshift"; and funky dance hits which include "Brick House", "Fancy Dancer", "Lady (You Bring Me Up)", and "Too Hot ta Trot".
The Commodores were inducted into the Alabama Music Hall of Fame and Vocal Group Hall of Fame. The band has also won one Grammy Award out of nine nominations. As well the Commodores have sold over 70 million albums worldwide.
The Commodores formed from two former student groups, the Mystics and the Jays. Richie described some members of the Mystics as "jazz buffs."
The new six-man band featured Lionel Richie, Thomas McClary, and William King from the Mystics, and Andre Callahan, Michael Gilbert, and Milan Williams from the Jays. To choose their name, William King opened a dictionary and randomly picked a word. "We lucked out," he remarked with a laugh when telling this story to "People" magazine. "We almost became 'The Commodes.'"
The bandmembers attended Tuskegee University in Alabama. After winning the university's annual freshman talent contest, they played at fraternity parties as well as a weekend gig at the Black Forest Inn, one of a few clubs in Tuskegee that catered to college students. They performed cover tunes and some original songs with their first singer, James Ingram (not the famous solo artist). Ingram, older than the rest of the band, left to serve in Vietnam, and was later replaced by Walter "Clyde" Orange, who wrote or co-wrote many of their hits. Lionel Richie and Orange alternated as lead singers. (Orange was the lead singer on the Top 10 hits "Brick House" and "Nightshift".)
The early band was managed by Benny Ashburn, who brought them to his family's vacation lodge on Martha's Vineyard in 1971 and 1972. There, Ashburn test-marketed the group by having them play in parking lots and summer festivals.
The Commodores made a brief appearance in the 1978 film "Thank God It's Friday". They performed the song "Too Hot ta Trot" during the dance contest; the songs "Brick House" and "Easy" were also played in the movie.
"Machine Gun," the instrumental title track from the band's debut album, became a staple at American sporting events, and is also heard in many films, including "Boogie Nights" and "Looking for Mr. Goodbar". It reached No. 22 on the "Billboard" Hot 100 in 1975. Another instrumental, "Cebu" (named after an island in the Philippines), later became a staple in the Quiet Storm radio format. Three albums released in 1975 and 1976 ("Caught in the Act", "Movin' On", and "Hot on the Tracks") are considered the peak of their harder funk period. After those recordings the group developed the mellower sound hinted at in their 1976 Top Ten hits "Sweet Love" and "Just to Be Close to You." In 1977 the Commodores released "Easy," which became the group's biggest hit yet, reaching No. 4 in the US, followed by "Brick House," also top 5, both from their album "The Commodores", as was "Zoom." The group reached No. 1 in 1978 with "Three Times a Lady." In 1979 the Commodores scored another top-five ballad, "Sail On," before reaching the top of the charts once again with another ballad, "Still." In 1981 they released two top-ten hits with "Oh No" (No. 4) and their first upbeat single in almost five years, "Lady (You Bring Me Up)" (No. 8).
In 1982, Lionel Richie left to pursue a solo career, and Skyler Jett replaced him as co-lead singer. Also in 1982, their manager Benjamin Ashburn died of a heart attack at the age of 54.
Founding member McClary left in 1983 (shortly after Richie) to pursue a solo career and to develop a gospel music company. McClary was replaced by guitarist-vocalist Sheldon Reynolds. Then LaPread left in 1986 and moved to Auckland, New Zealand. Reynolds departed for Earth, Wind & Fire in 1987, which prompted trumpeter William "WAK" King to take over primary guitar duties for live performances. Keyboardist Milan Williams exited the band in 1989 after allegedly refusing to tour South Africa.
The group gradually abandoned its funk roots and moved into the more commercial pop arena. In 1984 former Heatwave singer James Dean "J.D." Nicholas assumed co-lead vocal duties with drummer Walter Orange. That lineup was hitless until 1985 when their final Motown album "Nightshift", produced by Dennis Lambert (prior albums were produced by James Anthony Carmichael), delivered the title track "Nightshift," a loving tribute to Marvin Gaye and Jackie Wilson, both of whom had died the previous year. "Nightshift" hit no. 3 in the US and won The Commodores their first Grammy for Best R&B Performance by a Duo or Group With Vocals in 1985.
In 2010 a new version was recorded, dedicated to Michael Jackson. The Commodores were on a European tour performing at Wembley Arena, London, on June 25, 2009, when they walked off the stage after they were told that Michael Jackson had died. Initially the band thought it was a hoax. However, back in their dressing rooms they received confirmation and broke down in tears. The next night at Birmingham's NIA Arena, J.D. Nicholas added Jackson's name to the lyrics of the song, and henceforth The Commodores have mentioned Jackson and other deceased R&B singers. Thus came the inspiration upon the one-year anniversary of Jackson's death to re-record, with new lyrics, the hit song "Nightshift" as a tribute.
In 1990 the Commodores formed Commodores Records and re-recorded their 20 greatest hits as "Commodores Hits Vol. I & II". They have recorded a live album, "Commodores Live", along with a DVD of the same name, and a Christmas album titled "Commodores Christmas". In 2012, the band was working on new material, with some contributions written by current and former members.
The Commodores, as of 2020, consist of Walter "Clyde" Orange, James Dean "J.D." Nicholas, and William "WAK" King, along with their five-piece band The Mean Machine. They continue to perform, playing at arenas, theaters, and festivals around the world.
The Commodores have won one Grammy Award out of nine nominations.
During 1995 the Commodores were inducted into the Alabama Music Hall of Fame.
During 2003 the Commodores were also inducted into the Vocal Group Hall of Fame. | https://en.wikipedia.org/wiki?curid=6057 |
Collagen
Collagen is the main structural protein in the extracellular matrix in the various connective tissues in the body. As the main component of connective tissue, it is the most abundant protein in mammals, making up from 25% to 35% of the whole-body protein content. Collagen consists of amino acids bound together to form a triple helix of elongated fibril known as a collagen helix. It is mostly found in fibrous tissues such as tendons, ligaments, and skin.
Depending upon the degree of mineralization, collagen tissues may be rigid (bone), compliant (tendon), or have a gradient from rigid to compliant (cartilage). It is also abundant in corneas, blood vessels, the gut, intervertebral discs, and the dentin in teeth. In muscle tissue, it serves as a major component of the endomysium. Collagen constitutes one to two percent of muscle tissue and accounts for 6% of the weight of strong, tendinous, muscles. The fibroblast is the most common cell that creates collagen. Gelatin, which is used in food and industry, is collagen that has been irreversibly hydrolyzed. Collagen has many medical uses in treating complications of the bones and skin.
The name "collagen" comes from the Greek κόλλα ("kólla"), meaning "glue", and suffix -γέν, "-gen", denoting "producing". This refers to the compound's early use in the process of boiling the skin and tendons of horses and other animals to obtain glue.
Over 90% of the collagen in the human body is type I collagen. However, as of 2011, 30 types of collagen have been identified, described, and divided into several groups according to the structure they form: All of the types contain at least one triple helix. The number of types shows collagen's diverse functionality.
The five most common types are:
The collagenous cardiac skeleton which includes the four heart valve rings, is histologically, elastically and uniquely bound to cardiac muscle. The cardiac skeleton also includes the separating septa of the heart chambers – the interventricular septum and the atrioventricular septum. Collagen contribution to the measure of cardiac performance summarily represents a continuous torsional force opposed to the fluid mechanics of blood pressure emitted from the heart. The collagenous structure that divides the upper chambers of the heart from the lower chambers is an impermeable membrane that excludes both blood and electrical impulses through typical physiological means. With support from collagen, atrial fibrillation never deteriorates to ventricular fibrillation. Collagen is layered in variable densities with smooth muscle mass. The mass, distribution, age and density of collagen all contribute to the compliance required to move blood back and forth. Individual cardiac valvular leaflets are folded into shape by specialized collagen under variable pressure. Gradual calcium deposition within collagen occurs as a natural function of aging. Calcified points within collagen matrices show contrast in a moving display of blood and muscle, enabling methods of cardiac imaging technology to arrive at ratios essentially stating blood in (cardiac input) and blood out (cardiac output). Pathology of the collagen underpinning of the heart is understood within the category of connective tissue disease.
Collagen has been widely used in cosmetic surgery, as a healing aid for burn patients for reconstruction of bone and a wide variety of dental, orthopedic, and surgical purposes. Both human and bovine collagen is widely used as dermal fillers for treatment of wrinkles and skin aging. Some points of interest are:
As the skeleton forms the structure of the body, it is vital that it maintains its strength, even after breaks and injuries. Collagen is used in bone grafting as it has a triple helical structure, making it a very strong molecule. It is ideal for use in bones, as it does not compromise the structural integrity of the skeleton. The triple helical structure of collagen prevents it from being broken down by enzymes, it enables adhesiveness of cells and it is important for the proper assembly of the extracellular matrix.
Collagen scaffolds are used in tissue regeneration, whether in sponges, thin sheets, or gels. Collagen has the correct properties for tissue regeneration such as pore structure, permeability, hydrophilicity, and being stable in vivo. Collagen scaffolds are also ideal for the deposition of cells such as osteoblasts and fibroblasts, and once inserted, growth is able to continue as normal in the tissue.
Collagens are widely employed in the construction of the artificial skin substitutes used in the management of severe burns and wounds. These collagens may be derived from bovine, equine, porcine, or even human sources; and are sometimes used in combination with silicones, glycosaminoglycans, fibroblasts, growth factors and other substances.
Collagen is one of the body's key natural resources and a component of skin tissue that can benefit all stages of wound healing. When collagen is made available to the wound bed, closure can occur. Wound deterioration, followed sometimes by procedures such as amputation, can thus be avoided.
Collagen is a natural product and is thus used as a natural wound dressing and has properties that artificial wound dressings do not have. It is resistant against bacteria, which is of vital importance in a wound dressing. It helps to keep the wound sterile, because of its natural ability to fight infection. When collagen is used as a burn dressing, healthy granulation tissue is able to form very quickly over the burn, helping it to heal rapidly.
Throughout the 4 phases of wound healing, collagen performs the following functions in wound healing:
Collagen is used in laboratory studies for cell culture, studying cell behavior and cellular interactions with the extracellular environment.
The collagen protein is composed of a triple helix, which generally consists of two identical chains (α1) and an additional chain that differs slightly in its chemical composition (α2). The amino acid composition of collagen is atypical for proteins, particularly with respect to its high hydroxyproline content. The most common motifs in the amino acid sequence of collagen are glycine-proline-X and glycine-X-hydroxyproline, where X is any amino acid other than glycine, proline or hydroxyproline. The average amino acid composition for fish and mammal skin is given.
First, a three-dimensional stranded structure is assembled, with the amino acids glycine and proline as its principal components. This is not yet collagen but its precursor, procollagen. Procollagen is then modified by the addition of hydroxyl groups to the amino acids proline and lysine. This step is important for later glycosylation and the formation of the triple helix structure of collagen. Because the hydroxylase enzymes that perform these reactions require vitamin C as a cofactor, a long-term deficiency in this vitamin results in impaired collagen synthesis and scurvy. These hydroxylation reactions are catalyzed by two different enzymes: prolyl-4-hydroxylase and lysyl-hydroxylase. Vitamin C also serves with them in inducing these reactions. In this service, one molecule of vitamin C is destroyed for each H replaced by OH.
The synthesis of collagen occurs inside and outside of the cell. The formation of collagen which results in fibrillary collagen (most common form) is discussed here. Meshwork collagen, which is often involved in the formation of filtration systems, is the other form of collagen. All types of collagens are triple helices, and the differences lie in the make-up of the alpha peptides created in step 2.
Collagen has an unusual amino acid composition and sequence:
Cortisol stimulates degradation of (skin) collagen into amino acids.
Most collagen forms in a similar manner, but the following process is typical for type I:
Vitamin C deficiency causes scurvy, a serious and painful disease in which defective collagen prevents the formation of strong connective tissue. Gums deteriorate and bleed, with loss of teeth; skin discolors, and wounds do not heal. Prior to the 18th century, this condition was notorious among long-duration military, particularly naval, expeditions during which participants were deprived of foods containing vitamin C.
An autoimmune disease such as lupus erythematosus or rheumatoid arthritis may attack healthy collagen fibers.
Many bacteria and viruses secrete virulence factors, such as the enzyme collagenase, which destroys collagen or interferes with its production.
A single collagen molecule, tropocollagen, is used to make up larger collagen aggregates, such as fibrils. It is approximately 300 nm long and 1.5 nm in diameter, and it is made up of three polypeptide strands (called alpha peptides, see step 2), each of which has the conformation of a left-handed helix – this should not be confused with the right-handed alpha helix. These three left-handed helices are twisted together into a right-handed triple helix or "super helix", a cooperative quaternary structure stabilized by many hydrogen bonds. With type I collagen and possibly all fibrillar collagens, if not all collagens, each triple-helix associates into a right-handed super-super-coil referred to as the collagen microfibril. Each microfibril is interdigitated with its neighboring microfibrils to a degree that might suggest they are individually unstable, although within collagen fibrils, they are so well ordered as to be crystalline.
A distinctive feature of collagen is the regular arrangement of amino acids in each of the three chains of these collagen subunits. The sequence often follows the pattern Gly-Pro-X or Gly-X-Hyp, where X may be any of various other amino acid residues. Proline or hydroxyproline constitute about 1/6 of the total sequence. With glycine accounting for the 1/3 of the sequence, this means approximately half of the collagen sequence is not glycine, proline or hydroxyproline, a fact often missed due to the distraction of the unusual GX1X2 character of collagen alpha-peptides. The high glycine content of collagen is important with respect to stabilization of the collagen helix as this allows the very close association of the collagen fibers within the molecule, facilitating hydrogen bonding and the formation of intermolecular cross-links. This kind of regular repetition and high glycine content is found in only a few other fibrous proteins, such as silk fibroin.
Collagen is not only a structural protein. Due to its key role in the determination of cell phenotype, cell adhesion, tissue regulation, and infrastructure, many sections of its non-proline-rich regions have cell or matrix association/regulation roles. The relatively high content of proline and hydroxyproline rings, with their geometrically constrained carboxyl and (secondary) amino groups, along with the rich abundance of glycine, accounts for the tendency of the individual polypeptide strands to form left-handed helices spontaneously, without any intrachain hydrogen bonding.
Because glycine is the smallest amino acid with no side chain, it plays a unique role in fibrous structural proteins. In collagen, Gly is required at every third position because the assembly of the triple helix puts this residue at the interior (axis) of the helix, where there is no space for a larger side group than glycine's single hydrogen atom. For the same reason, the rings of the Pro and Hyp must point outward. These two amino acids help stabilize the triple helix—Hyp even more so than Pro; a lower concentration of them is required in animals such as fish, whose body temperatures are lower than most warm-blooded animals. Lower proline and hydroxyproline contents are characteristic of cold-water, but not warm-water fish; the latter tend to have similar proline and hydroxyproline contents to mammals. The lower proline and hydroxproline contents of cold-water fish and other poikilotherm animals leads to their collagen having a lower thermal stability than mammalian collagen. This lower thermal stability means that gelatin derived from fish collagen is not suitable for many food and industrial applications.
The tropocollagen subunits spontaneously self-assemble, with regularly staggered ends, into even larger arrays in the extracellular spaces of tissues. Additional assembly of fibrils is guided by fibroblasts, which deposit fully formed fibrils from fibripositors. In the fibrillar collagens, molecules are staggered to adjacent molecules by about 67 nm (a unit that is referred to as ‘D’ and changes depending upon the hydration state of the aggregate). In each D-period repeat of the microfibril, there is a part containing five molecules in cross-section, called the “overlap”, and a part containing only four molecules, called the "gap". These overlap and gap regions are retained as microfibrils assemble into fibrils, and are thus viewable using electron microscopy. The triple helical tropocollagens in the microfibrils are arranged in a quasihexagonal packing pattern.
There is some covalent crosslinking within the triple helices, and a variable amount of covalent crosslinking between tropocollagen helices forming well organized aggregates (such as fibrils). Larger fibrillar bundles are formed with the aid of several different classes of proteins (including different collagen types), glycoproteins, and proteoglycans to form the different types of mature tissues from alternate combinations of the same key players. Collagen's insolubility was a barrier to the study of monomeric collagen until it was found that tropocollagen from young animals can be extracted because it is not yet fully crosslinked. However, advances in microscopy techniques (i.e. electron microscopy (EM) and atomic force microscopy (AFM)) and X-ray diffraction have enabled researchers to obtain increasingly detailed images of collagen structure "in situ". These later advances are particularly important to better understanding the way in which collagen structure affects cell–cell and cell–matrix communication and how tissues are constructed in growth and repair and changed in development and disease. For example, using AFM–based nanoindentation it has been shown that a single collagen fibril is a heterogeneous material along its axial direction with significantly different mechanical properties in its gap and overlap regions, correlating with its different molecular organizations in these two regions.
Collagen fibrils/aggregates are arranged in different combinations and concentrations in various tissues to provide varying tissue properties. In bone, entire collagen triple helices lie in a parallel, staggered array. 40 nm gaps between the ends of the tropocollagen subunits (approximately equal to the gap region) probably serve as nucleation sites for the deposition of long, hard, fine crystals of the mineral component, which is hydroxylapatite (approximately) Ca10(OH)2(PO4)6. Type I collagen gives bone its tensile strength.
Collagen-related diseases most commonly arise from genetic defects or nutritional deficiencies that affect the biosynthesis, assembly, postranslational modification, secretion, or other processes involved in normal collagen production.
In addition to the above-mentioned disorders, excessive deposition of collagen occurs in scleroderma.
One thousand mutations have been identified in 12 out of more than 20 types of collagen. These mutations can lead to various diseases at the tissue level.
Osteogenesis imperfecta – Caused by a mutation in type 1 collagen, dominant autosomal disorder, results in weak bones and irregular connective tissue, some cases can be mild while others can be lethal. Mild cases have lowered levels of collagen type 1 while severe cases have structural defects in collagen.
Chondrodysplasias – Skeletal disorder believed to be caused by a mutation in type 2 collagen, further research is being conducted to confirm this.
Ehlers-Danlos syndrome – Thirteen different types of this disorder, which lead to deformities in connective tissue, are known. Some of the rarer types can be lethal, leading to the rupture of arteries. Each syndrome is caused by a different mutation. For example, the vascular type (vEDS) of this disorder is caused by a mutation in collagen type 3.
Alport syndrome – Can be passed on genetically, usually as X-linked dominant, but also as both an autosomal dominant and autosomal recessive disorder, sufferers have problems with their kidneys and eyes, loss of hearing can also develop during the childhood or adolescent years.
Knobloch syndrome – Caused by a mutation in the COL18A1 gene that codes for the production of collagen XVIII. Patients present with protrusion of the brain tissue and degeneration of the retina; an individual who has family members suffering from the disorder is at an increased risk of developing it themselves since there is a hereditary link.
Collagen is one of the long, fibrous structural proteins whose functions are quite different from those of globular proteins, such as enzymes. Tough bundles of collagen called "collagen fibers" are a major component of the extracellular matrix that supports most tissues and gives cells structure from the outside, but collagen is also found inside certain cells. Collagen has great tensile strength, and is the main component of fascia, cartilage, ligaments, tendons, bone and skin. Along with elastin and soft keratin, it is responsible for skin strength and elasticity, and its degradation leads to wrinkles that accompany aging. It strengthens blood vessels and plays a role in tissue development. It is present in the cornea and lens of the eye in crystalline form. It may be one of the most abundant proteins in the fossil record, given that it appears to fossilize frequently, even in bones from the Mesozoic and Paleozoic.
Collagen has a wide variety of applications, from food to medical. For instance, it is used in cosmetic surgery and burn surgery. It is widely used in the form of collagen casings for sausages.
If collagen is subject to sufficient denaturation, e.g. by heating, the three tropocollagen strands separate partially or completely into globular domains, containing a different secondary structure to the normal collagen polyproline II (PPII), e.g. random coils. This process describes the formation of gelatin, which is used in many foods, including flavored gelatin desserts. Besides food, gelatin has been used in pharmaceutical, cosmetic, and photography industries. It is also used as a dietary supplement.
From the Greek for glue, "kolla", the word collagen means "glue producer" and refers to the early process of boiling the skin and sinews of horses and other animals to obtain glue. Collagen adhesive was used by Egyptians about 4,000 years ago, and Native Americans used it in bows about 1,500 years ago. The oldest glue in the world, carbon-dated as more than 8,000 years old, was found to be collagen—used as a protective lining on rope baskets and embroidered fabrics, and to hold utensils together; also in crisscross decorations on human skulls. Collagen normally converts to gelatin, but survived due to dry conditions. Animal glues are thermoplastic, softening again upon reheating, so they are still used in making musical instruments such as fine violins and guitars, which may have to be reopened for repairs—an application incompatible with tough, synthetic plastic adhesives, which are permanent. Animal sinews and skins, including leather, have been used to make useful articles for millennia.
Gelatin-resorcinol-formaldehyde glue (and with formaldehyde replaced by less-toxic pentanedial and ethanedial) has been used to repair experimental incisions in rabbit lungs.
The evolution of collagens was a fundamental step in the early evolution of animals, supporting multicellular animal forms. Collagens are the most abundant proteins in vertebrates, making up some 30% of all proteins in the human body. Based on their molecular structures, collagen proteins are divided into two main classes – fibril-forming (or "fibrillar") collagens and non-fibril-forming ("non-fibrillar") collagens – which are further divided into 28 different types (as of 2017), based on individual structures and functions that the protein specifically has in the body. Fibrillar collagen, producing the three-dimensional frameworks in different tissues and organs, derived from a single common ancestor during evolution. Non-fibrillar collagen is the major supporting component of the extracellular matrix.
The morphology of fibrillar and non-fibrillar collagen types became differentiated during divergent evolution. These two collagen types arose from different mutation and gene duplication events that evolved to the current 28 types of collagen proteins providing the diversification of collagen-supporting structures in the body, such as the skeleton, which formed from the alpha (denoted as α) collagen gene. Fibrillar collagen was co-opted during evolutionary adaptation from existing genes by natural selection to construct new organ and tissue structures, enabling the emergence of evolved species with improved capabilities.
The molecular and packing structures of collagen have eluded scientists over decades of research. The first evidence that it possesses a regular structure at the molecular level was presented in the mid-1930s. Since that time, research concentrated on the conformation of the collagen monomer, producing several competing models, although correctly dealing with the conformation of each individual peptide chain. The triple-helical "Madras" model provided an accurate model of quaternary structure in collagen. This model was supported by further studies showing higher resolution in the late 20th century.
The packing structure of collagen has not been defined to the same degree outside of the fibrillar collagen types, although it has been long known to be hexagonal. As with its monomeric structure, several conflicting models alleged that either the packing arrangement of collagen molecules is 'sheet-like' or microfibrillar. The microfibrillar structure of collagen fibrils in tendon, cornea and cartilage was imaged directly by electron microscopy in the late 20th century and early 21st century. The microfibrillar structure of tail tendon was modeled as being closest to the observed structure, although it oversimplified the topological progression of neighboring collagen molecules, and so did not predict the correct conformation of the discontinuous D-periodic pentameric arrangement termed "microfibril". | https://en.wikipedia.org/wiki?curid=6058 |
Calvin and Hobbes
Calvin and Hobbes is a daily American comic strip created by cartoonist Bill Watterson that was syndicated from November 18, 1985 to December 31, 1995. Commonly cited as "the last great newspaper comic", "Calvin and Hobbes" has enjoyed broad and enduring popularity, influence, and academic and philosophical interest.
"Calvin and Hobbes" follows the humorous antics of the title characters: Calvin, a precocious, mischievous and adventurous six-year-old boy; and Hobbes, his sardonic stuffed tiger. Set in the contemporary suburban United States, the strip depicts Calvin's frequent flights of fancy and friendship with Hobbes. It also examines Calvin's relationships with family and classmates, especially the love/hate relationship between him and his classmate Susie Derkins. Hobbes' dual nature is a defining motif for the strip: to Calvin, Hobbes is a living anthropomorphic tiger, while all the other characters see Hobbes as an inanimate stuffed toy. Though the series does not frequently mention specific political figures or contemporary events, it does explore broad issues like environmentalism, public education, philosophical quandaries and the flaws of opinion polls.
At the height of its popularity, "Calvin and Hobbes" was featured in over 2,400 newspapers worldwide. In 2010, reruns of the strip appeared in more than 50 countries, and nearly 45 million copies of the "Calvin and Hobbes" books had been sold worldwide.
"Calvin and Hobbes" was conceived when Bill Watterson, while working in an advertising job he detested, began devoting his spare time to developing a newspaper comic for potential syndication. He explored various strip ideas but all were rejected by the syndicates. United Feature Syndicate finally responded positively to one strip called "The Doghouse", which featured a side character (the main character's little brother) who had a stuffed tiger. United identified these characters as the strongest, and encouraged Watterson to develop them as the centre of their own strip. Though United Feature ultimately rejected the new strip as lacking in marketing potential, Universal Press Syndicate took it up.
The first strip was published on November 18, 1985 in 35 newspapers. Watterson was warned by the syndicate not to give up his day job yet, but it was not long before the series had become a hit. Within a year of syndication, the strip was published in roughly 250 newspapers and was proving to have international appeal with translation and wide circulation outside the United States.
Although "Calvin and Hobbes" would undergo continual artistic development and creative innovation over the period of syndication, the earliest strips demonstrate a remarkable consistency with the latest. Watterson introduced all the major characters within the first three weeks, and made no changes to the central cast over the strip's 10-year history.
By April 5, 1987, Watterson was featured in an article in "The Los Angeles Times". "Calvin and Hobbes" earned Watterson the Reuben Award from the National Cartoonists Society in the Outstanding Cartoonist of the Year category, first in 1986 and again in 1988. He was nominated another time in 1992. The Society awarded him the Humor Comic Strip Award for 1988. "Calvin and Hobbes" has also won several more awards.
As his creation grew in popularity, Watterson underwent a long and emotionally draining battle with his syndicate editors over his refusal to license his characters for merchandising. By 1991, Watterson had achieved his goal of securing a new contract that granted him legal control over his creation and all future licensing arrangements.
Having achieved his objective of creative control, Watterson's desire for privacy subsequently reasserted itself and he ceased all media interviews, relocated to New Mexico, and largely disappeared from public engagements, refusing to attend the ceremonies of any of the cartooning awards he won. The pressures of the battle over merchandising led to Watterson taking an extended break from May 5, 1991, to February 1, 1992, a move that was virtually unprecedented in the world of syndicated cartoonists.
During Watterson's first sabbatical from the strip, Universal Press Syndicate continued to charge newspapers full price to re-run old "Calvin and Hobbes" strips. Few editors approved of the move, but the strip was so popular that they had no choice but to continue to run it for fear that competing newspapers might pick it up and draw its fans away. Watterson returned to the strip in 1992 with plans to produce his Sunday strip as an unbreakable half of a newspaper or tabloid page. This made him only the second cartoonist since Garry Trudeau to have sufficient popularity to demand more space and control over the presentation of his work.
Watterson took a second sabbatical from April 3 through December 31, 1994. When he returned, he had made the decision to end the strip. In 1995, Watterson sent a letter via his syndicate to all editors whose newspapers carried his strip announcing his plans to end the strip by the end of the year. Stating his belief that he had achieved everything that he wanted to within the medium, he announced his intention to work on future projects at a slower pace with fewer artistic compromises.
The final strip ran on Sunday, December 31, 1995. It depicted Calvin and Hobbes outside in freshly fallen snow, reveling in the wonder and excitement of the winter scene. "It's a magical world, Hobbes, ol' buddy... Let's go exploring!" Calvin exclaims as they zoom off over the snowy hills on their sled, leaving, according to one critic ten years later, "a hole in the comics page that no strip has been able to fill."
Syndicated comics were typically published six times a week in black and white, with a Sunday supplement version in a larger, full color format. This larger format version of the strip was constrained by mandatory layout requirements that made it possible for newspaper editors to format the strip for different page sizes and layouts.
Watterson grew increasingly frustrated by the shrinking of the available space for comics in the newspapers and the mandatory panel divisions that restricted his ability to produce better artwork and more creative storytelling. He lamented that without space for anything more than simple dialogue or sparse artwork, comics as an art form were becoming dilute, bland, and unoriginal.
Watterson longed for the artistic freedom allotted to classic strips such as "Little Nemo" and "Krazy Kat", and in 1989 he gave a sample of what could be accomplished with such liberty in the opening pages of the Sunday strip compilation, "The Calvin and Hobbes Lazy Sunday Book—"an 8-page previously unpublished Calvin story fully illustrated in watercolor. The same book contained an afterword from the artist himself, reflecting on a time when comic strips were allocated a whole page of the newspaper and every comic was like a "color poster".
Within two years, Watterson was ultimately successful in negotiating a deal that provided him more space and creative freedom. Following his 1991 sabbatical, Universal Press announced that Watterson had decided to sell his Sunday strip as an unbreakable half of a newspaper or tabloid page. Many editors and even a few cartoonists including Bil Keane ("The Family Circus") and Bruce Beattie ("Snafu") criticized him for what they perceived as arrogance and an unwillingness to abide by the normal practices of the cartoon business. Others, including Bill Amend ("Foxtrot"), Johnny Hart ("BC", "Wizard of Id") and Barbara Brandon ("Where I'm Coming From") supported him. The American Association of Sunday and Feature Editors even formally requested that Universal reconsider the changes. Watterson's own comments on the matter was that "editors will have to judge for themselves whether or not Calvin and Hobbes deserves the extra space. If they don't think the strip carries its own weight, they don't have to run it." Ultimately only 15 newspapers cancelled the strip in response to the layout changes.
Bill Watterson took two sabbaticals from the daily requirements of producing the strip. The first took place from May 5, 1991 to February 1, 1992, and second from April 3 through December 31, 1994. These sabbaticals were included in the new contract Watterson managed to negotiate with Universal Features in 1990. The sabbaticals were proposed by the syndicate themselves, who, fearing Watterson's complete burnout, endeavored to get another five years of work from their star artist.
Watterson remains only the third cartoonist with sufficient popularity and stature to receive a sabbatical from their syndicate, the first being Garry Trudeau ("Doonesbury") in 1983 and Gary Larson ("The Far Side") in 1989. Typically cartoonists are expected to produce sufficient strips to cover any period they may wish to take off. Watterson's lengthy sabbaticals received some mild criticism from his fellow cartoonists including Greg Evans ("Luann"), and Charles Schulz ("Peanuts"), one of Watterson's major artistic influences, even called it a "puzzle". Some cartoonists resented the idea that Watterson worked harder than others, while others supported it. At least one newspaper editor noted that the strip was the most popular in the country, and stated he "earned it". Following his second sabbatical, Watterson made the decision that he was going to retire from the comic strip entirely.
In spite of "Calvin and Hobbes"' popularity, the strip remains notable for the almost complete lack of official product merchandising. Bill Watterson held that comic strips should stand on their own as an art form and although he did not start out completely opposed to merchandising in all forms (or even for all comic strips), he did reject an early syndication deal that involved incorporating a more marketable, licensed character into his strip. In spite of being an unproven cartoonist, and having been flown all the way to New York to discuss the proposal, Watterson reflexively resented the idea of "cartooning by committee" and turned it down.
Later, when "Calvin" was accepted by Universal Syndicate, and began to grow in popularity, Watterson found himself at odds with the syndicate, which urged him to begin merchandising the characters and touring the country to promote the first collections of comic strips. Watterson refused. To him, the integrity of the strip and its artist would be undermined by commercialization, which he saw as a major negative influence in the world of cartoon art and he came to believe that licensing his character would only violate the spirit of his work. He gave an example of this in discussing his opposition to a Hobbes plush toy: that if the essence of Hobbes' nature in the strip is that it remain unresolved whether he is a real tiger or a stuffed toy, then creating a real stuffed toy would only destroy the magic. However, having initially signed away control over merchandising in his initial contract with the syndicate, Watterson would commence a lengthy and emotionally draining battle with Universal to gain control over his work. Ultimately Universal did not approve any products against Watterson's wishes, understanding that unlike other comic strips, it would be near impossible to separate the creator from the strip if Watterson chose to walk away.
One estimate places the value of licensing revenue forgone by Watterson at $300–$400 million. Almost no legitimate "Calvin and Hobbes" merchandise exists. Exceptions produced during the strip's original run include two 16-month calendars (1988–89 and 1989–90), a t-shirt for the Smithsonian Exhibit, "Great American Comics: 100 Years of Cartoon Art" (1990) and the textbook "Teaching with" Calvin and Hobbes, which has been described as "perhaps the most difficult piece of official "Calvin and Hobbes" memorabilia to find." In 2010, Watterson did allow his characters to be included in a series of United States Postal Service stamps honoring five classic American comics. Licensed prints of "Calvin and Hobbes" were made available, and have also been included in various academic works.
The strip's immense popularity has led to the appearance of various counterfeit items such as window decals and T-shirts that often feature crude humor, binge drinking and other themes that are not found in Watterson's work. Images from one strip in which Calvin and Hobbes dance to loud music at night were commonly used for copyright violations. After threat of a lawsuit alleging infringement of copyright and trademark, some sticker makers replaced Calvin with a different boy, while other makers made no changes. Watterson wryly commented, "I clearly miscalculated how popular it would be to show Calvin urinating on a Ford logo," but later added, "long after the strip is forgotten, [they] are my ticket to immortality".
Watterson has expressed admiration for animation as an artform. In a 1989 interview in "The Comics Journal" he described the appeal of being able to do things with a moving image that can't be done by a simple drawing: the distortion, the exaggeration and the control over the length of time an event is viewed. However, although the visual possibilities of animation appealed to Watterson, the idea of finding a voice for Calvin made him uncomfortable, as did the idea of working with a team of animators. Ultimately, "Calvin and Hobbes" was never made into an animated series. Watterson later stated in "The Calvin and Hobbes Tenth Anniversary Book" that he liked the fact that his strip was a "low-tech, one-man operation," and that he took great pride in the fact that he drew every line and wrote every word on his own. Calls from major Hollywood figures interested in an adaptation of his work, including Jim Henson, George Lucas and Steven Spielberg were never returned and in a 2013 interview Watterson stated that he had "zero interest" in an animated adaptation as there was really no upside for him in doing so.
The strip borrows several elements and themes from three major influences: Walt Kelly's "Pogo", George Herriman's "Krazy Kat" and Charles M. Schulz's "Peanuts". Schulz and Kelly particularly influenced Watterson's outlook on comics during his formative years.
Notable elements of Watterson's artistic style are his characters' diverse and often exaggerated expressions (particularly those of Calvin), elaborate and bizarre backgrounds for Calvin's flights of imagination, expressions of motion and frequent visual jokes and metaphors. In the later years of the strip, with more panel space available for his use, Watterson experimented more freely with different panel layouts, art styles, stories without dialogue and greater use of white space. He would also experiment with his tools, once inking a strip with a stick from his yard in order to achieve a particular look. He also makes a point of not showing certain things explicitly: the "Noodle Incident" and the children's book "Hamster Huey and the Gooey Kablooie" are left to the reader's imagination, where Watterson was sure they would be "more outrageous" than he could portray. However, since then, "Hamster Huey and the Gooey Kablooie" has been written and published by a different author.
Watterson's technique started with minimalist pencil sketches drawn with a light pencil (though the larger Sunday strips often required more elaborate work) on a piece of Bristol board, with his brand of choice being Strathmore because he felt it held the drawings better on the page as opposed to the cheaper brands (Watterson said he would use any cheap pad of Bristol board his local supply store had, but switched to Strathmore after he found himself growing more and more displeased with the results). He would then use a small sable brush and India ink to fill in the rest of the drawing, saying that he did not want to simply trace over his penciling and thus make the inking more spontaneous. He lettered dialogue with a Rapidograph fountain pen, and he used a crowquill pen for odds and ends. Mistakes were covered with various forms of correction fluid, including the type used on typewriters. Watterson was careful in his use of color, often spending a great deal of time in choosing the right colors to employ for the weekly Sunday strip; his technique was to cut the color tabs the syndicate sent him into individual squares, lay out the colors, and then paint a watercolor approximation of the strip on tracing paper over the Bristol board and then mark the strip accordingly before sending it on. When "Calvin and Hobbes" began there were 64 colors available for the Sunday strips. For the later Sunday strips Watterson had 125 colors as well as the ability to fade the colors into each other.
Calvin, named after the 16th-century theologian John Calvin, is a six-year-old boy with spiky blond hair and a distinctive red-and-black striped shirt, black pants and sneakers. Despite his poor grades in school, Calvin demonstrates his intelligence through a sophisticated vocabulary, philosophical mind and creative/artistic talent. Watterson described Calvin as having "not much of a filter between his brain and his mouth", a "little too intelligent for his age", lacking in restraint and not yet having the experience to "know the things that you shouldn't do." The comic strip largely revolves around Calvin's inner world, and his largely antagonistic experiences with those outside of it (fellow students, authority figures and his parents).
From Calvin's point of view, Hobbes is an anthropomorphic tiger much larger than Calvin and full of independent attitudes and ideas. When the scene includes any other human, presumably they see merely a stuffed animal, usually seated at an off-kilter angle and blankly staring into space. The true nature of the character is never resolved, instead as Watterson describes, a 'grown-up' version of reality is juxtaposed against Calvin's, with the reader left to "decide which is truer". Hobbes is based off a grey tabby cat named Sprite that was owned by Watterson. Sprite inspired the length of Hobbes' body as well as his personality. Although Hobbes' humor stems from acting like a human, Watterson maintained the feline attitude of his own cat, Sprite.
Hobbes is named after the 17th-century philosopher Thomas Hobbes, who held what Watterson describes as "a dim view of human nature." He typically exhibits a greater understanding of consequences than Calvin, although rarely intervenes in Calvin's activities beyond a few oblique warnings. The friendship between the two characters provides the core dynamic of the strip.
Calvin's unnamed mother and father are typical middle-class parents who are relatively down to earth and whose sensible attitudes serve as a foil for Calvin's outlandish behavior. Calvin's father is a patent attorney (like Watterson's own father), while his mother is a stay-at-home mom. As Watterson insists, "As far as the strip is concerned, they are important only as Calvin's mom and dad."
Watterson recounts that some fans are angered by the sometimes sardonic way that Calvin's parents respond to him. In response to his angry fan base, Watterson often defends what Calvin's parents do, remarking that in the case of parenting a kid like Calvin, "I think they do a better job than I would." Calvin's father is overly concerned with "character building" activities in a number of strips, either in the things he makes Calvin do or in the austere eccentricities of his own lifestyle.
Susie Derkins, who first appears early in the strip and is the only important character with both a first and last name, lives on Calvin's street and is one of his classmates. Her last name apparently derives from the pet beagle owned by Watterson's wife's family.
Susie is studious and polite (though she can be aggressive if sufficiently provoked), and she likes to play house or host tea parties with her stuffed animals. She also plays imaginary games with Calvin in which she acts as a high-powered lawyer or politician and wants Calvin to pretend to be her househusband. Though both of them are typically loath to admit it, Calvin and Susie exhibit many common traits and inclinations. For example, the reader occasionally sees Susie with a stuffed rabbit named "Mr. Bun." Much like Calvin, Susie has a mischievous (and sometimes aggressive) streak as well, which the reader witnesses whenever she subverts Calvin's attempts to cheat on school tests by feeding him incorrect answers, or whenever she fights back after Calvin attacks her with snowballs or water balloons.
Hobbes often openly expresses romantic feelings for Susie, to Calvin's disgust. In contrast, Calvin started a club (of which he and Hobbes are the only members) that he calls G.R.O.S.S. (Get Rid Of Slimy GirlS), and while holding "meetings" in Calvin's tree house or in the "box of secrecy" in Calvin's room, they usually come up with some plot against Susie. In one instance, Calvin steals one of Susie's dolls and holds it for ransom, only to have Susie retaliate by nabbing Hobbes. Watterson admits that Calvin and Susie have a nascent crush on each other and that Susie is a reference to the type of woman whom Watterson himself found attractive and eventually married.
Calvin also interacts with a handful of secondary characters. Several of these, including Rosalyn, his babysitter; Mrs Wormwood, his teacher; and Moe, the school bully, recur regularly through the duration of the strip.
Watterson used the strip to poke fun at the art world, principally through Calvin's unconventional creations of snowmen but also through other expressions of childhood art. When Miss Wormwood complains that he is wasting class time drawing impossible things (a "Stegosaurus" in a rocket ship, for example), Calvin proclaims himself "on the cutting edge of the "avant-garde"." He begins exploring the medium of snow when a warm day melts his snowman. His next sculpture "speaks to the horror of our own mortality, inviting the viewer to contemplate the evanescence of life." In later strips, Calvin's creative instincts diversify to include sidewalk drawings (or, as he terms them, examples of "suburban postmodernism").
Watterson also lampooned the academic world. In one example, Calvin carefully crafts an "artist's statement", claiming that such essays convey more messages than artworks themselves ever do (Hobbes blandly notes, "You misspelled "Weltanschauung""). He indulges in what Watterson calls "pop psychobabble" to justify his destructive rampages and shift blame to his parents, citing "toxic codependency." In one instance, he pens a book report based on the theory that the purpose of academic writing is to "inflate weak ideas, obscure poor reasoning and inhibit clarity," entitled "The Dynamics of Interbeing and Monological Imperatives in Dick and Jane: A Study in Psychic Transrelational Gender Modes". Displaying his creation to Hobbes, he remarks, "Academia, here I come!" Watterson explains that he adapted this jargon (and similar examples from several other strips) from an actual book of art criticism.
Overall, Watterson's satirical essays serve to attack both sides, criticizing both the commercial mainstream and the artists who are supposed to be "outside" it. The strip on Sunday, June 21, 1992, criticized the naming of The Big Bang theory as not evocative of the wonders behind it, and coined the term "Horrendous Space Kablooie", an alternative that achieved some informal popularity among scientists and was often shortened to "the HSK." The term has also been referred to in newspapers, books and university courses.
There are many recurring gags in the strip, some in reality and others in Calvin's imagination. These are as follows:
Calvin imagines himself as many great creatures and other people, including dinosaurs, elephants, jungle-farers and superheroes. Three of his alter egos are well-defined and recurrent:
Calvin also has several adventures involving corrugated cardboard boxes, which he adapts for many imaginative and elaborate uses. In one strip, when Calvin shows off his Transmogrifier, a device that transforms its user into any desired creature or item, Hobbes remarks, "It's amazing what they do with corrugated cardboard these days." Calvin is able to change the function of the boxes by rewriting the label and flipping the box onto another side. In this way, a box can be used not only for its conventional purposes (a storage container for water balloons, for example), but also as a flying time machine, a duplicator, a transmogrifier or, with the attachment of a few wires and a colander, a "Cerebral Enhance-o-tron."
In the real world, Calvin's antics with his box have had varying effects. When he transmogrified into a tiger, he still appeared as a regular human child to his parents. However, in a story where he made several duplicates of himself, his parents are seen interacting with what does seem like multiple Calvins, including in a strip where two of him are seen in the same panel as his father. It is ultimately unknown what his parents do or do not see, as Calvin tries to hide most of his creations (or conceal their effects) so as not to traumatize them.
In addition, Calvin uses a cardboard box as a sidewalk kiosk to sell things. Often, Calvin offers merchandise no one would want, such as "suicide drink", "a swift kick in the butt" for one dollar or a "frank appraisal of your looks" for fifty cents. In one strip, he sells "happiness" for ten cents, hitting the customer in the face with a water balloon and explaining that he meant his own happiness. In another strip, he sold "insurance", firing a slingshot at those who refused to buy it. In some strips, he tried to sell "great ideas", and in one earlier strip, he attempted to sell the family car to obtain money for a grenade launcher. In yet another strip, he sells "life" for five cents, where the customer receives nothing in return, which, in Calvin's opinion, is life.
The box has also functioned as an alternate secret meeting place for G.R.O.S.S., as the "Box of Secrecy".
Calvinball is an improvisational sport/game introduced in a 1990 storyline that involved Calvin's negative experience of joining the school baseball team. Calvinball is a nomic or self-modifying game, a contest of wits, skill and creativity rather than stamina or athletic skill. The game is portrayed as a rebellion against conventional team sports and became a staple of the final 5 years of the comic. The only consistent rules of the game are that Calvinball may never be played with the same rules twice and that each participant must wear a mask.
When asked how to play, Watterson states: "It's pretty simple: you make up the rules as you go." In most appearances of the game, a comical array of conventional and non-conventional sporting equipment is involved, including a croquet set, a badminton set, assorted flags, bags, signs, a hobby horse, water buckets and balloons, with humorous allusions to unseen elements such as "time-fracture wickets". Scoring is portrayed as arbitrary and nonsensical ("Q to 12" and "oogy to boogy") and the lack of fixed rules leads to lengthy argument between the participants as to who scored, where the boundaries are, and when the game is finished. Usually, the contest results in Calvin being outsmarted by Hobbes. The game has been described in one academic work not as a new game based on fragments of an older one, but as the "constant connecting and disconnecting of parts, the constant evasion of rules or guidelines based on collective creativity."
Calvin often creates horrendous/dark humor scenes with his snowmen and other snow sculptures. He uses the snowman for social commentary, revenge or pure enjoyment. Examples include Snowman Calvin being yelled at by Snowman Dad to shovel the snow; one snowman eating snow cones scooped out of a second snowman, who is lying on the ground with an ice-cream scoop in his back; a "snowman house of horror"; and snowmen representing people he hates. "The ones I "really" hate are small, so they'll melt faster," he says. There was even an occasion on which Calvin accidentally brought a snowman to life and it made itself and a small army into "deranged mutant killer monster snow goons."
Calvin's snow art is often used as a commentary on art in general. For example, Calvin has complained more than once about the lack of originality in other people's snow art and compared it with his own grotesque snow sculptures. In one of these instances, Calvin and Hobbes claim to be the sole guardians of high culture; in another, Hobbes admires Calvin's willingness to put artistic integrity above marketability, causing Calvin to reconsider and make an ordinary snowman.
Calvin and Hobbes frequently ride downhill in a wagon or sled (depending on the season), as a device to add some physical comedy to the strip and because, according to Watterson, "it's a lot more interesting ... than talking heads." While the ride is sometimes the focus of the strip, it also frequently serves as a counterpoint or visual metaphor while Calvin ponders the meaning of life, death, God, philosophy or a variety of other weighty subjects. Many of their rides end in spectacular crashes which leave them battered, beaten up and broken, a fact which convinces Hobbes to sometimes hop off before a ride even begins. In the final strip, Calvin and Hobbes depart on their sled to go exploring. This theme is similar (perhaps even homage) to scenarios in Walt Kelly's "Pogo".
G.R.O.S.S. (which stands for Get Rid Of Slimy GirlS or "otherwise it doesn't spell anything") is a club in which Calvin and Hobbes are the only members. The club was founded in the garage of their house, but to clear space for its activities, Calvin and (purportedly) Hobbes push Calvin's parents' car, causing it to roll into a ditch (but not suffer damage); the incident prompts the duo to change the club's location to Calvin's treehouse. They hold meetings that involve annoying and discomforting Susie Derkins, a girl and enemy of their club. Notable actions include planting a fake secret tape near her in attempt to draw her in to a trap, trapping her in a closet at their house and creating elaborate water balloon traps. Calvin gave himself and Hobbes important positions in the club, Calvin being "Dictator-for-Life" and Hobbes being "President-and-First-Tiger". They go into Calvin's treehouse for their club meetings and often get into fights during them. The password to get into the treehouse is intentionally long and difficult, which has on at least one occasion ruined Calvin's plans. As Hobbes is able to climb the tree without the rope, he is usually the one who comes up with the password, which often involves heaping praise upon tigers. An example of this can be seen in the comic strip where Calvin, rushing to get into the treehouse to throw things at a passing Susie Derkins, insults Hobbes, who is in the treehouse and thus has to let down the rope. Hobbes forces Calvin to say the password for insulting him. By the time Susie arrives, in time to hear Calvin saying some of the password, causing him to stumble, Calvin is on ""Verse Seven:" Tigers are perfect!/The E-pit-o-me/of good looks and grace/and quiet..uh..um..dignity". The opportunity to pelt Susie with something having passed, Calvin threatens to turn Hobbes into a rug. G.R.O.S.S. is one of the most common adventures that Calvin has. The club anthem begins: "Ohhhh Gross, best club in the cosmos..."
There are 18 "Calvin and Hobbes" books, published from 1987 to 1997. These include 11 collections, which form a complete archive of the newspaper strips, except for a single daily strip from November 28, 1985. (The collections "do" contain a strip for this date, but it is not the same strip that appeared in some newspapers. Treasuries usually combine the two preceding collections with bonus material and include color reprints of Sunday comics.)
Watterson included some new material in the treasuries. In "The Essential Calvin and Hobbes", which includes cartoons from the collections "Calvin and Hobbes" and "Something Under the Bed Is Drooling", the back cover features a scene of a giant Calvin rampaging through a town. The scene is based on Watterson's home town of Chagrin Falls, Ohio, and Calvin is holding the Chagrin Falls Popcorn Shop, an iconic candy and ice cream shop overlooking the town's namesake falls. Several of the treasuries incorporate additional poetry; "The Indispensable Calvin and Hobbes" book features a set of poems, ranging from just a few lines to an entire page, that cover topics such as Calvin's mother's "hindsight" and exploring the woods. In "The Essential Calvin and Hobbes", Watterson presents a long poem explaining a night's battle against a monster from Calvin's perspective. "The Authoritative Calvin and Hobbes" includes a story based on Calvin's use of the Transmogrifier to finish his reading homework.
A complete collection of "Calvin and Hobbes" strips, in three hardcover volumes totaling 1440 pages, was released on October 4, 2005, by Andrews McMeel Publishing. It includes color prints of the art used on paperback covers, the treasuries' extra illustrated stories and poems and a new introduction by Bill Watterson in which he talks about his inspirations and his story leading up to the publication of the strip. The alternate 1985 strip is still omitted, and three other strips (January 7 and November 24, 1987, and November 25, 1988) have altered dialogue. A four-volume paperback version was released November 13, 2012.
To celebrate the release (which coincided with the strip's 20th anniversary and the tenth anniversary of its absence from newspapers), Bill Watterson answered 15 questions submitted by readers.
Early books were printed in smaller format in black and white. These were later reproduced in twos in color in the "Treasuries" ("Essential", "Authoritative" and "Indispensable"), except for the contents of "Attack of the Deranged Mutant Killer Monster Snow Goons". Those Sunday strips were not reprinted in color until the "Complete" collection was finally published in 2005.
Watterson claims he named the books the ""Essential", "Authoritative" and "Indispensable"" because, as he says in "The Calvin and Hobbes Tenth Anniversary Book", the books are "obviously none of these things."
Reviewing "Calvin and Hobbes" in 1990, "Entertainment Weekly"'s Ken Tucker gave the strip an A+ rating, writing "Watterson summons up the pain and confusion of childhood as much as he does its innocence and fun."
In 1993, paleontologist and paleoartist Gregory S. Paul praised Bill Watterson for the scientific accuracy of the dinosaurs appearing in "Calvin and Hobbes".
In her 1994 book "When Toys Come Alive", Lois Rostow Kuznets theorizes that Hobbes serves both as a figure of Calvin's childish fantasy life and as an outlet for the expression of libidinous desires more associated with adults. Kuznets also analyzes Calvin's other fantasies, suggesting that they are a second tier of fantasies utilized in places like school where transitional objects such as Hobbes would not be socially acceptable.
Political scientist James Q. Wilson, in a paean to "Calvin and Hobbes" upon Watterson's decision to end the strip in 1995, characterized it as "our only popular explication of the moral philosophy of Aristotle."
Alisa White Coleman analyzed the strip's underlying messages concerning ethics and values in "'Calvin and Hobbes': A Critique of Society's Values," published in the "Journal of Mass Media Ethics" in 2000.
A collection of original Sunday strips was exhibited at Ohio State University's Billy Ireland Cartoon Library & Museum in 2001. Watterson himself selected the strips and provided his own commentary for the exhibition catalog, which was later published by Andrews McMeel as "Calvin and Hobbes: Sunday Pages 1985–1995".
Since the discontinuation of "Calvin and Hobbes", individual strips have been licensed for reprint in schoolbooks, including the Christian homeschooling book "The Fallacy Detective" in 2002, and the university-level philosophy reader "Open Questions: Readings for Critical Thinking and Writing" in 2005; in the latter, the ethical views of Watterson and his characters Calvin and Hobbes are discussed in relation to the views of professional philosophers. Since 2009, Twitter users have indicated that "Calvin and Hobbes" strips have appeared in textbooks for subjects in the sciences, social sciences, mathematics, philosophy and foreign language.
In a 2009 evaluation of the entire body of "Calvin and Hobbes" strips using grounded theory methodology, Christijan D. Draper found that: "Overall, "Calvin and Hobbes" suggests that meaningful time use is a key attribute of a life well lived," and that "the strip suggests one way to assess the meaning associated with time use is through preemptive retrospection by which a person looks at current experiences through the lens of an anticipated future..."
Jamey Heit's "Imagination and Meaning in Calvin and Hobbes", a critical and academic analysis of the strip, was published in 2012.
"Calvin and Hobbes" strips were again exhibited at the Billy Ireland Cartoon Library & Museum at Ohio State University in 2014, in an exhibition entitled "Exploring Calvin and Hobbes". An exhibition catalog by the same title, which also contained an interview with Watterson conducted by Jenny Robb, the curator of the museum, was published by Andrews McMeel in 2015.
Years after its original newspaper run, "Calvin and Hobbes" has continued to exert influence in entertainment, art and fandom.
In television, Calvin and Hobbes have been satirically depicted in stop motion animation in the 2006 "Robot Chicken" episode "Lust for Puppets," and in traditional animation in the 2009 "Family Guy" episode "Not All Dogs Go to Heaven." In the 2013 "Community" episode "Paranormal Parentage," the characters Abed Nadir (Danny Pudi) and Troy Barnes (Donald Glover) dress as Calvin and Hobbes, respectively, for Halloween.
British artists, merchandisers, booksellers and philosophers were interviewed for a 2009 BBC Radio 4 half-hour programme about the abiding popularity of the comic strip, narrated by Phill Jupitus.
The first book-length study of the strip, "Looking for Calvin and Hobbes: The Unconventional Story of Bill Watterson and His Revolutionary Comic Strip" by Nevin Martell, was first published in 2009; an expanded edition was published in 2010. The book chronicles Martell's quest to tell the story of "Calvin and Hobbes" and Watterson through research and interviews with people connected to the cartoonist and his work. The director of the later documentary "Dear Mr. Watterson" referenced "Looking for Calvin and Hobbes" in discussing the production of the movie, and Martell appears in the film.
The American documentary film "Dear Mr. Watterson", released in 2013, explores the impact and legacy of "Calvin and Hobbes" through interviews with authors, curators, historians, and numerous professional cartoonists.
The enduring significance of "Calvin and Hobbes" to international cartooning was recognized by the jury of the Angoulême International Comics Festival in 2014 by the awarding of its Grand Prix to Watterson, only the fourth American to ever receive the honor (after Will Eisner, Robert Crumb, and Art Spiegelman).
In 2016, 2017 and 2019, author Berkeley Breathed included "Calvin and Hobbes" in various "Bloom County" cartoons. He launched the first cartoon on April Fool's Day 2016 and jokingly issued a statement suggesting that he had acquired "Calvin and Hobbes" from Bill Watterson, who was "out of the Arizona facility, continent and looking forward to some well-earned financial security." While bearing Watterson's signature and drawing style, as well as featuring characters from both "Calvin and Hobbes" and Breathed's "Bloom County", it is unclear whether Watterson had any input into these cartoons or not.
"Calvin and Hobbes" remains the most viewed comic on GoComics, which cycles through old strips with an approximately 30-year delay.
A number of artists and cartoonists have created unofficial works portraying Calvin as a teenager/adult; the concept has also inspired writers.
In 2011, a comic strip appeared by cartoonists Dan and Tom Heyerman called "Hobbes and Bacon". The strip depicts Calvin as an adult, married to Susie Derkins with a young daughter named after philosopher Francis Bacon, to whom Calvin gives Hobbes. Though consisting of only four strips originally, "Hobbes and Bacon" received considerable attention when it appeared and was continued by other cartoonists and artists.
A novel entitled "Calvin" by CLA Young Adult Book Award–winning author Martine Leavitt was published in 2015. The story tells of seventeen-year-old Calvin—who was born on the day that "Calvin and Hobbes" ended, and who has now been diagnosed with schizophrenia—and his hallucination of Hobbes, his childhood stuffed tiger. With his friend Susie, who might also be a hallucination, Calvin sets off to find Bill Watterson in the hope that the cartoonist can provide aid for Calvin's condition. | https://en.wikipedia.org/wiki?curid=6059 |
Campaign for Real Ale
The Campaign for Real Ale (CAMRA) is an independent voluntary consumer organisation headquartered in St Albans, England, which promotes real ale, real cider, real perry and traditional British pubs and clubs. With nearly 190,000 members, it is now the largest single-issue consumer group in the UK, and is a founding member of the European Beer Consumers Union (EBCU).
The organisation was founded on 16 March 1971 in Kruger's bar in Dunquin, Kerry, Ireland by Michael Hardman, Graham Lees, Jim Makin, and Bill Mellor, who were opposed to the growing mass production of beer and the homogenisation of the British brewing industry. The original name was the Campaign for the Revitalisation of Ale. Following the formation of the Campaign, the first annual general meeting took place in 1972, at the Rose Inn in Coton Road, Nuneaton. Early membership consisted of the four founders and their friends. Interest in CAMRA and its objectives spread rapidly, with 5,000 members signed up by 1973. Other early influential members included Christopher Hutt, author of "Death of the English Pub", who succeeded Hardman as chairman, Frank Baillie, author of "The Beer Drinker's Companion", and later the many times "Good Beer Guide" editor, Roger Protz.
In the 1980s, CAMRA was influential in campaigning against the "Big Six" breweries who had a stranglehold on the UK's pubs. This resulted in The Supply of Beer (Tied Estate) Order 1989, more commonly known as the Beer Orders. The orders limited the number of pubs that a brewery could own but an unintended consequence was that many breweries ceased producing beer to focus solely on being pub-owning businesses or PubCos. In the late 1980s, CAMRA added cider campaigning to its remit with the creation of APPLE.
In 1991, CAMRA reached 30,000 members across the UK and abroad and, a year later, helped to launch the European Beer Consumers Union. CAMRA remains EBCU's largest contributor, despite the UK's exit from the European Union.
On 31 March 2016, founder Michael Hardman returned to chair a Revitalisation Project Steering Group. The aim of the Revitalisation Project was to review the organisation's purpose and future direction. Consultation meetings took place in the spring and summer of 2016, and further discussion took place at the Bournemouth AGM and Conference in spring 2017 leading to possible refinement of proposals then subject to a final vote of all members at spring 2018's AGM and Conference. At that 2018 meeting, all but one of the proposed special resolutions were passed by voting, which all members could participate in. The failed special resolution related to a proposal for the Campaign to act as the voice and represent the interests of all pub-goers and beer, cider and perry drinkers.
CAMRA's stated aims are:
CAMRA's campaigns include promoting small brewing and pub businesses, reforming licensing laws, reducing tax on beer, and stopping continued consolidation among local British brewers. It also makes an effort to promote less common varieties of beer, including stout, porter, and mild, as well as traditional cider and perry.
CAMRA's states that real ale can only be served without the use of additional carbonation. This means that "any beer brand which is produced in both cask and keg versions" is not admitted to CAMRA festivals if the brewery's marketing is deemed to imply an equivalence of quality or character between the two versions.
CAMRA is organised on a federal basis, over 200 local branches, each covering a particular geographical area of the UK, that contribute to the central body of the organisation based in St Albans. It is governed by a National Executive, made up of 12 voluntary unpaid directors elected by the membership. The local branches are grouped into 16 regions across the UK, such as the West Midlands or Wessex.
In 2009, CAMRA announced that it had reached the 100,000 members mark and subsequently went on to pass the 150,000 members mark in 2013. Member benefits include a monthly newspaper, "What's Brewing" and a quarterly "BEER" magazine, and free or reduced price admission to CAMRA-organised beer festivals. In recent times CAMRA has obtained benefits for its members from some commercial organisations and increasingly some licensed premises offer members price reductions on real ale (and sometimes cider and perry).
CAMRA publishes the "Good Beer Guide", an annually compiled directory of the best 4,500 real ale outlets and listing of real ale brewers.
CAMRA members receive a monthly newspaper called "What's Brewing" and a quarterly colour magazine called "Beer". In 2013 CAMRA launched public access to a national pub database, WhatPub. This database is maintained by local volunteers, covering all pubs in the UK and listing information such as opening times, location and beers available. It also maintains a National Inventory of Historic Pub Interiors to help bring greater recognition and protection to Britain's most historic pubs.
CAMRA supports and promotes beer and cider festivals around the country, which are organised by local CAMRA branches. Generally, each festival charges an entry fee which either covers entry only or also includes a commemorative glass showing the details of the festival. A festival programme is usually also provided, with a list and description of the drinks available. Members may get discounted entrance to CAMRA festivals.
The Campaign also organises the annual Great British Beer Festival in August. It is now held in the Great, National & West Halls at the Olympia Exhibition Centre, in Kensington, London, having been held for a few years at Earl's Court as well as regionally in the past at venues such as Brighton and Leeds. This is the UK's largest beer festival, with over 900 beers, ciders and perrys available over the week long event.
For many years, CAMRA also organised the National Winter Ales Festival. However, in 2017 this was re-branded as the Great British Beer Festival Winter. Unlike the Great British Beer Festival, the Winter event does not have a permanent venue and is rotated throughout the country every three years. Recent hosts have been Derby and Norwich, with the event currently held each February in Birmingham. In 2020 CAMRA also launched the Great Welsh Beer Festival, to be held in Cardiff in April.
CAMRA presents awards for beers and pubs, such as the National Pub of the Year. The competition begins in the preceding year with branches choosing their local pub of the year through either a ballot or a panel of judges. The branch winners are entered into 16 regional competitions which are then visited by several individuals who agree the best using a scoring system that looks at beer quality, aesthetic, and welcome. The four finalists are announced each year before a ceremony to crown the winner in the spring. There are also the Pub Design Awards, which are held in association with English Heritage and the Victorian Society. These comprise several categories, including new build, refurbished and converted pubs.
The best known CAMRA award is the Champion Beer of Britain, which is selected at the Great British Beer Festival. Other awards include the Champion Beer of Scotland and the Champion Beer of Wales.
CAMRA developed the National Beer Scoring Scheme (NBSS) as an easy to use scheme for judging beer quality in pubs, to assist CAMRA branches in selecting pubs for the "Good Beer Guide". CAMRA members input their beer scores online via WhatPub or through the Good Beer Guide app.
The CAMRA Pub Heritage Group identifies, records and helps to protect pub interiors of historic and/or architectural importance, and seeks to get them listed.
The group maintains two inventories of Heritage pubs, the National Inventory (NI), which contains only those pubs that have been maintained in their original condition (or have been modified very little) for at least thirty years, but usually since at least World War II. The second, larger, inventory is the Regional Inventory (RI), which is broken down by county and contains both those pubs listed in the NI and other pubs that are not eligible for the NI, for reasons such as having been overly modified, but are still considered historically important, or have particular architectural value.
The NI contains 289 pubs .
The LocAle scheme was launched in 2007 to promote locally brewed beers. The scheme functions slightly differently in each area, and is managed by each branch, but each is similar: if the beer is to be promoted as a LocAle it must come from a brewery within a predetermined number of miles set by each CAMRA branch, generally around 20, although the North London branch has set it at 30 miles from brewery to pub, even if it comes from a distribution centre further away; in addition, each participating pub must keep at least one LocAle for sale at all times.
CAMRA members may join the CAMRA Members' Investment Club which, since 1989, has invested in real ale breweries and pub chains, although all investors must be CAMRA members. | https://en.wikipedia.org/wiki?curid=6060 |
CNO cycle
The CNO cycle (for carbon–nitrogen–oxygen) is one of the two known sets of fusion reactions by which stars convert hydrogen to helium, the other being the proton–proton chain reaction (pp-chain reaction). Unlike the latter, the CNO cycle is a catalytic cycle. It is hypothesized to be dominant in stars that are more than 1.3 times as massive as the Sun.
In the CNO cycle, four protons fuse, using carbon, nitrogen, and oxygen isotopes as catalysts, to produce one alpha particle, two positrons and two electron neutrinos. Although there are various paths and catalysts involved in the CNO cycles, all these cycles have the same net result:
The positrons will almost instantly annihilate with electrons, releasing energy in the form of gamma rays. The neutrinos escape from the star carrying away some energy. One nucleus goes on to become carbon, nitrogen, and oxygen isotopes through a number of transformations in an endless loop.
The proton–proton chain is more prominent in stars the mass of the Sun or less. This difference stems from temperature dependency differences between the two reactions; pp-chain reaction starts at temperatures around (4 megakelvin), making it the dominant energy source in smaller stars. A self-maintaining CNO chain starts at approximately , but its energy output rises much more rapidly with increasing temperatures so that it becomes the dominant source of energy at approximately .
The Sun has a core temperature of around , and only of nuclei produced in the Sun are
born in the CNO cycle. The CNO-I process was independently proposed by Carl von Weizsäcker and Hans Bethe in the late 1930s.
Under typical conditions found in stars, catalytic hydrogen burning by the CNO cycles is limited by proton captures. Specifically, the timescale for beta decay of the radioactive nuclei produced is faster than the timescale for fusion. Because of the long timescales involved, the cold CNO cycles convert hydrogen to helium slowly, allowing them to power stars in quiescent equilibrium for many years.
The first proposed catalytic cycle for the conversion of hydrogen into helium was initially called the carbon–nitrogen cycle (CN-cycle), also referred to as the Bethe–Weizsäcker cycle in honor of the independent work of Carl von Weizsäcker in 1937-38 and Hans Bethe. Bethe's 1939 papers on the CN-cycle drew on three earlier papers written in collaboration with Robert Bacher and Milton Stanley Livingston and which came to be known informally as "Bethe's Bible." It was considered the standard work on nuclear physics for many years and was a significant factor in his being awarded the 1967 Nobel Prize in Physics. Bethe's original calculations suggested the CN-cycle was the Sun's primary source of energy. This conclusion arose from what is now-known as a mistaken belief: that the abundance of nitrogen in the sun is approximately 10%, when it is actually less than half a percent. The CN-cycle, named as it contains no stable isotope of oxygen involves the following cycle of transformations: → → → → → → . This cycle is now understood as being the first part of a larger process, the CNO-cycle, and the main reactions in this part of the cycle (CNO-I) are:
where the carbon-12 nucleus used in the first reaction is regenerated in the last reaction. After the two positrons emitted annihilate with two ambient electrons producing an additional 2.04 MeV, the total energy released in one cycle is 26.73 MeV; in some texts, authors are erroneously including the positron annihilation energy in with the beta-decay Q-value and then neglecting the equal amount of energy released by annihilation, leading to possible confusion. All values are calculated with reference to the Atomic Mass Evaluation 2003.
The limiting (slowest) reaction in the CNO-I cycle is the proton capture on . In 2006 it was experimentally measured down to stellar energies, revising the calculated age of globular clusters by around 1 billion years.
The neutrinos emitted in beta decay will have a spectrum of energy ranges, because although momentum is conserved, the momentum can be shared in any way between the positron and neutrino, with either emitted at rest and the other taking away the full energy, or anything in between, so long as all the energy from the Q-value is used. The total momentum received by the electron and the neutrino is not great enough to cause a significant recoil of the much heavier daughter nucleus and hence, its contribution to kinetic energy of the products, for the precision of values given here, can be neglected. Thus the neutrino emitted during the decay of nitrogen-13 can have an energy from zero up to 1.20 MeV, and the neutrino emitted during the decay of oxygen-15 can have an energy from zero up to 1.73 MeV. On average, about 1.7 MeV of the total energy output is taken away by neutrinos for each loop of the cycle, leaving about 25 MeV available for producing luminosity.
In a minor branch of the above reaction, occurring in the Sun's core 0.04% of the time, the final reaction involving shown above does not produce carbon-12 and an alpha particle, but instead produces oxygen-16 and a photon and continues →→→→→→:
Like the carbon, nitrogen, and oxygen involved in the main branch, the fluorine produced in the minor branch is merely an intermediate product and at steady state, does not accumulate in the star.
This subdominant branch is significant only for massive stars. The reactions are started when one of the reactions in CNO-II results in fluorine-18 and gamma instead of nitrogen-14 and alpha, and continues →→→→→→:
Like the CNO-III, this branch is also only significant in massive stars. The reactions are started when one of the reactions in CNO-III results in fluorine-19 and gamma instead of nitrogen-15 and alpha, and continues →→→→→→:
In some instances can combine with a helium nucleus to start a sodium-neon cycle
Under conditions of higher temperature and pressure, such as those found in novae and x-ray bursts, the rate of proton captures exceeds the rate of beta-decay, pushing the burning to the proton drip line. The essential idea is that a radioactive species will capture a proton before it can beta decay, opening new nuclear burning pathways that are otherwise inaccessible. Because of the higher temperatures involved, these catalytic cycles are typically referred to as the hot CNO cycles; because the timescales are limited by beta decays instead of proton captures, they are also called the beta-limited CNO cycles.
The difference between the CNO-I cycle and the HCNO-I cycle is that captures a proton instead of decaying, leading to the total sequence →→→→→→:
The notable difference between the CNO-II cycle and the HCNO-II cycle is that captures a proton instead of decaying, and neon is produced in a subsequent reaction on , leading to the total sequence →→→→→→:
An alternative to the HCNO-II cycle is that captures a proton moving towards higher mass and using the same helium production mechanism as the CNO-IV cycle as →→→→→→:
While the total number of "catalytic" nuclei are conserved in the cycle, in stellar evolution the relative proportions of the nuclei are altered. When the cycle is run to equilibrium, the ratio of the carbon-12/carbon-13 nuclei is driven to 3.5, and nitrogen-14 becomes the most numerous nucleus, regardless of initial composition. During a star's evolution, convective mixing episodes moves material, within which the CNO cycle has operated, from the star's interior to the surface, altering the observed composition of the star. Red giant stars are observed to have lower carbon-12/carbon-13 and carbon-12/nitrogen-14 ratios than do main sequence stars, which is considered to be convincing evidence for the operation of the CNO cycle. | https://en.wikipedia.org/wiki?curid=6061 |
Craps
Craps is a dice game in which the players make wagers on the outcome of the roll, or a series of rolls, of a pair of dice. Players may wager money against each other (playing "street craps") or a bank (playing "casino craps", also known as "table craps", or often just "craps"). Because it requires little equipment, "street craps" can be played in informal settings. While shooting craps, players may use slang terminology to place bets and actions.
In 1788 "Krabs" (later spelled Crabs) was an English variation on the dice game Hasard (also spelled Hazard).
Craps developed in the United States from a simplification of the western European game of hazard. The origins of hazard are obscure and may date to the Crusades. Hazard was brought from London to New Orleans in approximately 1805 by the returning Bernard Xavier Philippe de Marigny de Mandeville, the young gambler and scion of a family of wealthy landowners in colonial Louisiana. Although in hazard the dice shooter may choose any number from five to nine to be his main number, de Marigny simplified the game such that the main number is always seven, which is the mathematically optimal choice. Both hazard and its simpler derivative were unfamiliar to and rejected by Americans of his social class, leading de Marigny to introduce his novelty to the local underclass. Fieldhands taught their friends and deckhands carried the new game up the Mississippi River and its tributaries. Celebrating the popular success of his novelty, de Marigny gave the name Rue de Craps to a street in his new subdivision in New Orleans.
The central game, called Pass from the French word Pas meaning pace or step, has been gradually supplemented over the decades by many companion games which can be played simultaneously with Pass. Now applied to the entire collection of games, the name craps derives from an underclass Louisiana mispronunciation of the word "crabs", which in aristocratic London had been the epithet for the numbers two and three. In hazard, both "crabs" are always instant losing numbers for the first dice roll regardless of the shooter's selected main number. Also in hazard, if the main number is seven then the number twelve is added to the "crabs" as a losing number on the first dice roll. This structure is retained in the simplified game called Pass. All three losing numbers on the first roll of Pass are jointly called the craps numbers.
For a century after its invention, craps was abused by casinos using unfair dice. To remedy the problem, in approximately 1907 a dice maker named John H. Winn in Philadelphia introduced a layout which featured bets on Don't Pass as well as Pass. Virtually all modern casinos use his innovation, which provides incentive for casinos to use perfectly fair dice.
Craps exploded in popularity during World War II, which brought most young American men of every social class into the military. The street version of craps was popular among servicemembers who often played it using a blanket as a shooting surface. Their military memories led to craps becoming the dominant casino game in postwar Las Vegas and the Caribbean. After 1960 a few casinos in Europe, Australia, and Macau began offering craps and after 2004 online casinos extended its spread globally.
Bank craps or casino craps is played by one or more players betting against the casino rather than each other. Both the players and the dealers stand around a large rectangular craps table. Sitting is discouraged by most casinos unless a player has medical reasons for requiring a seat.
Players use casino chips rather than cash to bet on the Craps "layout," a fabric surface which displays the various bets. The bets vary somewhat among casinos in availability, locations, and payouts. The tables roughly resemble bathtubs and come in various sizes. In some locations, chips may be called checks, tokens, or plaques.
Against one long side is the casino's table bank: as many as two thousand casino chips in stacks of 20. The opposite long side is usually a long mirror. The U-shaped ends of the table have duplicate layouts and standing room for approximately eight players. In the center of the layout is an additional group of bets which are used by players from both ends. The vertical walls at each end are usually covered with a rubberized target surface covered with small pyramid shapes to randomize the dice which strike them. The top edges of the table walls have one or two horizontal grooves in which players may store their reserve chips.
The table is run by up to four casino employees: a boxman seated (usually the only seated employee) behind the casino's bank, who manages the chips, supervises the dealers, and handles "coloring up" players (exchanging small chip denominations for larger denominations in order to preserve the chips at a table); two base dealers who stand to either side of the boxman and collect and pay bets to players around their half of the table; and a stickman who stands directly across the table from the boxman, takes and pays (or directs the base dealers to do so) the bets in the center of the table, announces the results of each roll (usually with a distinctive patter), and moves the dice across the layout with an elongated wooden stick.
Each employee also watches for mistakes by the others because of the sometimes large number of bets and frantic pace of the game. In smaller casinos or at quiet times of day, one or more of these employees may be missing, and have their job covered by another, or cause player capacity to be reduced.
Some smaller casinos have introduced "mini-craps" tables which are operated with only two dealers; rather than being two essentially identical sides and the center area, a single set of major bets is presented, split by the center bets. Responsibility of the dealers is adjusted: the stickman continuing to handle the center bets, and the base dealer handling the other bets as well as cash and chip exchanges.
By contrast, in "street craps", there is no marked table and often the game is played with no back-stop against which the dice are to hit. (Despite the name "street craps", this game is often played in houses, usually on an un-carpeted garage or kitchen floor.) The wagers are made in cash, never in chips, and are usually thrown down onto the ground or floor by the players. There are no attendants, and so the progress of the game, fairness of the throws, and the way that the payouts are made for winning bets are self-policed by the players.
Each casino may set which bets are offered and different payouts for them, though a core set of bets and payouts is typical. Players take turns rolling two dice and whoever is throwing the dice is called the "shooter". Players can bet on the various options by placing chips directly on the appropriately-marked sections of the layout, or asking the base dealer or stickman to do so, depending on which bet is being made.
While acting as the shooter, a player must have a bet on the "Pass" line and/or the "Don't Pass" line. "Pass" and "Don't Pass" are sometimes called "Win" and "Don't Win" or "Right" and "Wrong" bets. The game is played in rounds and these "Pass" and "Don't Pass" bets are betting on the outcome of a round. The shooter is presented with multiple dice (typically five) by the "stickman", and must choose two for the round. The remaining dice are returned to the stickman's bowl and are not used.
Each round has two phases: "come-out" and "point". Dice are passed to the left. To start a round, the shooter makes one or more "come-out" rolls. The shooter must shoot toward the farther back wall and is generally required to hit the farther back wall with both dice. Casinos may allow a few warnings before enforcing the dice to hit the back wall and are generally lenient if at least one die hits the back wall. Both dice must be tossed in one throw. If only one die is thrown the shot is invalid. A come-out roll of 2, 3 or 12 is called "craps" or "crapping out", and anyone betting the Pass line loses. On the other hand, anyone betting the Don't Pass line on come out wins with a roll of 2 or 3 and ties (pushes) if a 12 is rolled. Shooters may keep rolling after crapping out; the dice are only required to be passed if a shooter sevens out (rolls a seven after a point has been established). A come-out roll of 7 or 11 is a "natural"; the Pass line wins and Don't Pass loses. The other possible numbers are the point numbers: 4, 5, 6, 8, 9, and 10. If the shooter rolls one of these numbers on the come-out roll, this establishes the "point" – to "pass" or "win", the point number must be rolled again before a seven.
The dealer flips a button to the "On" side and moves it to the point number signifying the second phase of the round. If the shooter "hits" the point value again (any value of the dice that sum to the point will do; the shooter doesn't have to exactly repeat the exact combination of the come-out roll) before rolling a seven, the Pass line wins and a new round starts. If the shooter rolls any seven before repeating the point number (a "seven-out"), the Pass line loses, the Don't Pass line wins, and the dice pass clockwise to the next new shooter for the next round. Once a point has been established any multi-roll bet (including Pass and/or Don't Pass line bets and odds) are unaffected by the 2, 3, 11 or 12; the only numbers which affect the round are the established point, any specific bet on a number, or any 7. Any single roll bet is always affected (win or lose) by the outcome of any roll.
While the come-out roll may specifically refer to the first roll of a new shooter, any roll where no point is established may be referred to as a come-out. By this definition the start of any new round regardless if it is the shooter's first toss can be referred to as a come-out roll.
Any player can make a bet on pass or don't pass as long as a point has not been established, or come or don't come as long as a point is established. All other bets, including an increase in odds behind the pass and don't pass lines, may be made at any time. All bets other than pass line and come may be removed or reduced any time before the bet loses. This is known as "taking it down" in craps.
The maximum bet for Place, Buy, Lay, Pass and Come bets are generally equal to table maximum. Lay bet maximum are equal to the table maximum win, so if a player wishes to lay the 4 or 10, he or she may bet twice at amount of the table maximum for the win to be table maximum. Odds behind Pass, Come, Don't Pass and Don't Come may be however larger than the odds offered allows and can be greater than the table maximum in some casinos. Don't odds are capped on the maximum allowed win some casino allow the odds bet itself to be larger than the maximum bet allowed as long as the win is capped at maximum odds. Single rolls bets can be lower than the table minimum, but the maximum bet allowed is also lower than the table maximum. The maximum allowed single roll bet is based on the maximum allowed win from a single roll.
In all the above scenarios, whenever the Pass line wins, the Don't Pass line loses, and vice versa, with one exception: on the come-out roll, a roll of 12 will cause Pass Line bets to lose, but Don't Pass bets are pushed (or "barred"), neither winning nor losing. (The same applies to "Come" and "Don't Come" bets, discussed below.)
A player wishing to play craps without being the shooter should approach the craps table and first check to see if the dealer's "On" button is on any of the point numbers.
In either case, all single or multi-roll proposition bets may be placed in either of the two rounds.
Between dice rolls there is a period for dealers to make payouts and collect losing bets, after which players can place new bets. The stickman monitors the action at a table and decides when to give the shooter the dice, after which no more betting is allowed.
When joining the game, one should place money on the table rather than passing it directly to a dealer. The dealer's exaggerated movements during the process of "making change" or "change only" (converting currency to an equivalent in casino cheques) are required so that any disputes can be later reviewed against security camera footage.
The dealers will insist that the shooter roll with one hand and that the dice bounce off the far wall surrounding the table. These requirements are meant to keep the game fair (preventing switching the dice or making a "controlled shot"). If a die leaves the table, the shooter will usually be asked to select another die from the remaining three but can request permission to use the same die if it passes the boxman's inspection. This requirement exists to keep the game fair and reduce the chance of loaded dice.
There are many local variants of the calls made by the stickman for rolls during a craps game. These frequently incorporate a reminder to the dealers as to which bets to pay or collect.
Rolls of 4, 6, 8, and 10 are called "hard" or "easy" (e.g. "six the hard way", "easy eight", "hard ten") depending on whether they were rolled as a "double" or as any other combination of values, because of their significance in center table bets known as the "hard ways". Hard way rolls are so named because there is only one way to roll them (i.e., the value on each die is the same when the number is rolled). Consequently, it is more likely to roll the number in combinations (easy) rather than as a double (hard).
The shooter is required to make either a pass line bet or a Don't Pass bet if he wants to shoot. On the come out roll each player may only make one bet on the Pass or Don't Pass, but may bet both if desired. The Pass Line and Don't Pass bet is optional for any player not shooting. In rare cases, some casinos require all players to make a minimum Pass Line or Don't Pass bet (if they want to make any other bet), whether they are currently shooting or not.
The fundamental bet in craps is the pass line bet, which is a bet for the shooter to win. This bet must be at least the table minimum and at most the table maximum.
The pass line bet pays even money.
The pass line bet is a contract bet. Once a pass line bet is made, it is always working and cannot be turned "Off", taken down, or reduced until a decision is reached – the point is made, or the shooter sevens out. A player may increase any corresponding odds (up to the table limit) behind the Pass line at any time after a point is established. Players may only bet the pass line on the come out roll when no point has been established, unless the casino allows put betting where the player can bet Pass line or increase an existing Pass line bet whenever desired and may take odds immediately if the point is already on.
A don't pass bet is a bet for the shooter to lose ("seven out, line away") and is almost the opposite of the pass line bet. Like the Pass bet, this bet must be at least the table minimum and at most the table maximum.
The don't pass bet pays even money.
The don't pass bet is a no-contract bet. After a point is established, a player may take down or reduce a don't pass bet and any corresponding odds at any time because odds of rolling a 7 before the point is in the player's favor. Once taken down or reduced, however, the don't pass bet may not be restored or increased. Because the shooter must have a line bet the shooter generally may not reduce a don't pass bet below the table minimum. In Las Vegas, a majority of casinos will allow the shooter to move the bet to the pass line in lieu of taking it down, however in other areas such as Pennsylvania and Atlantic City, this is not allowed. Even though players are allowed to remove the don't pass line bet after a point has been established, the bet cannot be turned "Off" without being removed. If a player chooses to remove the don't pass line bet, he or she can no longer lay odds behind the don't pass line. The player can, however, still make standard lay bets on any of the point numbers (4, 5, 6, 8, 9, 10).
There are two different ways to calculate the odds and house edge of this bet. The table below gives the numbers considering that the game ends in a push when a 12 is rolled, rather than being undetermined. Betting on don't pass is often called "playing the dark side", and it is considered by some players to be in poor taste, or even taboo, because it goes directly against conventional play, winning when most of the players lose.
If a 4, 5, 6, 8, 9, or 10 is thrown on the come-out roll (i.e., if a point is established), most casinos allow pass line players to take odds by placing up to some predetermined multiple of the pass line bet, behind the pass line. This additional bet wins if the point is rolled again before a 7 is rolled (the point is made) and pays at the true odds of 2-to-1 if 4 or 10 is the point, 3-to-2 if 5 or 9 is the point, or 6-to-5 if 6 or 8 is the point. Unlike the pass line bet itself, the pass line odds bet can be turned "Off" (not working), removed or reduced anytime before it loses. In Las Vegas, generally odds bets are required to be the table minimum. In Atlantic City and Pennsylvania, the combine odds and pass bet must be table minimum so players can bet the minimum single unit on odds depending on the point. If the point is a 4 or 10 players can bet as little as $1 on odds if the table minimum is low such as is $5, $10 or $15. If the player requests the pass odds be not working ("Off") and the shooter sevens-out or hits the point, the pass line bet will be lost or doubled and the pass odds returned.
Individual casinos (and sometimes tables within a casino) vary greatly in the maximum odds they offer, from single or double odds (one or two times the pass line bet) up to 100x or even unlimited odds. A variation often seen is "3-4-5X Odds", where the maximum allowed odds bet depends on the point: three times if the point is 4 or 10; four times on points of 5 or 9; or five times on points of 6 or 8. This rule simplifies the calculation of winnings: a maximum pass odds bet on a 3–4–5× table will always be paid at six times the pass line bet regardless of the point.
As odds bets are paid at true odds, in contrast with the pass line which is always even money, taking odds on a minimum pass line bet lessens the house advantage compared with betting the same total amount on the pass line only. A maximum odds bet on a minimum pass line bet often gives the lowest house edge available in any game in the casino. However, the odds bet cannot be made independently, so the house retains an edge on the pass line bet itself.
If a player is playing don't pass instead of pass, they may also lay odds by placing chips behind the don't pass line. If a 7 comes before the point is rolled, the odds pay at true odds of 1-to-2 if 4 or 10 is the point, 2-to-3 if 5 or 9 is the point, or 5-to-6 if 6 or 8 is the point. Typically the maximum lay bet will be expressed such that a player may win up to an amount equal to the maximum odds multiple at the table. If a player lays maximum odds with a point of 4 or 10 on a table offering five-times odds, he would be able to lay a maximum of ten times the amount of his Don't Pass bet. At 5x odds table, the maximum amount the combined bet can win will always be 6x the amount of the Don't Pass bet. Players can bet table minimum odds if desired and win less than table minimum. Like the Don't Pass bet the odds can be removed or reduced. Unlike the don't pass bet itself, the don't pass odds can be turned "Off" (not working). In Las Vegas generally odds bets are required to be the table minimum. In Atlantic City and Pennsylvania, the combine lay odds and Don't Pass bet must be table minimum so players may bet as little as the minimum two units on odds depending on the point. If the point is a 4 or 10 players can bet as little as $2 if the table minimum is low such as $5, $10 or $15 tables. If the player requests the don't pass odds to be not working ("Off") and the shooter hits the point or sevens-out, the don't pass bet will be lost or doubled and the don't pass odds returned. Unlike a standard lay bet on a point, lay odds behind the Don't Pass line does not charge commission (vig).
A Come bet can be visualized as starting an entirely new pass line bet, unique to that player. Like the Pass Line each player may only make one Come bet per roll, this does not exclude a player from betting odds on an already established Come point. This bet must be at least the table minimum and at most the table maximum. Players may bet both the Come and Don't Come on the same roll if desired. Come bets can only be made after a point has been established since, on the come-out roll, a Come bet would be the same thing as a pass line bet. A player making a Come bet will bet on the first point number that "comes" from the shooter's next roll, regardless of the table's round. If a 7 or 11 is rolled on the first round, it wins. If a 2, 3, or 12 is rolled, it loses. If instead the roll is 4, 5, 6, 8, 9, or 10, the Come bet will be moved by the base dealer onto a box representing the number the shooter threw. This number becomes the "come-bet point" and the player is allowed to take odds, just like a pass line bet. Also like a pass line bet, the come bet is a contract bet and is always working, and cannot be turned "Off", removed or reduced until it wins or loses. However, the odds taken behind a Come bet can be turned "Off" (not working), removed or reduced anytime before the bet loses. In Las Vegas generally odds bets are required to be the table minimum. In Atlantic City and Pennsylvania, the combine odds and pass bet must be table minimum so players can bet the minimum single unit depending on the point. If the point is a 4 or 10 players can bet as little as $1 if the table minimum is low such as $5, $10 or $15 minimums. If the player requests the Come odds to be not working ("Off") and the shooter sevens-out or hits the Come bet point, the Come bet will be lost or doubled and the Come odds returned. If the casino allows put betting a player may increase a Come bet after a point has been established and bet larger odds behind if desired. Put betting also allows a player to bet on a Come and take odds immediately on a point number without a Come bet point being established.
The dealer will place the odds on top of the come bet, but slightly off center in order to differentiate between the original bet and the odds. The second round wins if the shooter rolls the come bet point again before a seven. Winning come bets are paid the same as winning pass line bets: even money for the original bet and true odds for the odds bet. If, instead, the seven is rolled before the come-bet point, the come bet (and any odds bet) loses.
Because of the come bet, if the shooter makes their point, a player can find themselves in the situation where they still have a come bet (possibly with odds on it) and the next roll is a come-out roll. In this situation, odds bets on the come wagers are usually presumed to be not working for the come-out roll. That means that if the shooter rolls a 7 on the come-out roll, any players with active come bets waiting for a come-bet point lose their initial wager but will have their odds bets returned to them.
If the come-bet point is rolled on the come-out roll, the odds do not win but the come bet does and the odds bet is returned (along with the come bet and its payoff). The player can tell the dealer that they want their odds working, such that if the shooter rolls a number that matches the come point, the odds bet will win along with the come bet, and if a seven is rolled, both lose.
Many players will use a come bet as "insurance" against sevening out: if the shooter rolls a seven, the come bet pays 1:1, offsetting the loss of the pass line bet. The risk in this strategy is the situation where the shooter does not hit a seven for several rolls, leading to multiple come bets that will be lost if the shooter eventually sevens out.
In the same way that a come bet is similar to a pass line bet, a don't come bet is similar to a don't pass bet. Like the come, the don't come can only be bet after a point has already been established as it is the same as a don't pass line bet when no point is established. This bet must be at least the table minimum and at most the table maximum. A don't come bet is played in two rounds. If a 2 or 3 is rolled in the first round, it wins. If a 7 or 11 is rolled, it loses. If a 12 is rolled, it is a push (subject to the same 2/12 switch described above for the don't pass bet). If, instead, the roll is 4, 5, 6, 8, 9, or 10, the don't come bet will be moved by the base dealer onto a box representing the number the shooter threw. The second round wins if the shooter rolls a seven before the don't come point. Like the Don't Pass each player may only make one Don't Come bet per roll, this does not exclude a player from laying odds on an already established Don't Come points. Players may bet both the Don't Come and Come on the same roll if desired.
The player may lay odds on a don't come bet, just like a don't pass bet; in this case, the dealer (not the player) places the odds bet on top of the bet in the box, because of limited space, slightly offset to signify that it is an odds bet and not part of the original don't come bet. Lay odds behind a Don't Come are subject to the same rules as Don't Pass lay odds. Unlike a standard lay bet on a point, lay odds behind a don't come point does not charge commission (vig) and gives the player true odds. Like the don't pass line bet, don't come bets are no-contract, and can be removed or reduced after a don't come point has been established, but cannot be turned off ("not working") without being removed. A player may also call, "No Action" when a point is established, and the bet will not be moved to its point. This play is not to the player's advantage. If the bet is removed, the player can no longer lay odds behind the don't come point and cannot restore or increase the same don't come bet. Players must wait until next roll as long as a pass line point has been established (players cannot bet don't come on come out rolls) before they can make a new don't come bet. Las Vegas casinos which allow put betting allows players to move the Don't Come directly to any Come point as a put, however this is not allowed in Atlantic City or Pennsylvania. Unlike the don't come bet itself, the don't come odds can be turned "Off" (not working), removed or reduced if desired. In Las Vegas, players generally must lay at least table minimum on odds if desired and win less than table minimum; in Atlantic City and Pennsylvania a player's combined bet must be at least table minimum, so depending on the point number players may lay as little as 2 minimum units (e.g. if the point is 4 or 10). If the player requests the don't come odds be not working ("Off") and the shooter hits the don't come point or sevens-out, the don't come bet will be lost or doubled and the don't come odds returned.
Winning don't come bets are paid the same as winning don't pass bets: even money for the original bet and true odds for the odds lay. Unlike come bets, the odds laid behind points established by don't come bets are always working including come out rolls unless the player specifies otherwise.
These are bets that may not be settled on the first roll and may need any number of subsequent rolls before an outcome is determined.
Most multi-roll bets may fall into the situation where a point is made by the shooter before the outcome of the multi-roll bet is decided. These bets are often considered "not working" on the new come-out roll until the next point is established, unless the player calls the bet as "working."
Casino rules vary on this; some of these bets may not be callable, while others may be considered "working" during the come-out. Dealers will usually announce if bets are working unless otherwise called off. If a non-working point number placed, bought or laid becomes the new point as the result of a come-out, the bet is usually refunded, or can be moved to another number for free.
Players can bet any point number (4, 5, 6, 8, 9, 10) by placing their wager in the come area and telling the dealer how much and on what number(s), "30 on the 6", "5 on the 5" or "25 on the 10". These are typically "Place Bets to Win". These are bets that the number bet on will be rolled before a 7 is rolled. These bets are considered working bets, and will continue to be paid out each time a shooter rolls the number bet. On a come-out roll, a place bet is considered to be not in effect unless the player who made it specifies otherwise. This bet may be removed or reduced at any time until it loses; in the latter case, the player must abide by any table minimums.
Place bets to win pay out at slightly worse than the true odds: 9-to-5 on points 4 or 10, 7-to-5 on points 5 or 9, and 7-to-6 on points 6 or 8. The place bets on the outside numbers (4,5,9,10) should be made in units of $5, (on a $5 minimum table), in order to receive the correct exact payout of $5 paying $7 or $5 paying $9. The place bets on the 6 & 8 should be made in units of $6, (on a $5 minimum table), in order to receive the correct exact payout of $6 paying $7. For the 4 and 10, it is to the player's advantage to 'buy' the bet (see below).
An alternative form, rarely offered by casinos, is the "place bet to lose." This bet is the opposite of the place bet to win and pays off if a 7 is rolled before the specific point number. The place bet to lose typically carries a lower house edge than a place bet to win. Payouts are 4–5 on points 6 or 8, 5–8 on 5 or 9, and 5–11 on 4 or 10.
Players can also buy a bet which are paid at true odds, but a 5% commission is charged on the amount of the bet. Buy bets are placed with the shooter betting at a specific number will come out before a player sevens out. The buy bet must be at least table minimum excluding commission, however some casinos require the minimum buy bet amount to be at least $20 to match the $1 charged on the 5% commission. Traditionally, the buy bet commission is paid no matter what, but in recent years a number of casinos have changed their policy to charge the commission only when the buy bet wins. Some casinos charge the commission as a one-time fee to buy the number; payouts are then always at true odds. Most casinos usually charge only $1 for a $25 green-chip bet (4% commission), or $2 for $50 (two green chips), reducing the house advantage a bit more. Players may remove or reduce this bet (bet must be at least table minimum excluding vig) anytime before it loses. Buy bets like place bets are not working when no point has been established unless the player specifies otherwise.
Where commission is charged only on wins, the commission is often deducted from the winning payoff—a winning $25 buy bet on the 10 would pay $49, for instance. The house edges stated in the table assume the commission is charged on all bets. They are reduced by at least a factor of two if commission is charged on winning bets only.
A lay bet is the opposite of a buy bet, where a player bets on a 7 to roll before the number that is laid. Players may only lay the 4, 5, 6, 8, 9 or 10 and may lay multiple numbers if desired. Just like the buy bet lay bets pay true odds, but because the lay bet is the opposite of the buy bet, the payout is reversed. Therefore, players get 1 to 2 for the numbers 4 and 10, 2 to 3 for the numbers 5 and 9, and 5 to 6 for the numbers 6 and 8. A 5% commission (vigorish, vig, juice) is charged up front on the possible winning amount. For example: A $40 Lay Bet on the 4 would pay $20 on a win. The 5% vig would be $1 based on the $20 win. (not $2 based on the $40 bet as the way buy bet commissions are figured.) Like the buy bet the commission is adjusted to suit the betting unit such that fraction of a dollar payouts are not needed. Casinos may charge the vig up front thereby requiring the player to pay a vig win or lose, other casinos may only take the vig if the bet wins. Taking vig only on wins lowers house edge. Players may removed or reduce this bet (bet must be at least table minimum) anytime before it loses. Some casinos in Las Vegas allow players to lay table minimum plus vig if desired and win less than table minimum. Lay bet maximums are equal to the table maximum win, so if a player wishes to lay the 4 or 10, he or she may bet twice at amount of the table maximum for the win to be table maximum. Other casinos require the minimum bet to win at $20 even at the lowest minimum tables in order to match the $1 vig, this requires a $40 bet. Similar to buy betting, some casinos only take commission on win reducing house edge. Unlike place and buy bets, lay bets are always working even when no point has been established. The player must specify otherwise if he or she wishes to have the bet not working.
If a player is unsure of whether a bet is a single or multi-roll bet, it can be noted that all single-roll bets will be displayed on the playing surface in one color (usually red), while all multi-roll bets will be displayed in a different color (usually yellow).
A put bet is a bet which allows players to increase or make a Pass line bet after a point has been established (after come-out roll). Players may make a put bet on the Pass line and take odds immediately or increase odds behind if a player decides to add money to an already existing Pass line bet. Put betting also allows players to increase an existing come bet for additional odds after a come point has been established or make a new come bet and take odds immediately behind if desired without a come bet point being established. If increased or added put bets on the Pass line and Come cannot be turned "Off", removed or reduced, but odds bet behind can be turned "Off", removed or reduced. The odds bet is generally required to be the table minimum. Player cannot put bet the Don't Pass or Don't Come. Put betting may give a larger house edge over place betting unless the casino offers high odds.
Put bets are generally allowed in Las Vegas, but not allowed in Atlantic City and Pennsylvania.
Put bets are better than place bets (to win) when betting more than 5-times odds over the flat bet portion of the put bet. For example, a player wants a $30 bet on the six. Looking at two possible bets: 1) Place the six, or 2) Put the six with odds. A $30 place bet on the six pays $35 if it wins. A $30 put bet would be a $5 flat line bet plus $25 (5-times) in odds, and also would pay $35 if it wins. Now, with a $60 bet on the six, the place bet wins $70, where the put bet ($5 + $55 in odds) would pay $71. The player needs to be at a table which not only allows put bets, but also high-times odds, to take this advantage.
This bet can only be placed on the numbers 4, 6, 8, and 10. In order for this bet to win, the chosen number must be rolled the "hard way" (as doubles) before a 7 or any other non-double combination ("easy way") totaling that number is rolled. For example, a player who bets a hard 6 can only win by seeing a 3–3 roll come up before any 7 or any easy roll totaling 6 (4–2 or 5–1); otherwise, he/she loses.
In Las Vegas casinos, this bet is generally working, including when no point has been established, unless the player specifies otherwise. In other casinos such as those in Atlantic City, hard ways are not working when the point is off unless the player requests to have it working on the come out roll.
Like single-roll bets, hard way bets can be lower than the table minimum; however, the maximum bet allowed is also lower than the table maximum. The minimum hard way bet can be a minimum one unit. For example, lower stake table minimums of $5 or $10, generally allow minimum hard ways bets of $1. The maximum bet is based on the maximum allowed win from a single roll.
Easy way is not a specific bet offered in standard casinos, but a term used to define any number combination which has two ways to roll. For example, (6–4, 4–6) would be a "10 easy". The 4, 6, 8 or 10 can be made both hard and easy ways. Betting point numbers (which pays off on easy or hard rolls of that number) or single-roll ("hop") bets (e.g., "hop the 2–4" is a bet for the next roll to be an easy six rolled as a two and four) are methods of betting easy ways.
A player can choose either the 6 or 8 being rolled before the shooter throws a seven. These wagers are usually avoided by experienced craps players since they pay even money (1:1) while a player can make place bets on the 6 or the 8, which pay more (7:6). Some casinos (especially all those in Atlantic City) do not even offer the Big 6 & 8. The bets are located in the corners behind the pass line, and bets may be placed directly by players.
The only real advantage offered by the Big 6 & 8 is that they can be bet for the table minimum, whereas a place bet minimum may sometimes be greater than the table minimum (e.g. $6 place bet on a $3 minimum game.) In addition place bets are usually not working, except by agreement, when the shooter is "coming out" i.e. shooting for a point, and Big 6 and 8 bets always work. Some modern layouts no longer show the Big 6/Big 8 bet.
Single-roll (proposition) bets are resolved in one dice roll by the shooter. Most of these are called "service bets", and they are located at the center of most craps tables. Only the stickman or a dealer can place a service bet. Single-roll bets can be lower than the table minimum, but the maximum bet allowed is also lower than the table maximum. The maximum bet is based on the maximum allowed win from a single roll. The lowest single-roll bet can be a minimum one unit bet. For example, tables with minimums of $5 or $10 generally allow minimum single-roll bets of $1. Single bets are always working by default unless the player specifies otherwise. The bets include:
2 (snake eyes, or Aces): Wins if shooter rolls a 2.
3 (ace-deuce): Wins if the shooter rolls a 3.
Yo: Wins if the shooter rolls 11.
12 (boxcars, midnight, or cornrows): Wins if shooter rolls a 12.
2 or 12 (hi-lo): Wins if shooter rolls a 2 or 12. The stickman places this bet on the line dividing the 2 and 12 bets.
Any Craps (Three-Way): Wins if the shooter rolls 2, 3 or 12.
C & E: A combined bet, a player is betting half their bet on craps (2,3,12) and the other half on 11 (yo). The combine payout is 3:1 on craps and 7:1 on 11 (yo). Another method of calculating the payout is to divide the total bet in half. The player would receive 7:1 minus half the total bet payout on half the total bet for craps and 15:1 minus half the total bet payout on half the total bet for 11 (yo). For example, using this method if a player were to bet $2 on C & E, $1 would receive 7:1 payout on craps minus $1 for the bet on 11 so the total profit would be $6. If an 11 was rolled the player would receive 15:1 minus $1 for the bet on craps so the player's total profit is $14. Both methods of calculation yield the same result so either method can be used. If a player wishes to take the bet down after a win the player would receive the whole bet not half even though only one of the two bets can win per roll. The minimum bet on C & E is double the lowest unit bet allowed at the table. So if the minimum single roll bet is $1 the lowest C & E bet allowed would be $2. Players are, however, able to make odd number bets larger than $2 if desired. One of the two bets will always lose, the other may win.
Any seven: Wins if the shooter rolls a 7 with 4:1 payout. This bet is also nicknamed Big Red, since the 7 on its betting space on the layout is usually large and red, and it is considered bad luck and a breach of etiquette to speak the word "seven" at the table.
Horn: This is a bet that involves betting on 1 unit each for 2, 3, 11 and 12 at the same time for the next roll. The bet is actually four separate bets, and pays off depending on which number is actually rolled. The combined payout is 27:4 for 2, 12 and 3:1 for 3, 11. Each individual bet has the same payout as a single bet on the specific numbers, 30:1 for 2 and 12 minus the other three bets, 15:1 for 3 and 11 minus the other three bets. If a player wins the bet he can take down all four bets instead of a single bet even though only one bet can win per roll. Many players, in order to eliminate the confusion of tossing four chips to the center of the table or having change made while bets are being placed, will make a five-unit Horn High bet, which is a four-way bet with the extra unit going to one specific number. For example, if one tosses a $5 chip into the center and says "horn high yo", you are placing four $1 bets on each of the horn numbers and the extra dollar will go on the yo (11). Horn bets are generally required to be in multiples of 4 or 5 with the minimum bet being 4 times the minimum unit allowed. For example, if the single roll minimum at the table is $1 the Horn bet must be $4 or more.
Whirl or World: A five-unit bet that is a combination of a horn and any-seven bet, with the idea that if a seven is rolled the bet is a push, because the money won on the seven is lost on the horn portions of the bet. The combine odds are 26:5 on the 2, 12, 11:5 on the 3, 11, and a push on the 7. Like the C & E and Horn bet, if a player wishes to take down the bet after a win he or she would receive all five units back. The minimum bet is five of the minimum units. For example, if the minimum single roll bet is $1, the minimum World/Whirl bet is $5.
On the Hop, Hop or Hopping: A single roll bet on any particular combination of the two dice on the next roll including combinations whose sum is 7 (e.g. 4 and 3). For example, if you bet on "5 and 1" on the hop, you are betting that the next roll will have a 5 on one die and a 1 on the other die. The bet pays 15:1 on easy ways (same as a bet on 3 or 11). Hard ways hop pays 30:1 (e.g., 3 and 3 on the hop, same as a bet on 2 or 12). The true odds are 17:1 and 35:1, resulting in a house edge of 11.11% and 13.89% respectively. When presented, hop bets are located at the center of the craps layout with the other proposition bets. If hop bets are not on the craps layout, they still may be bet on by players but they become the responsibility of the boxman to book the bet. Sometimes players may request to hop a whole number. In this case the money on the bet different combinations. For example, if a player says "hop the tens" (6–4, 5–5, 4–6) the player must give the dealer an even number bet so it can be divided among the hard and easy ways. If the player gives $10, $5 would be placed on the easy ways 10 with 15:1 odds and $5 would be placed on the hard way with 30:1 odds. If a player wishes to "hop the sevens" there would be three different combinations and six possible ways to roll a 7 (6–1, 5–2, 4–3, 3–4, 2–5, 1–6) therefore the player should bet in multiples of 3 so the bet can be divided among each combination with a 15:1 payout minus the other two bets, otherwise if players does not bet in multiples of 3, they would specific which combination has additional units.
Field: This bet is a wager that one of the numbers 2, 3, 4, 9, 10, 11, or 12 will appear on the next roll of the dice. This bet typically pays more (2:1 or 3:1) if 2 or 12 is rolled, and 1:1 if 3, 4, 9, 10 or 11 is rolled.
The Field bet is a "Self-Service" Bet. Unlike the other proposition bets which are handled by the dealers or stickman, the field bet is placed directly by the player.
Players identify their Field bets by placing them in the Field area directly in front of them or as close to their position as possible. The initial bet and/or any payouts can "ride" through several rolls until they lose, and are assumed to be "riding" by dealers. It is thus the player's responsibility to collect their bet and/or winnings immediately upon payout, before the next dice roll, if they do not wish to let it ride.
Fire Bet: Before the shooter begins, some casinos will allow a bet known as a fire bet to be placed. A fire bet is a bet of as little as $1 and generally up to a maximum of $5 to $10 sometimes higher, depending on casino, made in the hope that the next shooter will have a hot streak of setting and getting many points of different values. As different individual points are made by the shooter, they will be marked on the craps layout with a fire symbol.
The first three points will not pay out on the fire bet, but the fourth, fifth and sixth will pay out at increasing odds. The fourth point pays at 24-to-1, the fifth point pays at 249-to-1 and the 6th point pays at 999-to-1. Note that the points must all be different numbers for them to count towards the fire bet. For example, a shooter who successfully hits a point of 10 twice will only garner credit for the first one on the fire bet. Players must hit the established point in order for it to count toward the fire bet. The payout is determine by the number of points which have been established and hit after the shooter sevens out.
Bonus Craps: Prior to the initial "come out roll", players may place an optional wager (usually a $1 minimum to a maximum $25) on one or more of the three Bonus Craps wagers, "All Small", "All Tall", or "All or Nothing at All." For players to win the "All Small" wager, the shooter must hit all five small numbers (2, 3, 4, 5, 6) before a seven is rolled; similarly, "All Tall" wins if all five high numbers (8, 9, 10, 11, 12) are hit before a seven is rolled.
These bets pay 35-for-1, for a house advantage of 7.76%. “All or Nothing at All” wins if the shooter hits all 10 numbers before a seven is rolled. This pays 176-for-1, for a house edge of 7.46%. For all three wagers, the order in which the numbers are hit does not matter. Whenever a seven is hit, including on the come out roll, all bonus bets lose, the bonus board is reset, and new bonus bets may be placed.
A player may wish to make multiple different bets. For example, a player may be wish to bet $1 on all hard ways and the horn. If one of the bets win the dealer may automatically replenish the losing bet with profits from the winning bet. In this example, if the shooter rolls a hard 8 (pays 9:1), the horn loses. The dealer may return $5 to the player and place the other $4 on the horn bet which lost. If the player does not want the bet replenished, he or she should request any or all bets be taken down.
A working bet is a live bet. Bets may also be on the board, but not in play and therefore not working. Pass line and come bets are always working meaning the chips are in play and the player is therefore wagering live money. Other bets may be working or not working depending whether a point has been established or player's choice. Place and buy bets are working by default when a point is established and not working when the point is off unless the player specifies otherwise. Lay bets are always working even if a point has not been established unless the player requests otherwise. At any time, a player may wish to take any bet or bets out of play. The dealer will put an "Off" button on the player's specific bet or bets; this allows the player to keep his chips on the board without a live wager. For example, if a player decides not to wager a place bet mid-roll but wishes to keep the chips on the number, he or she may request the bet be "not working" or "Off". The chips remain on the table, but the player cannot win from or lose chips which are not working.
The opposite is also allowed. By default place and buy bets are not working without an established point; a player may wish to wager chips before a point has been established. In this case, the player would request the bet be working in which the dealer will place an "On" button on the specified chips.
The probability of dice combinations determine the odds of the payout. The following chart shows the dice combinations needed to roll each number. The two and twelve are the hardest to roll since only one combination of dice is possible. The game of craps is built around the dice roll of seven, since it is the most easily rolled dice combination.
Viewed another way:
The expected value of all bets is usually negative, such that the average player will always lose money. This is because the house always sets the paid odds to below the actual odds. The only exception is the "odds" bet that the player is allowed to make after a point is established on a pass/come don't pass/don't come bet (the odds portion of the bet has a long-term expected value of 0). However, this "free odds" bet cannot be made independently, so the expected value of the entire bet, including odds, is still negative. Since there is no correlation between die rolls, there is normally no possible long-term winning strategy in craps.
There are occasional promotional variants that provide either no house edge or even a player edge. One example is a field bet that pays 3:1 on 12 and 2:1 on either 3 or 11. Overall, given the 5:4 true odds of this bet, and the weighted average paid odds of approximately 7:5, the player has a 5% advantage on this bet. This is sometimes seen at casinos running limited-time incentives, in jurisdictions or gaming houses that require the game to be fair, or in layouts for use in informal settings using play money. No casino currently runs a craps table with a bet that yields a player edge full-time.
Maximizing the size of the odds bet in relation to the line bet will reduce, but never eliminate the house edge, and will increase variance. Most casinos have a limit on how large the odds bet can be in relation to the line bet, with single, double, and five times odds common. Some casinos offer 3–4–5 odds, referring to the maximum multiple of the line bet a player can place in odds for the points of 4 and 10, 5 and 9, and 6 and 8, respectively. During promotional periods, a casino may even offer 100x odds bets, which reduces the house edge to almost nothing, but dramatically increases variance, as the player will be betting in large betting units.
Since several of the multiple roll bets pay off in ratios of fractions on the dollar, it is important that the player bets in multiples that will allow a correct payoff in complete dollars. Normally, payoffs will be rounded down to the nearest dollar, resulting in a higher house advantage. These bets include all place bets, taking odds, and buying on numbers 6, 8, 5, and 9, as well as laying all numbers.
These variants depend on the casino and the table, and sometimes a casino will have different tables that use or omit these variants and others.
When craps is played in a casino, all bets have a house advantage. That is, it can be shown mathematically that a player will (with 100% probability) lose all his or her money to the casino in the long run, while in the short run the player is more likely to lose money than make money. There may be players who are lucky and get ahead for a period of time, but in the long run these winning streaks are eroded away. One can slow, but not eliminate, one's average losses by only placing bets with the smallest house advantage.
The pass/don't pass line, come/don't come line, place 6, place 8, buy 4 and buy 10 (only under the casino rules where commission is charged only on wins) have the lowest house edge in the casino, and all other bets will, on average, lose money between three and twelve times faster because of the difference in house edges.
The place bets and buy bets differ from the pass line and come line, in that place bets and buy bets can be removed at any time, since, while they are multi-roll bets, their odds of winning do not change from roll to roll, whereas pass line bets and come line bets are a combination of different odds on their first roll and subsequent rolls. The first roll of a pass line bet is 2:1 advantage for the player (8 wins, 4 losses), but it's "paid for" by subsequent rolls that are at the same disadvantage to the player as the don't pass bets were at an advantage. As such, they cannot profitably let you take down the bet after the first roll. Players can bet or lay odds behind an established point depending on whether it was a Pass/Come or Don't Pass/Don't Come to lower house edge by receiving true odds on the point. Casinos which allow put betting allows players to increase or make new pass/come bets after the come-out roll. This bet generally has a higher house edge than place betting, unless the casino offers high odds.
Conversely, you can take back (pick up) a don't pass or don't come bet after the first roll, but this cannot be recommended, because you already endured the disadvantaged part of the combination – the first roll. On that come-out roll, you win just 3 times (2 and 3), while losing 8 of them (7 and 11) and pushing one (12) out of the 36 possible rolls. On the other 24 rolls that become a point, your don't pass bet is now to your advantage by 6:3 (4 and 10), 6:4 (5 and 9) and 6:5 (6 and 8). If a player chooses to remove the initial don't come and/or don't pass line bet, he or she can no longer lay odds behind the bet and cannot re-bet the same don't pass and/or don't come number (players must make a new don't pass or come bets if desired). However, players can still make standard lay bets odds on any of the point numbers (4,5,6,8,9,10).
Among these, and the remaining numbers and possible bets, there are a myriad of systems and progressions that can be used with many combinations of numbers.
An important alternative metric is house advantage per roll (rather than per bet), which may be expressed in loss per hour. The typical pace of rolls varies depending on the number of players, but 102 rolls per hour is a cited rate for a nearly full table. This same reference states that only "29.6% of total rolls are come out rolls, on average", so for this alternative metric, needing extra rolls to resolve the pass line bet, for example, is factored. This number then permits calculation of rate of loss per hour, and per the 4 day/5 hour per day gambling trip:
Besides the rules of the game itself, a number of formal and informal rules are commonly applied in the table form of Craps, especially when played in a casino.
To reduce the potential opportunity for switching dice by sleight-of-hand, players are not supposed to handle the dice with more than one hand (such as shaking them in cupped hands before rolling) nor take the dice past the edge of the table. If a player wishes to change shooting hands, they may set the dice on the table, let go, then take them with the other hand.
When throwing the dice, the player is expected to hit the farthest wall at the opposite end of the table (these walls are typically augmented with pyramidal structures to ensure highly unpredictable bouncing after impact). Casinos will sometimes allow a roll that does not hit the opposite wall as long as the dice are thrown past the middle of the table; a very short roll will be nullified as a "no roll". The dice may not be slid across the table and must be tossed. These rules are intended to prevent dexterous players from physically influencing the outcome of the roll.
Players are generally asked not to throw the dice above a certain height (such as the eye level of the dealers). This is both for the safety of those around the table, and to eliminate the potential use of such a throw as a distraction device in order to cheat.
Dice are still considered "in play" if they land on players' bets on the table, the dealer's working stacks, on the marker puck, or with one die resting on top of the other. The roll is invalid if either or both dice land in the boxman's bank, the stickman's bowl (where the extra three dice are kept between rolls), or in the rails around the top of the table where players chips are kept. If one or both dice hits a player or dealer and rolls back onto the table, the roll counts as long as the person being hit did not intentionally interfere with either of the dice, though some casinos will rule "no roll" for this situation. If one or both leave the table, it is also a "no roll", and the dice may either be replaced or examined by the boxman and returned to play.
Shooters may wish to "set" the dice to a particular starting configuration before throwing (such as showing a particular number or combination, stacking the dice, or spacing them to be picked up between different fingers), but if they do, they are often asked to be quick about it so as not to delay the game. Some casinos disallow such rituals to speed up the pace of the game. Some may also discourage or disallow unsanitary practices such as kissing or spitting on the dice.
In most casinos, players are not allowed to hand anything directly to dealers, and vice versa. Items such as cash, checks, and chips are exchanged by laying them down on the table; for example, when "buying in" (paying cash for chips), players are expected to place the cash on the layout: the dealer will take it and then place the chips in front of the player. This rule is enforced in order to allow the casino to easily monitor and record all transfers via overhead surveillance cameras, and to reduce the opportunity for cheating via sleight-of-hand.
Most casinos prohibit "call bets", and may have a warning such as "No Call Bets" printed on the layout to make this clear. This means a player may not call out a bet without also placing the corresponding chips on the table. Such a rule reduces the potential for misunderstanding in loud environments, as well as disputes over the amount that the player intended to bet after the outcome has been decided. Some casinos choose to allow call bets once players have bought-in. When allowed, they are usually made when a player wishes to bet at the last second, immediately before the dice are thrown, to avoid the risk of obstructing the roll.
Craps is among the most social and most superstitious of all gambling games, which leads to an enormous variety of informal rules of etiquette that players may be expected to follow. An exhaustive list of these is beyond the scope of this article, but the guidelines below are most commonly given.
Tipping the dealers is universal and expected in Craps. As in most other casino games, a player may simply place (or toss) chips onto the table and say, "For the dealers", "For the crew", "etc." In craps, it is also common to place a bet for the dealers. This is usually done one of three ways: by placing an ordinary bet and simply declaring it for the dealers, as a "two-way", or "on top". A "Two-Way" is a bet for both parties: for example, a player may toss in two chips and say "Two Way Hard Eight", which will be understood to mean one chip for the player and one chip for the dealers. Players may also place a stack of chips for a bet as usual, but leave the top chip off-center and announce "on top for the dealers". The dealer's portion is often called a "toke" bet, which comes from the practice of using $1 slot machine tokens to place dealer bets in some casinos.
In some cases, players may also tip each other, for example as a show of gratitude to the thrower for a roll on which they win a substantial bet.
Craps players routinely practice a wide range of superstitious behaviors, and may expect or demand these from other players as well.
Most prominently, it is universally considered bad luck to say the word "seven" (after the "come-out", a roll of 7 is a loss for "pass" bets). Dealers themselves often make significant efforts to avoid calling out the number. When necessary, participants may refer to seven with a "nickname" such as "Big Red" (or just "Red"), "the S-word", etc.
Although no wagering system can consistently beat casino games based on independent trials such as craps, that does not stop gamblers from believing in them. One of the best known systems is the Martingale System. In this strategy, the gambler doubles his bet after every loss. After a win, the bet is reset to the original bet. The theory is that the first win would recover all previous losses plus win a profit equal to the original stake.
Other systems depend on the gambler's fallacy, which in craps terms is the belief that past dice rolls influence the probabilities of future dice rolls. For example, the gambler's fallacy indicates that a craps player should bet on eleven if an eleven has not appeared or has appeared too often in the last 20 rolls. In practice this can be observed as players respond to a roll such as a Hard Six with an immediate wager on the Hard Six.
In reality, each roll of the dice is an independent event, so the probability of rolling eleven is exactly 1/18 on every roll, regardless of the number of times eleven has come up in the last x rolls. Even if the dice are actually biased toward particular results ("loaded"), each roll is still independent of all the previous ones. The common term to describe this is "dice have no memory".
Another approach is to "set" the dice in a particular orientation, and then throw them in such a manner that they do not tumble randomly. The theory is that given exactly the same throw from exactly the same starting configuration, the dice will tumble in the same way and therefore show the same or similar values every time.
Casinos take steps to prevent this. The dice are usually required to hit the back wall of the table, which is normally faced with an angular texture such as pyramids, making controlled spins more difficult. There has been no independent evidence that such methods can be successfully applied in a real casino.
Bank craps is a variation of the original craps game and is sometimes known as Las Vegas Craps. This variant is quite popular in Nevada gambling houses, and its availability online has now made it a globally played game. Bank craps uses a special table layout and all bets must be made against the house. In Bank Craps, the dice are thrown over a wire or a string that is normally stretched a few inches from the table's surface. The lowest house edge (for the pass/don't pass) in this variation is around 1.4%. Generally, if the word "craps" is used without any modifier, it can be inferred to mean this version of the game, to which most of this article refers.
Crapless craps, also known as Bastard Craps, is a simple version of the original craps game, and is normally played as an online private game. The biggest difference between crapless craps and original craps, is that the shooter (person throwing the dice) is at a far greater disadvantage and has a house edge of 5.38%. Another difference is that this is one of the craps games in which a player can bet on rolling a 2, 3, 11 or 12 before a 7 is thrown. In crapless craps, 2 and 12 have odds of 11:2 and have a house edge of 7.143% while 3 and 11 have odds of 11:4 with a house edge of 6.25%.
New York Craps is one of the variations of craps played mostly in the Eastern coast of the US, true to its name. History states that this game was actually found and played in casinos in Yugoslavia, the UK and the Bahamas. In this craps variant, the house edge is greater than Las Vegas Craps or Bank craps. The table layout is also different, and is called a double-end-dealer table. This variation is different from the original craps game in several ways, but the primary difference is that New York craps doesn't allow Come or Don't Come bets. New York Craps Players bet on box numbers like 4, 5, 6, 8, 9 or 10. The overall house edge in New York craps is 5%.
In order to get around California laws barring the payout of a game being directly related to the roll of dice, Indian reservations have adapted the game to substitute cards for dice.
In one variation, there are no dice at all. Two shoes are used, each containing some number of regular card decks that have been stripped down to just the Aces and deuces through sixes. The boxman simply deals one card from each shoe and that is the roll on which bets are settled. Since a card-counting scheme is easily devised to make use of the information of cards that have already been dealt, a relatively small portion (less than 50%) of each shoe is usually dealt in order to protect the house.
In a similar variation, cards representing dice are dealt directly from a continuous shuffling machine (CSM). Typically, the CSM will hold approximately 264 cards, or 44 sets of 1 through 6 spot cards. Two cards are dealt from the CSM for each roll. The game is played exactly as regular craps, but the roll distribution of the remaining cards in the CSM is slightly skewed from the normal symmetric distribution of dice.
Even if the dealer were to shuffle each roll back into the CSM, the effect of buffering a number of cards in the chute of the CSM provides information about the skew of the next roll. Analysis shows this type of game is biased towards the don't pass and don't come bets. A player betting don't pass and don't come every roll and laying 10x odds receives a 2% profit on the initial don't pass / don't come bet each roll. Using a counting system allows the player to attain a similar return at lower variance.
To replicate the original dice odds exactly without dice or possibility of card-counting, another scheme uses two shuffle machines with just one deck of Ace through 6 each. Each machine selects one of the 6 cards at random and this is the roll. The selected cards are replaced and the decks are reshuffled for the next roll.
In this game variation, one red deck and one blue deck of six cards each (A through 6), and a red die and a blue die are used. Each deck is shuffled separately, usually by machine. Each card is then dealt onto the layout, into the 6 red and 6 blue numbered boxes. The shooter then shoots the dice. The red card in the red-numbered box corresponding to the red die, and the blue card in the blue-numbered box corresponding to the blue die are then turned over to form the roll on which bets are settled.
Another variation uses a red and a blue deck of 36 custom playing cards each. Each card has a picture of a two-die roll on it – from 1–1 to 6–6. The shooter shoots what looks like a red and a blue die, called "cubes". They are numbered such that they can never throw a pair, and that the blue one will show a higher value than the red one exactly half the time. One such scheme could be 222555 on the red die and 333444 on the blue die.
One card is dealt from the red deck and one is dealt from the blue deck. The shooter throws the "cubes" and the color of the cube that is higher selects the color of the card to be used to settle bets. On one such table, an additional one-roll prop bet was offered: If the card that was turned over for the "roll" was either 1–1 or 6–6, the other card was also turned over. If the other card was the "opposite" (6–6 or 1–1, respectively) of the first card, the bet paid 500:1 for this 647:1 proposition.
And additional variation uses a single set of 6 cards, and regular dice. The roll of the dice maps to the card in that position, and if a pair is rolled, then the mapped card is used twice, as a pair.
Recreational or informal playing of craps outside of a casino is referred to as street craps or private craps. The most notable difference between playing street craps and bank craps is that there is no bank or house to cover bets in street craps. Players must bet against each other by covering or fading each other's bets for the game to be played. If money is used instead of chips and depending on the laws of where it is being played, street craps can be an illegal form of gambling.
There are many variations of street craps. The simplest way is to either agree on or roll a number as the point, then roll the point again before you roll a seven. Unlike more complex proposition bets offered by casinos, street craps has more simplified betting options. The shooter is required to make either a Pass or a Don't Pass bet if he wants to roll the dice. Another player must choose to cover the shooter to create a stake for the game to continue.
If there are several players, the rotation of the player who must cover the shooter may change with the shooter (comparable to a blind in poker). The person covering the shooter will always bet against the shooter. For example, if the shooter made a "Pass" bet, the person covering the shooter would make a "Don't Pass" bet to win. Once the shooter is covered, other players may make Pass/Don't Pass bets, or any other proposition bets, as long as there is another player willing to cover.
Due to the random nature of the game, in popular culture a "crapshoot" is often used to describe an action with an unpredictable outcome.
The prayer or invocation "Baby needs a new pair of shoes!" is associated with shooting craps.
Floating craps is an illegal operation of craps. The term "floating" refers to the practice of the game's operators using portable tables and equipment to quickly move the game from location to location to stay ahead of the law enforcement authorities. The term may have originated in the 1930s when Benny Binion (later known for founding the downtown Las Vegas hotel Binions) set up an illegal craps game utilizing tables created from portable crates for the Texas Centennial Exposition.
The 1950 Broadway musical "Guys and Dolls" features a major plot point revolving around a floating craps game.
In the 1950s and 1960s The Sands Hotel in Las Vegas had a craps table that floated in the swimming pool, as a joke reference to the notoriety of the term.
A Golden Arm is a craps player who rolls the dice for longer than one hour without losing. The first Golden Arm was Oahu native, Stanley Fujitake, who rolled 118 times without sevening out in 3 hours and 6 minutes at the California Hotel and Casino on May 28, 1989.
The current record for length of a "hand" (successive rounds won by the same shooter) is 154 rolls including 25 passes by Patricia DeMauro of New Jersey, lasting 4 hours and 18 minutes, at the Borgata in Atlantic City, New Jersey, on May 23–24, 2009. She bested by over an hour the record held for almost 20 years – that of Fujitake. | https://en.wikipedia.org/wiki?curid=6062 |
Carl von Clausewitz
Carl Philipp Gottfried (or Gottlieb) von Clausewitz (; 1 June 1780 – 16 November 1831) was a Prussian general and military theorist who stressed the "moral" (meaning, in modern terms, psychological) and political aspects of war. His most notable work, "Vom Kriege" ("On War"), was unfinished at his death.
Clausewitz was a realist in many different senses and, while in some respects a romantic, also drew heavily on the rationalist ideas of the European Enlightenment.
Clausewitz's thinking is often described as Hegelian because of his dialectical method; but, although he was probably personally acquainted with Hegel, there remains debate as to whether or not Clausewitz was in fact influenced by him. He stressed the dialectical interaction of diverse factors, noting how unexpected developments unfolding under the "fog of war" (i.e., in the face of incomplete, dubious, and often completely erroneous information and high levels of fear, doubt, and excitement) call for rapid decisions by alert commanders. He saw history as a vital check on erudite abstractions that did not accord with experience. In contrast to the early work of Antoine-Henri Jomini, he argued that war could not be quantified or reduced to mapwork, geometry, and graphs. Clausewitz had many aphorisms, of which the most famous is "War is the continuation of politics by other means."
Clausewitz's Christian names are sometimes given in non-German sources as "Karl," "Carl Philipp Gottlieb," or "Carl Maria." He spelled his own given name with a "C" in order to identify with the classical Western tradition; writers who use "Karl" are often seeking to emphasize his German (rather than European) identity. "Carl Philipp Gottfried" appears on Clausewitz's tombstone. Nonetheless, sources such as military historian Peter Paret and "Encyclopædia Britannica" continue to use Gottlieb instead of Gottfried.
Clausewitz was born on 1 June 1780 in Burg bei Magdeburg in the Prussian Duchy of Magdeburg as the fourth and youngest son of a family that made claims to noble status which Carl accepted. Clausewitz's family claimed descent from the Barons of Clausewitz in Upper Silesia, though scholars question the connection. His grandfather, the son of a Lutheran pastor, had been a professor of theology. Clausewitz's father, once a lieutenant in the army of Frederick the Great, King of Prussia, held a minor post in the Prussian internal-revenue service. Clausewitz entered the Prussian military service at the age of twelve as a lance-corporal, eventually attaining the rank of major general.
Clausewitz served in the Rhine Campaigns (1793–1794) including the Siege of Mainz, when the Prussian army invaded France during the French Revolution, and fought in the Napoleonic Wars from 1806 to 1815. He entered the "Kriegsakademie" (also cited as "The German War School", the "Military Academy in Berlin", and the "Prussian Military Academy," later the "War College") in Berlin in 1801 (aged 21), probably studied the writings of the philosophers Immanuel Kant and/or Fichte and Schleiermacher and won the regard of General Gerhard von Scharnhorst, the future first chief-of-staff of the newly reformed Prussian Army (appointed 1809). Clausewitz, Hermann von Boyen (1771–1848) and Karl von Grolman (1777–1843) were among Scharnhorst's primary allies in his efforts to reform the Prussian army between 1807 and 1814.
Clausewitz served during the Jena Campaign as aide-de-camp to Prince August. At the Battle of Jena-Auerstedt on 14 October 1806—when Napoleon invaded Prussia and defeated the massed Prussian-Saxon army commanded by Karl Wilhelm Ferdinand, Duke of Brunswick—he was captured, one of the 25,000 prisoners taken that day as the Prussian army disintegrated. He was 26. Clausewitz was held prisoner with his prince in France from 1807 to 1808. Returning to Prussia, he assisted in the reform of the Prussian army and state.
On 10 December 1810 he married the socially prominent Countess Marie von Brühl, whom he had first met in 1803. She was a member of the noble German von Brühl family originating in Thuringia. The couple moved in the highest circles, socialising with Berlin's political, literary and intellectual élite. Marie was well-educated and politically well-connected—she played an important role in her husband's career progress and intellectual evolution. She also edited, published, and introduced his collected works.
Opposed to Prussia's enforced alliance with Napoleon I, Clausewitz left the Prussian army and served in the Imperial Russian Army from 1812 to 1813 during the Russian Campaign, taking part in the Battle of Borodino (1812). Like many Prussian officers serving in Russia, he joined the Russian-German Legion in 1813. In the service of the Russian Empire, Clausewitz helped negotiate the Convention of Tauroggen (1812), which prepared the way for the coalition of Prussia, Russia, and the United Kingdom that ultimately defeated Napoleon and his allies.
In 1815 the Russian-German Legion became integrated into the Prussian Army and Clausewitz re-entered Prussian service as a colonel. He was soon appointed chief-of-staff of Johann von Thielmann's III Corps. In that capacity he served at the Battle of Ligny and the Battle of Wavre during the Waterloo Campaign in 1815. An army led personally by Napoleon defeated the Prussians at Ligny (south of Mont-Saint-Jean and the village of Waterloo) on 16 June 1815, but they withdrew in good order. Napoleon's failure to destroy the Prussian forces led to his defeat a few days later at the Battle of Waterloo (18 June 1815), when the Prussian forces arrived on his right flank late in the afternoon to support the Anglo-Dutch-Belgian forces pressing his front. Napoleon had convinced his troops that the field grey uniforms were those of Marshall Grouchy's Granadiers. Clausewitz's unit fought heavily outnumbered at Wavre (18–19 June 1815), preventing large reinforcements from reaching Napoleon at Waterloo. After the war Clausewitz served as the director of the "Kriegsakademie", where he served until 1830. In that year he returned to active duty with the army. Soon afterwards, the outbreak of several revolutions around Europe and a crisis in Poland appeared to presage another major European war. Clausewitz was appointed chief of staff of the only army Prussia was able to mobilise in this emergency, which was sent to the Polish border. Its commander, Gneisenau, died of cholera (August 1831), and Clausewitz took command of the Prussian army's efforts to construct a "cordon sanitaire" to contain the great cholera outbreak (the first time cholera had appeared in modern heartland Europe, causing a continent-wide panic). Clausewitz himself died of the same disease shortly afterwards, on 17 November 1831.
His widow edited, published, and wrote the introduction to his "magnum opus" on the philosophy of war in 1832. (He had started working on the text in 1816, but had not completed it.) She wrote the preface for "On War" and by 1835 had published most of his collected works. She died in January 1836.
Clausewitz was a professional combat soldier who was involved in numerous military campaigns, but he is famous primarily as a military theorist interested in the examination of war, utilising the campaigns of Frederick the Great and Napoleon as frames of reference for his work. He wrote a careful, systematic, philosophical examination of war in all its aspects. The result was his principal book, "On War," a major work on the philosophy of war. It was unfinished when Clausewitz died and contains material written at different stages in his intellectual evolution, producing some significant contradictions between different sections. The sequence and precise character of that evolution is a source of much debate, as are exact meaning behind his seemingly contradictory claims (discussions pertinent to the tactical, operational and strategic levels of war are one example). Clausewitz constantly sought to revise the text, particularly between 1827 and his departure on his last field assignments, to include more material on "people's war" and forms of war other than high-intensity warfare between states, but relatively little of this material was included in the book. Soldiers before this time had written treatises on various military subjects, but none had undertaken a great philosophical examination of war on the scale of those written by Clausewitz and Leo Tolstoy, both of which were inspired by the events of the Napoleonic Era.
Clausewitz's work is still studied today, demonstrating its continued relevance. More than sixteen major English-language books that focused specifically on his work were published between 2005 and 2014, whereas his 19th-century rival Jomini faded from influence. The historian Lynn Montross said the outcome, "may be explained by the fact that Jomini produced a system of war, Clausewitz a philosophy. The one has been outdated by new weapons, the other still influences the strategy behind those weapons." Jomini did not attempt to define war but Clausewitz did, providing (and dialectically comparing) a number of definitions. The first is his dialectical thesis: "War is thus an act of force to compel our enemy to do our will." The second, often treated as Clausewitz's 'bottom line,' is in fact merely his dialectical antithesis: "War is merely the continuation of politics with other means." The synthesis of his dialectical examination of the nature of war is his famous "trinity," saying that war is "a fascinating trinity—composed of primordial violence, hatred, and enmity, which are to be regarded as a blind natural force; the play of chance and probability, within which the creative spirit is free to roam; and its element of subordination, as an instrument of policy, which makes it subject to pure reason." Christopher Bassford says the best shorthand for Clausewitz's trinity should be something like "violent emotion/chance/rational calculation." However, it is frequently presented as "people/army/government," a misunderstanding based on a later paragraph in the same chapter. This misrepresentation was popularised by U.S. Army Colonel Harry Summers' Vietnam-era interpretation, facilitated by weaknesses in the 1976 Howard/Paret translation.
The degree to which Clausewitz managed to revise his manuscript to reflect that synthesis is the subject of much debate. His final reference to war and "Politik", however, goes beyond his widely quoted antithesis: "War is simply the continuation of political intercourse with the addition of other means. We deliberately use the phrase 'with the addition of other means' because we also want to make it clear that war in itself does not suspend political intercourse or change it into something entirely different. In essentials that intercourse continues, irrespective of the means it employs. The main lines along which military events progress, and to which they are restricted, are political lines that continue throughout the war into the subsequent peace."
Clausewitz introduced systematic philosophical contemplation into Western military thinking, with powerful implications not only for historical and analytical writing but also for practical policy, military instruction, and operational planning. He relied on his own experiences, contemporary writings about Napoleon, and on deep historical research. His historiographical approach is evident in his first extended study, written when he was 25, of the Thirty Years War. He rejects the Enlightenment's view of the war as a chaotic muddle and instead explains its drawn-out operations by the economy and technology of the age, the social characteristics of the troops, and the commanders' politics and psychology. In "On War", Clausewitz sees all wars as the sum of decisions, actions, and reactions in an uncertain and dangerous context, and also as a socio-political phenomenon. He also stressed the complex nature of war, which encompasses both the socio-political and the operational and stresses the primacy of state policy.
The word "strategy" had only recently come into usage in modern Europe, and Clausewitz's definition is quite narrow: "the use of engagements for the object of war." Clausewitz conceived of war as a political, social, and military phenomenon which might — depending on circumstances — involve the entire population of a nation at war. In any case, Clausewitz saw military force as an instrument that states and other political actors use to pursue the ends of policy, in a dialectic between opposing wills, each with the aim of imposing his policies and will upon his enemy.
Clausewitz's emphasis on the inherent superiority of the defence suggests that habitual aggressors are likely to end up as failures. The inherent superiority of the defence obviously does not mean that the defender will always win, however: there are other asymmetries to be considered. He was interested in co-operation between the regular army and militia or partisan forces, or citizen soldiers, as one possible — sometimes the only — method of defence. In the circumstances of the Wars of the French Revolution and with Napoleon, which were energised by a rising spirit of nationalism, he emphasised the need for states to involve their entire populations in the conduct of war. This point is especially important, as these wars demonstrated that such energies could be of decisive importance and for a time led to a democratisation of the armed forces much as universal suffrage democratised politics.
While Clausewitz was intensely aware of the value of intelligence at all levels, he was also very sceptical of the accuracy of much military intelligence: "Many intelligence reports in war are contradictory; even more are false, and most are uncertain... In short, most intelligence is false." This circumstance is generally described as part of the fog of war. Such sceptical comments apply only to intelligence at the tactical and operational levels; at the strategic and political levels he constantly stressed the requirement for the best possible understanding of what today would be called strategic and political intelligence. His conclusions were influenced by his experiences in the Prussian Army, which was often in an intelligence fog due partly to the superior abilities of Napoleon's system but even more to the nature of war. Clausewitz acknowledges that friction creates enormous difficulties for the realisation of any plan, and the "fog of war" hinders commanders from knowing what is happening. It is precisely in the context of this challenge that he develops the concept of military genius, whose capabilities are seen above all in the execution of operations. 'Military genius' is not simply a matter of intellect, but a combination of qualities of intellect, experience, personality, and temperament (and there are many possible such combinations) that create a very highly developed mental aptitude for the waging of war.
Key ideas discussed in "On War" include:
Clausewitz used a dialectical method to construct his argument, leading to frequent misinterpretation of his ideas. British military theorist B. H. Liddell Hart contends that the enthusiastic acceptance by the Prussian military establishment – especially Moltke the Elder, a former student of his – of what they believed to be Clausewitz's ideas, and the subsequent widespread adoption of the Prussian military system worldwide, had a deleterious effect on military theory and practice, due to their egregious misinterpretation of his ideas:
As so often happens, Clausewitz's disciples carried his teaching to an extreme which their master had not intended... [Clausewitz's] theory of war was expounded in a way too abstract and involved for ordinary soldier-minds, essentially concrete, to follow the course of his argument – which often turned back from the direction in which it was apparently leading. Impressed yet befogged, they grasped at his vivid leading phrases, seeing only their surface meaning, and missing the deeper current of his thought.
As described by Christopher Bassford, then-professor of strategy at the National War College of the United States:
One of the main sources of confusion about Clausewitz's approach lies in his dialectical method of presentation. For example, Clausewitz's famous line that "War is a mere continuation of politics by other means," ("Der Krieg ist eine bloße Fortsetzung der Politik mit anderen Mitteln") while accurate as far as it goes, was not intended as a statement of fact. It is the antithesis in a dialectical argument whose thesis is the point – made earlier in the analysis – that "war is nothing but a duel [or wrestling match, a better translation of the German "Zweikampf"] on a larger scale." His synthesis, which resolves the deficiencies of these two bold statements, says that war is neither "nothing but" an act of brute force nor "merely" a rational act of politics or policy. This synthesis lies in his "fascinating trinity" [wunderliche Dreifaltigkeit]: a dynamic, inherently unstable interaction of the forces of violent emotion, chance, and rational calculation.
Another example of this confusion is the idea that Clausewitz was a proponent of total war as used in the Third Reich's propaganda in the 1940s. In fact, he never used the term "total war": rather, he discussed "absolute war," a concept which evolved into the much more abstract notion of "ideal war" discussed at the very beginning of "Vom Kriege"—the purely "logical" result of the forces underlying a "pure," Platonic "ideal" of war. In what he called a "logical fantasy," war cannot be waged in a limited way: the rules of competition will force participants to use all means at their disposal to achieve victory. But in the "real world", he said, such rigid logic is unrealistic and dangerous. As a practical matter, the military objectives in "real" war that support political objectives generally fall into two broad types: "war to achieve limited aims"; and war to "disarm" the enemy, "to render [him] politically helpless or militarily impotent." Thus the complete defeat of the enemy may not be necessary, desirable, or even possible.
In modern times the reconstruction of Clausewitzian theory has been a matter of much dispute. One analysis was that of Panagiotis Kondylis, a Greek writer and philosopher, who opposed the interpretations of Raymond Aron in "Penser la Guerre, Clausewitz", and other liberal writers. According to Aron, Clausewitz was one of the first writers to condemn the militarism of the Prussian general staff and its war-proneness, based on Clausewitz's argument that "war is a continuation of politics by other means." In "Theory of War," Kondylis claims that this is inconsistent with Clausewitzian thought. He claims that Clausewitz was morally indifferent to war (though this probably reflects a lack of familiarity with personal letters from Clausewitz, which demonstrate an acute awareness of war's tragic aspects) and that his advice regarding politics' dominance over the conduct of war has nothing to do with pacifist ideas. For Clausewitz, war is simply one unique means that is sometimes applied to the eternal quest for power, of "raison d'État" in an anarchic and unsafe world.
Other notable writers who have studied Clausewitz's texts and translated them into English are historians Peter Paret of the Institute for Advanced Study and Sir Michael Howard, and the philosopher, musician, and game theorist Anatol Rapoport. Howard and Paret edited the most widely used edition of "On War" (Princeton University Press, 1976/1984) and have produced comparative studies of Clausewitz and other theorists, such as Tolstoy. Bernard Brodie's "A Guide to the Reading of "On War"", in the 1976 Princeton translation, expressed his interpretations of the Prussian's theories and provided students with an influential synopsis of this vital work.
The British military historian John Keegan attacked Clausewitz's theory in his book "A History of Warfare". Keegan argued that Clausewitz assumed the existence of states, yet 'war antedates the state, diplomacy and strategy by many millennia.'
Clausewitz died without completing "On War", but despite this his ideas have been widely influential in military theory and have had a strong influence on German military thought specifically. Later Prussian and German generals, such as Helmuth Graf von Moltke, were clearly influenced by Clausewitz: Moltke's widely quoted statement that "No campaign plan survives first contact with the enemy" is a classic reflection of Clausewitz's insistence on the roles of chance, friction, "fog," uncertainty, and interactivity in war.
Clausewitz's influence spread to British thinking as well, though at first more as a historian and analyst than as a theorist. See for example Wellington's extended essay discussing Clausewitz's study of the Campaign of 1815—Wellington's only serious written discussion of the battle, which was widely discussed in 19th-century Britain. Clausewitz's broader thinking came to the fore following Britain's military embarrassments in the Boer War (1899–1902). One example of a heavy Clausewitzian influence in that era is Spenser Wilkinson, journalist, the first Chichele Professor of Military History at Oxford University, and perhaps the most prominent military analyst in Britain from c. 1885 until well into the interwar period. Another is naval historian Julian Corbett (1854–1922), whose work reflected a deep if idiosyncratic adherence to Clausewitz's concepts and frequently an emphasis on Clausewitz's ideas about 'limited war' and the inherent strengths of the defensive form of war. Corbett's practical strategic views were often in prominent public conflict with Wilkinson's – see, for example, Wilkinson's article "Strategy at Sea," "The Morning Post", 12 February 1912. Following the First World War, however, the influential British military commentator B. H. Liddell Hart in the 1920s erroneously attributed to him the doctrine of "total war" that during the First World War had been embraced by many European general staffs and emulated by the British. More recent scholars typically see that war as so confused in terms of political rationale that it in fact contradicts much of "On War." One of the most influential British Clausewitzians today is Colin S. Gray; historian Hew Strachan (like Wilkinson also the Chichele Professor of Military History at Oxford University, since 2001) has been an energetic proponent of the "study" of Clausewitz, but his own views on Clausewitz's ideas are somewhat ambivalent.
With some interesting exceptions (e.g., John McAuley Palmer, Robert M. Johnston, Hoffman Nickerson), Clausewitz had little influence on American military thought before 1945 other than via British writers, though Generals Eisenhower and Patton were avid readers. He did influence Karl Marx, Friedrich Engels, Vladimir Lenin, Leon Trotsky and Mao Zedong, and thus the Communist Soviet and Chinese traditions, as Lenin emphasised the inevitability of wars among capitalist states in the age of imperialism and presented the armed struggle of the working class as the only path toward the eventual elimination of war. Because Lenin was an admirer of Clausewitz and called him "one of the great military writers", his influence on the Red Army was immense. The Russian historian A.N. Mertsalov commented that "It was an irony of fate that the view in the USSR was that it was Lenin who shaped the attitude towards Clausewitz, and that Lenin's dictum that war is a continuation of politics is taken from the work of this [allegedly] anti-humanist anti-revolutionary." The American mathematician Anatol Rapoport wrote in 1968 that Clausewitz as interpreted by Lenin formed the basis of all Soviet military thinking since 1917, and quoted the remarks by Marshal V.D. Sokolovsky:
In describing the essence of war, Marxism-Leninism takes as its point of departure the premise that war is not an aim in itself, but rather a tool of politics. In his remarks on Clausewitz's "On War", Lenin stressed that "Politics is the reason, and war is only the tool, not the other way around. Consequently, it remains only to subordinate the military point of view to the political".
Henry A. Kissinger, however, described Lenin's approach as being that politics is a continuation of war by other means, thus turning Clausewitz's argument "on its head."
Rapoport argued that:
As for Lenin's approval of Clausewitz, it probably stems from his obsession with the struggle for power. The whole Marxist conception of history is that of successive struggles for power, primarily between social classes. This was constantly applied by Lenin in a variety of contexts. Thus the entire history of philosophy appears in Lenin's writings as a vast struggle between "idealism" and "materialism". The fate of the socialist movement was to be decided by a struggle between the revolutionists and the reformers. Clausewitz's acceptance of the struggle for power as the essence of international politics must have impressed Lenin as starkly realistic.
Clausewitz directly influenced Mao Zedong, who read "On War" in 1938 and organised a seminar on Clausewitz for the Party leadership in Yan'an. Thus the "Clausewitzian" content in many of Mao's writings is not merely a regurgitation of Lenin but reflects Mao's study.
The idea that war involves inherent "friction" that distorts, to a greater or lesser degree, all prior arrangements, has become common currency in fields such as business strategy and sport. The phrase "fog of war" derives from Clausewitz's stress on how confused warfare can seem while immersed within it. The term center of gravity, used in a military context derives from Clausewitz's usage, which he took from Newtonian mechanics. In U.S. military doctrine, "center of gravity" refers to the basis of an opponent's power at the operational, strategic, or political level, though this is only one aspect of Clausewitz's use of the term.
The deterrence strategy of the United States in the 1950s was closely inspired by President Dwight Eisenhower’s reading of Clausewitz as a young officer in the 1920s. Eisenhower was greatly impressed by Clausewitz’s example of a theoretical, idealised “absolute war” in "Vom Krieg" as a way of demonstrating how absurd it would be to attempt such a strategy in practice. For Eisenhower, the age of nuclear weapons had made what was for Clausewitz in the early 19th century only a theoretical vision an all too real possibility in the mid-20th century. From Eisenhower's viewpoint, the best deterrent to war was to show the world just how appalling and horrific a nuclear “absolute war” would be if it should ever occur, hence a series of much publicised nuclear tests in the Pacific, giving first priority in the defence budget to nuclear weapons and delivery systems over conventional weapons, and making repeated statements in public that the United States was able and willing at all times to use nuclear weapons. In this way through the massive retaliation doctrine and the closely related foreign policy concept of brinkmanship, Eisenhower hoped to hold out a creditable vision of Clausewitzian nuclear “absolute war” in order to deter the Soviet Union and/or China from ever risking a war or even conditions that might lead to a war with the United States.
After 1970, some theorists claimed that nuclear proliferation made Clausewitzian concepts obsolete after the 20th-century period in which they dominated the world. John E. Sheppard, Jr., argues that by developing nuclear weapons, state-based conventional armies simultaneously both perfected their original purpose, to destroy a mirror image of themselves, and made themselves obsolete. No two powers have used nuclear weapons against each other, instead using conventional means or proxy wars to settle disputes. If such a conflict did occur, presumably both combatants would be annihilated. Heavily influenced by the war in Vietnam and by antipathy to American strategist Henry Kissinger, the American biologist, musician, and game-theorist Anatol Rapoport argued in 1968 that a Clausewitzian view of war was not only obsolete in the age of nuclear weapons, but also highly dangerous as it promoted a "zero-sum paradigm" to international relations and a "dissolution of rationality" amongst decision-makers.
The end of the 20th century and the beginning of the 21st century have seen many instances of state armies attempting to suppress insurgencies, terrorism, and other forms of asymmetrical warfare. Clausewitz did not focus solely on wars between countries with well-defined armies. The era of the French Revolution and Napoleon was full of revolutions, rebellions, and violence by "non-state actors", such as the wars in the French Vendée and in Spain. Clausewitz wrote a series of “Lectures on Small War” and studied the rebellion in the Vendée (1793–1796) and the Tyrolean uprising of 1809. In his famous “Bekenntnisdenkschrift” of 1812, he called for a “Spanish war in Germany” and laid out a comprehensive guerrilla strategy to be waged against Napoleon. In "On War" he included a famous chapter on “The People in Arms.”
One prominent critic of Clausewitz is the Israeli military historian Martin van Creveld. In his book "The Transformation of War", Creveld argued that Clausewitz's famous "Trinity" of people, army, and government was an obsolete socio-political construct based on the state, which was rapidly passing from the scene as the key player in war, and that he (Creveld) had constructed a new "non-trinitarian" model for modern warfare. Creveld's work has had great influence. Daniel Moran replied, 'The most egregious misrepresentation of Clausewitz's famous metaphor must be that of Martin van Creveld, who has declared Clausewitz to be an apostle of Trinitarian War, by which he means, incomprehensibly, a war of 'state against state and army against army,' from which the influence of the people is entirely excluded." Christopher Bassford went further, noting that one need only "read" the paragraph in which Clausewitz defined his Trinity to see "that the words 'people,' 'army,' and 'government' appear nowhere at all in the list of the Trinity’s components... Creveld's and Keegan's assault on Clausewitz's Trinity is not only a classic 'blow into the air,' i.e., an assault on a position Clausewitz doesn't occupy. It is also a pointless attack on a concept that is quite useful in its own right. In any case, their failure to read the actual wording of the theory they so vociferously attack, and to grasp its deep relevance to the phenomena they describe, is hard to credit."
Some have gone further and suggested that Clausewitz's best-known aphorism, that war is a continuation of politics with other means, is not only irrelevant today but also inapplicable historically. For an opposing view see the sixteen essays presented in "Clausewitz in the Twenty-First Century" edited by Hew Strachan and Andreas Herberg-Rothe.
In military academies, schools, and universities worldwide, Clausewitz's literature is often mandatory reading.
Literature
Film
Video games
August Otto Rühle von Lilienstern – Prussian officer from whom Clausewitz allegedly took, without acknowledgement, several important ideas (including that about war as pursuing political aims) made famous in "On War". However, such ideas as Clausewitz and Lilienstern shared in common derived from a common influence, i.e., Scharnhorst, who was Clausewitz's "second father" and professional mentor.
Informational notes
Citations
Further reading | https://en.wikipedia.org/wiki?curid=6066 |
Common Lisp
Common Lisp (CL) is a dialect of the Lisp programming language, published in ANSI standard document "ANSI INCITS 226-1994 (R2004)" (formerly "X3.226-1994 (R1999)"). The Common Lisp HyperSpec, a hyperlinked HTML version, has been derived from the ANSI Common Lisp standard.
The Common Lisp language was developed as a standardized and improved successor of Maclisp. By the early 1980s several groups were already at work on diverse successors to MacLisp: Lisp Machine Lisp (aka ZetaLisp), Spice Lisp, NIL and S-1 Lisp. Common Lisp sought to unify, standardise, and extend the features of these MacLisp dialects. Common Lisp is not an implementation, but rather a language specification. Several implementations of the Common Lisp standard are available, including free and open-source software and proprietary products.
Common Lisp is a general-purpose, multi-paradigm programming language. It supports a combination of procedural, functional, and object-oriented programming paradigms. As a dynamic programming language, it facilitates evolutionary and incremental software development, with iterative compilation into efficient run-time programs. This incremental development is often done interactively without interrupting the running application.
It also supports optional type annotation and casting, which can be added as necessary at the later profiling and optimization stages, to permit the compiler to generate more efficient code. For instance, codice_1 can hold an unboxed integer in a range supported by the hardware and implementation, permitting more efficient arithmetic than on big integers or arbitrary precision types. Similarly, the compiler can be told on a per-module or per-function basis which type of safety level is wanted, using "optimize" declarations.
Common Lisp includes CLOS, an object system that supports multimethods and method combinations. It is often implemented with a Metaobject Protocol.
Common Lisp is extensible through standard features such as "Lisp macros" (code transformations) and "reader macros" (input parsers for characters).
Common Lisp provides some backwards compatibility to Maclisp and to John McCarthy's original Lisp. This allows older Lisp software to be ported to Common Lisp.
Work on Common Lisp started in 1981 after an initiative by ARPA manager Bob Engelmore to develop a single community standard Lisp dialect. Much of the initial language design was done via electronic mail. In 1982, Guy L. Steele, Jr. gave the first overview of Common Lisp at the 1982 ACM Symposium on LISP and functional programming.
The first language documentation was published in 1984 as Common Lisp the Language (known as CLtL1), first edition. A second edition (known as CLtL2), published in 1990, incorporated many changes to the language, made during the ANSI Common Lisp standardization process: extended LOOP syntax, the Common Lisp Object System, the Condition System for error handling, an interface to the pretty printer and much more. But CLtL2 does not describe the final ANSI Common Lisp standard and thus is not a documentation of ANSI Common Lisp. The final ANSI Common Lisp standard then was published in 1994. Since then no update to the standard has been published. Various extensions and improvements to Common Lisp (examples are Unicode, Concurrency, CLOS-based IO) have been provided by implementations and libraries (many available via Quicklisp).
Common Lisp is a dialect of Lisp. It uses S-expressions to denote both code and data structure. Function calls, macro forms and special forms are written as lists, with the name of the operator first, as in these examples:
Common Lisp has many data types.
"Number" types include integers, ratios, floating-point numbers, and complex numbers. Common Lisp uses bignums to represent numerical values of arbitrary size and precision. The ratio type represents fractions exactly, a facility not available in many languages. Common Lisp automatically coerces numeric values among these types as appropriate.
The Common Lisp "character" type is not limited to ASCII characters. Most modern implementations allow Unicode characters.
The "symbol" type is common to Lisp languages, but largely unknown outside them. A symbol is a unique, named data object with several parts: name, value, function, property list, and package. Of these, "value cell" and "function cell" are the most important. Symbols in Lisp are often used similarly to identifiers in other languages: to hold the value of a variable; however there are many other uses. Normally, when a symbol is evaluated, its value is returned. Some symbols evaluate to themselves, for example, all symbols in the keyword package are self-evaluating. Boolean values in Common Lisp are represented by the self-evaluating symbols T and NIL. Common Lisp has namespaces for symbols, called 'packages'.
A number of functions are available for rounding scalar numeric values in various ways. The function codice_2 rounds the argument to the nearest integer, with halfway cases rounded to the even integer. The functions codice_3, codice_4, and codice_5 round towards zero, down, or up respectively. All these functions return the discarded fractional part as a secondary value. For example, codice_6 yields −3, 0.5; codice_7 yields −2, −0.5; codice_8 yields 2, 0.5; and codice_9 yields 4, −0.5.
"Sequence" types in Common Lisp include lists, vectors, bit-vectors, and strings. There are many operations that can work on any sequence type.
As in almost all other Lisp dialects, "lists" in Common Lisp are composed of "conses", sometimes called "cons cells" or "pairs". A cons is a data structure with two slots, called its "car" and "cdr". A list is a linked chain of conses or the empty list. Each cons's car refers to a member of the list (possibly another list). Each cons's cdr refers to the next cons—except for the last cons in a list, whose cdr refers to the codice_10 value. Conses can also easily be used to implement trees and other complex data structures; though it is usually advised to use structure or class instances instead. It is also possible to create circular data structures with conses.
Common Lisp supports multidimensional "arrays", and can dynamically resize "adjustable" arrays if required. Multidimensional arrays can be used for matrix mathematics. A "vector" is a one-dimensional array. Arrays can carry any type as members (even mixed types in the same array) or can be specialized to contain a specific type of members, as in a vector of bits. Usually, only a few types are supported. Many implementations can optimize array functions when the array used is type-specialized. Two type-specialized array types are standard: a "string" is a vector of characters, while a "bit-vector" is a vector of bits.
"Hash tables" store associations between data objects. Any object may be used as key or value. Hash tables are automatically resized as needed.
"Packages" are collections of symbols, used chiefly to separate the parts of a program into namespaces. A package may "export" some symbols, marking them as part of a public interface. Packages can use other packages.
"Structures", similar in use to C structs and Pascal records, represent arbitrary complex data structures with any number and type of fields (called "slots"). Structures allow single-inheritance.
"Classes" are similar to structures, but offer more dynamic features and multiple-inheritance. (See CLOS). Classes have been added late to Common Lisp and there is some conceptual overlap with structures. Objects created of classes are called "Instances". A special case is Generic Functions. Generic Functions are both functions and instances.
Common Lisp supports first-class functions. For instance, it is possible to write functions that take other functions as arguments or return functions as well. This makes it possible to describe very general operations.
The Common Lisp library relies heavily on such higher-order functions. For example, the codice_11 function takes a relational operator as an argument and key function as an optional keyword argument. This can be used not only to sort any type of data, but also to sort data structures according to a key.
The evaluation model for functions is very simple. When the evaluator encounters a form codice_12 then it presumes that the symbol named f is one of the following:
If f is the name of a function, then the arguments a1, a2, ..., an are evaluated in left-to-right order, and the function is found and invoked with those values supplied as parameters.
The macro codice_14 defines functions where a function definition gives the name of the function, the names of any arguments, and a function body:
Function definitions may include compiler directives, known as "declarations", which provide hints to the compiler about optimization settings or the data types of arguments. They may also include "documentation strings" (docstrings), which the Lisp system may use to provide interactive documentation:
Anonymous functions (function literals) are defined using codice_13 expressions, e.g. codice_16 for a function that squares its argument. Lisp programming style frequently uses higher-order functions for which it is useful to provide anonymous functions as arguments.
Local functions can be defined with codice_17 and codice_18.
There are several other operators related to the definition and manipulation of functions. For instance, a function may be compiled with the codice_19 operator. (Some Lisp systems run functions using an interpreter by default unless instructed to compile; others compile every function).
The macro codice_20 defines generic functions. Generic functions are a collection of methods.
The macro codice_21 defines methods.
Methods can specialize their parameters over CLOS "standard classes", "system classes", "structure classes" or individual objects. For many types, there are corresponding "system classes".
When a generic function is called, multiple-dispatch will determine the effective method to use.
Generic Functions are also a first class data type. There are many more features to Generic Functions and Methods than described above.
The namespace for function names is separate from the namespace for data variables. This is a key difference between Common Lisp and Scheme. For Common Lisp, operators that define names in the function namespace include codice_14, codice_17, codice_18, codice_21 and codice_20.
To pass a function by name as an argument to another function, one must use the codice_27 special operator, commonly abbreviated as codice_28. The first codice_11 example above refers to the function named by the symbol codice_30 in the function namespace, with the code codice_31. Conversely, to call a function passed in such a way, one would use the codice_32 operator on the argument.
Scheme's evaluation model is simpler: there is only one namespace, and all positions in the form are evaluated (in any order) – not just the arguments. Code written in one dialect is therefore sometimes confusing to programmers more experienced in the other. For instance, many Common Lisp programmers like to use descriptive variable names such as "list" or "string" which could cause problems in Scheme, as they would locally shadow function names.
Whether a separate namespace for functions is an advantage is a source of contention in the Lisp community. It is usually referred to as the "Lisp-1 vs. Lisp-2 debate". Lisp-1 refers to Scheme's model and Lisp-2 refers to Common Lisp's model. These names were coined in a 1988 paper by Richard P. Gabriel and Kent Pitman, which extensively compares the two approaches.
Common Lisp supports the concept of "multiple values", where any expression always has a single "primary value", but it might also have any number of "secondary values", which might be received and inspected by interested callers. This concept is distinct from returning a list value, as the secondary values are fully optional, and passed via a dedicated side channel. This means that callers may remain entirely unaware of the secondary values being there if they have no need for them, and it makes it convenient to use the mechanism for communicating information that is sometimes useful, but not always necessary. For example,
Multiple values are supported by a handful of standard forms, most common of which are the codice_35 special form for accessing secondary values and codice_36 for returning multiple values:
Other data types in Common Lisp include:
Like programs in many other programming languages, Common Lisp programs make use of names to refer to variables, functions, and many other kinds of entities. Named references are subject to scope.
The association between a name and the entity which the name refers to is called a binding.
Scope refers to the set of circumstances in which a name is determined to have a particular binding.
The circumstances which determine scope in Common Lisp include:
To understand what a symbol refers to, the Common Lisp programmer must know what kind of reference is being expressed, what kind of scope it uses if it is a variable reference (dynamic versus lexical scope), and also the run-time situation: in what environment is the reference resolved, where was the binding introduced into the environment, et cetera.
Some environments in Lisp are globally pervasive. For instance, if a new type is defined, it is known everywhere thereafter. References to that type look it up in this global environment.
One type of environment in Common Lisp is the dynamic environment. Bindings established in this environment have dynamic extent, which means that a binding is established at the start of the execution of some construct, such as a codice_51 block, and disappears when that construct finishes executing: its lifetime is tied to the dynamic activation and deactivation of a block. However, a dynamic binding is not just visible within that block; it is also visible to all functions invoked from that block. This type of visibility is known as indefinite scope. Bindings which exhibit dynamic extent (lifetime tied to the activation and deactivation of a block) and indefinite scope (visible to all functions which are called from that block) are said to have dynamic scope.
Common Lisp has support for dynamically scoped variables, which are also called special variables. Certain other kinds of bindings are necessarily dynamically scoped also, such as restarts and catch tags. Function bindings cannot be dynamically scoped using codice_17 (which only provides lexically scoped function bindings), but function objects (a first-level object in Common Lisp) can be assigned to dynamically scoped variables, bound using codice_51 in dynamic scope, then called using codice_32 or codice_57.
Dynamic scope is extremely useful because it adds referential clarity and discipline to global variables. Global variables are frowned upon in computer science as potential sources of error, because they can give rise to ad-hoc, covert channels of communication among modules that lead to unwanted, surprising interactions.
In Common Lisp, a special variable which has only a top-level binding behaves just like a global variable in other programming languages. A new value can be stored into it, and that value simply replaces what is in the top-level binding. Careless replacement of the value of a global variable is at the heart of bugs caused by the use of global variables. However, another way to work with a special variable is to give it a new, local binding within an expression. This is sometimes referred to as "rebinding" the variable. Binding a dynamically scoped variable temporarily creates a new memory location for that variable, and associates the name with that location. While that binding is in effect, all references to that variable refer to the new binding; the previous binding is hidden. When execution of the binding expression terminates, the temporary memory location is gone, and the old binding is revealed, with the original value intact. Of course, multiple dynamic bindings for the same variable can be nested.
In Common Lisp implementations which support multithreading, dynamic scopes are specific to each thread of execution. Thus special variables serve as an abstraction for thread local storage. If one thread rebinds a special variable, this rebinding has no effect on that variable in other threads. The value stored in a binding can only be retrieved by the thread which created that binding. If each thread binds some special variable codice_58, then codice_58 behaves like thread-local storage. Among threads which do not rebind codice_58, it behaves like an ordinary global: all of these threads refer to the same top-level binding of codice_58.
Dynamic variables can be used to extend the execution context with additional context information which is implicitly passed from function to function without having to appear as an extra function parameter. This is especially useful when the control transfer has to pass through layers of unrelated code, which simply cannot be extended with extra parameters to pass the additional data. A situation like this usually calls for a global variable. That global variable must be saved and restored, so that the scheme doesn't break under recursion: dynamic variable rebinding takes care of this. And that variable must be made thread-local (or else a big mutex must be used) so the scheme doesn't break under threads: dynamic scope implementations can take care of this also.
In the Common Lisp library, there are many standard special variables. For instance, all standard I/O streams are stored in the top-level bindings of well-known special variables. The standard output stream is stored in *standard-output*.
Suppose a function foo writes to standard output:
To capture its output in a character string, *standard-output* can be bound to a string stream and called:
Common Lisp supports lexical environments. Formally, the bindings in a lexical environment have lexical scope and may have either an indefinite extent or dynamic extent, depending on the type of namespace. Lexical scope means that visibility is physically restricted to the block in which the binding is established. References which are not textually (i.e. lexically) embedded in that block simply do not see that binding.
The tags in a TAGBODY have lexical scope. The expression (GO X) is erroneous if it is not embedded in a TAGBODY which contains a label X. However, the label bindings disappear when the TAGBODY terminates its execution, because they have dynamic extent. If that block of code is re-entered by the invocation of a lexical closure, it is invalid for the body of that closure to try to transfer control to a tag via GO:
When the TAGBODY is executed, it first evaluates the setf form which stores a function in the special variable *stashed*. Then the (go end-label) transfers control to end-label, skipping the code (print "Hello"). Since end-label is at the end of the tagbody, the tagbody terminates, yielding NIL. Suppose that the previously remembered function is now called:
This situation is erroneous. One implementation's response is an error condition containing the message, "GO: tagbody for tag SOME-LABEL has already been left". The function tried to evaluate (go some-label), which is lexically embedded in the tagbody, and resolves to the label. However, the tagbody isn't executing (its extent has ended), and so the control transfer cannot take place.
Local function bindings in Lisp have lexical scope, and variable bindings also have lexical scope by default. By contrast with GO labels, both of these have indefinite extent. When a lexical function or variable binding is established, that binding continues to exist for as long as references to it are possible, even after the construct which established that binding has terminated. References to lexical variables and functions after the termination of their establishing construct are possible thanks to lexical closures.
Lexical binding is the default binding mode for Common Lisp variables. For an individual symbol, it can be switched to dynamic scope, either by a local declaration, by a global declaration. The latter may occur implicitly through the use of a construct like DEFVAR or DEFPARAMETER. It is an important convention in Common Lisp programming that special (i.e. dynamically scoped) variables have names which begin and end with an asterisk sigil codice_62 in what is called the "earmuff convention". If adhered to, this convention effectively creates a separate namespace for special variables, so that variables intended to be lexical are not accidentally made special.
Lexical scope is useful for several reasons.
Firstly, references to variables and functions can be compiled to efficient machine code, because the run-time environment structure is relatively simple. In many cases it can be optimized to stack storage, so opening and closing lexical scopes has minimal overhead. Even in cases where full closures must be generated, access to the closure's environment is still efficient; typically each variable becomes an offset into a vector of bindings, and so a variable reference becomes a simple load or store instruction with a base-plus-offset addressing mode.
Secondly, lexical scope (combined with indefinite extent) gives rise to the lexical closure, which in turn creates a whole paradigm of programming centered around the use of functions being first-class objects, which is at the root of functional programming.
Thirdly, perhaps most importantly, even if lexical closures are not exploited, the use of lexical scope isolates program modules from unwanted interactions. Due to their restricted visibility, lexical variables are private. If one module A binds a lexical variable X, and calls another module B, references to X in B will not accidentally resolve to the X bound in A. B simply has no access to X. For situations in which disciplined interactions through a variable are desirable, Common Lisp provides special variables. Special variables allow for a module A to set up a binding for a variable X which is visible to another module B, called from A. Being able to do this is an advantage, and being able to prevent it from happening is also an advantage; consequently, Common Lisp supports both lexical and dynamic scope.
A "macro" in Lisp superficially resembles a function in usage. However, rather than representing an expression which is evaluated, it represents a transformation of the program source code. The macro gets the source it surrounds as arguments, binds them to its parameters and computes a new source form. This new form can also use a macro. The macro expansion is repeated until the new source form does not use a macro. The final computed form is the source code executed at runtime.
Typical uses of macros in Lisp:
Various standard Common Lisp features also need to be implemented as macros, such as:
Macros are defined by the "defmacro" macro. The special operator "macrolet" allows the definition of local (lexically scoped) macros. It is also possible to define macros for symbols using "define-symbol-macro" and "symbol-macrolet".
Paul Graham's book On Lisp describes the use of macros in Common Lisp in detail. Doug Hoyte's book Let Over Lambda extends the discussion on macros, claiming "Macros are the single greatest advantage that lisp has as a programming language and the single greatest advantage of any programming language." Hoyte provides several examples of iterative development of macros.
Macros allow Lisp programmers to create new syntactic forms in the language. One typical use is to create new control structures. The example macro provides an codice_73 looping construct. The syntax is:
The macro definition for "until":
"tagbody" is a primitive Common Lisp special operator which provides the ability to name tags and use the "go" form to jump to those tags. The backquote "`" provides a notation that provides code templates, where the value of forms preceded with a comma are filled in. Forms preceded with comma and at-sign are "spliced" in. The tagbody form tests the end condition. If the condition is true, it jumps to the end tag. Otherwise the provided body code is executed and then it jumps to the start tag.
An example form using above "until" macro:
The code can be expanded using the function "macroexpand-1". The expansion for above example looks like this:
(TAGBODY
During macro expansion the value of the variable "test" is "(= (random 10) 0)" and the value of the variable "body" is "((write-line "Hello"))". The body is a list of forms.
Symbols are usually automatically upcased. The expansion uses the TAGBODY with two labels. The symbols for these labels are computed by GENSYM and are not interned in any package. Two "go" forms use these tags to jump to. Since "tagbody" is a primitive operator in Common Lisp (and not a macro), it will not be expanded into something else. The expanded form uses the "when" macro, which also will be expanded. Fully expanding a source form is called "code walking".
In the fully expanded ("walked") form, the "when" form is replaced by the primitive "if":
(TAGBODY
All macros must be expanded before the source code containing them can be evaluated or compiled normally. Macros can be considered functions that accept and return S-expressions – similar to abstract syntax trees, but not limited to those. These functions are invoked before the evaluator or compiler to produce the final source code.
Macros are written in normal Common Lisp, and may use any Common Lisp (or third-party) operator available.
Common Lisp macros are capable of what is commonly called "variable capture", where symbols in the macro-expansion body coincide with those in the calling context, allowing the programmer to create macros wherein various symbols have special meaning. The term "variable capture" is somewhat misleading, because all namespaces are vulnerable to unwanted capture, including the operator and function namespace, the tagbody label namespace, catch tag, condition handler and restart namespaces.
"Variable capture" can introduce software defects. This happens in one of the following two ways:
The Scheme dialect of Lisp provides a macro-writing system which provides the referential transparency that eliminates both types of capture problem. This type of macro system is sometimes called "hygienic", in particular by its proponents (who regard macro systems which do not automatically solve this problem as unhygienic).
In Common Lisp, macro hygiene is ensured one of two different ways.
One approach is to use gensyms: guaranteed-unique symbols which can be used in a macro-expansion without threat of capture. The use of gensyms in a macro definition is a manual chore, but macros can be written which simplify the instantiation and use of gensyms. Gensyms solve type 2 capture easily, but they are not applicable to type 1 capture in the same way, because the macro expansion cannot rename the interfering symbols in the surrounding code which capture its references. Gensyms could be used to provide stable aliases for the global symbols which the macro expansion needs. The macro expansion would use these secret aliases rather than the well-known names, so redefinition of the well-known names would have no ill effect on the macro.
Another approach is to use packages. A macro defined in its own package can simply use internal symbols in that package in its expansion. The use of packages deals with type 1 and type 2 capture.
However, packages don't solve the type 1 capture of references to standard Common Lisp functions and operators. The reason is that the use of packages to solve capture problems revolves around the use of private symbols (symbols in one package, which are not imported into, or otherwise made visible in other packages). Whereas the Common Lisp library symbols are external, and frequently imported into or made visible in user-defined packages.
The following is an example of unwanted capture in the operator namespace, occurring in the expansion of a macro:
The codice_73 macro will expand into a form which calls codice_75 which is intended to refer to the standard Common Lisp macro codice_75. However, in this context, codice_75 may have a completely different meaning, so codice_73 may not work properly.
Common Lisp solves the problem of the shadowing of standard operators and functions by forbidding their redefinition. Because it redefines the standard operator codice_75, the preceding is actually a fragment of non-conforming Common Lisp, which allows implementations to diagnose and reject it.
The "condition system" is responsible for exception handling in Common Lisp. It provides "conditions", "handler"s and "restart"s. "Condition"s are objects describing an exceptional situation (for example an error). If a "condition" is signaled, the Common Lisp system searches for a "handler" for this condition type and calls the handler. The "handler" can now search for restarts and use one of these restarts to automatically repair the current problem, using information such as the condition type and any relevant information provided as part of the condition object, and call the appropriate restart function.
These restarts, if unhandled by code, can be presented to users (as part of a user interface, that of a debugger for example), so that the user can select and invoke one of the available restarts. Since the condition handler is called in the context of the error (without unwinding the stack), full error recovery is possible in many cases, where other exception handling systems would have already terminated the current routine. The debugger itself can also be customized or replaced using the codice_80 dynamic variable. Code found within "unwind-protect" forms such as finalizers will also be executed as appropriate despite the exception.
In the following example (using Symbolics Genera) the user tries to open a file in a Lisp function "test" called from the Read-Eval-Print-LOOP (REPL), when the file does not exist. The Lisp system presents four restarts. The user selects the "Retry OPEN using a different pathname" restart and enters a different pathname (lispm-init.lisp instead of lispm-int.lisp). The user code does not contain any error handling code. The whole error handling and restart code is provided by the Lisp system, which can handle and repair the error without terminating the user code.
Command: (test ">zippy>lispm-int.lisp")
Error: The file was not found.
LMFS:OPEN-LOCAL-LMFS-1
s-A, : Retry OPEN of lispm:>zippy>lispm-int.lisp.newest
s-B: Retry OPEN using a different pathname
s-C, : Return to Lisp Top Level in a TELNET server
s-D: Restart process TELNET terminal
-> Retry OPEN using a different pathname
Use what pathname instead [default lispm:>zippy>lispm-int.lisp.newest]:
...the program continues
Common Lisp includes a toolkit for object-oriented programming, the Common Lisp Object System or CLOS, which is one of the most powerful object systems available in any language. For example, Peter Norvig explains how many Design Patterns are simpler to implement in a dynamic language with the features of CLOS (Multiple Inheritance, Mixins, Multimethods, Metaclasses, Method combinations, etc.).
Several extensions to Common Lisp for object-oriented programming have been proposed to be included into the ANSI Common Lisp standard, but eventually CLOS was adopted as the standard object-system for Common Lisp. CLOS is a dynamic object system with multiple dispatch and multiple inheritance, and differs radically from the OOP facilities found in static languages such as C++ or Java. As a dynamic object system, CLOS allows changes at runtime to generic functions and classes. Methods can be added and removed, classes can be added and redefined, objects can be updated for class changes and the class of objects can be changed.
CLOS has been integrated into ANSI Common Lisp. Generic functions can be used like normal functions and are a first-class data type. Every CLOS class is integrated into the Common Lisp type system. Many Common Lisp types have a corresponding class. There is more potential use of CLOS for Common Lisp. The specification does not say whether conditions are implemented with CLOS. Pathnames and streams could be implemented with CLOS. These further usage possibilities of CLOS for ANSI Common Lisp are not part of the standard. Actual Common Lisp implementations use CLOS for pathnames, streams, input–output, conditions, the implementation of CLOS itself and more.
A Lisp interpreter directly executes Lisp source code provided as Lisp objects (lists, symbols, numbers, ...) read from s-expressions. A Lisp compiler generates bytecode or machine code from Lisp source code. Common Lisp allows both individual Lisp functions to be compiled in memory and the compilation of whole files to externally stored compiled code ("fasl" files).
Several implementations of earlier Lisp dialects provided both an interpreter and a compiler. Unfortunately often the semantics were different. These earlier Lisps implemented lexical scoping in the compiler and dynamic scoping in the interpreter. Common Lisp requires that both the interpreter and compiler use lexical scoping by default. The Common Lisp standard describes both the semantics of the interpreter and a compiler. The compiler can be called using the function "compile" for individual functions and using the function "compile-file" for files. Common Lisp allows type declarations and provides ways to influence the compiler code generation policy. For the latter various optimization qualities can be given values between 0 (not important) and 3 (most important): "speed", "space", "safety", "debug" and "compilation-speed".
There is also a function to evaluate Lisp code: codice_81. codice_81 takes code as pre-parsed s-expressions and not, like in some other languages, as text strings. This way code can be constructed with the usual Lisp functions for constructing lists and symbols and then this code can be evaluated with the function codice_81. Several Common Lisp implementations (like Clozure CL and SBCL) are implementing codice_81 using their compiler. This way code is compiled, even though it is evaluated using the function codice_81.
The file compiler is invoked using the function "compile-file". The generated file with compiled code is called a "fasl" (from "fast load") file. These "fasl" files and also source code files can be loaded with the function "load" into a running Common Lisp system. Depending on the implementation, the file compiler generates byte-code (for example for the Java Virtual Machine), C language code (which then is compiled with a C compiler) or, directly, native code.
Common Lisp implementations can be used interactively, even though the code gets fully compiled. The idea of an Interpreted language thus does not apply for interactive Common Lisp.
The language makes a distinction between read-time, compile-time, load-time, and run-time, and allows user code to also make this distinction to perform the wanted type of processing at the wanted step.
Some special operators are provided to especially suit interactive development; for instance, codice_86 will only assign a value to its provided variable if it wasn't already bound, while codice_87 will always perform the assignment. This distinction is useful when interactively evaluating, compiling and loading code in a live image.
Some features are also provided to help writing compilers and interpreters. Symbols consist of first-level objects and are directly manipulable by user code. The codice_88 special operator allows to create lexical bindings programmatically, while packages are also manipulable. The Lisp compiler is available at runtime to compile files or individual functions. These make it easy to use Lisp as an intermediate compiler or interpreter for another language.
The following program calculates the smallest number of people in a room for whom the probability of unique birthdays is less than 50% (the birthday paradox, where for 1 person the probability is obviously 100%, for 2 it is 364/365, etc.). The answer is 23.
By convention, constants in Common Lisp are enclosed with + characters.
Calling the example function using the REPL (Read Eval Print Loop):
CL-USER > (birthday-paradox 1.0 1)
23
We define a class codice_89 and a method for displaying the name and age of a person.
Next we define a group of persons as a list of codice_89 objects.
Then we iterate over the sorted list.
(defparameter *group*
It prints the three names with descending age.
Bob (33)
Ash (23)
Chris (16)
Use of the LOOP macro is demonstrated:
Example use:
CL-USER > (power 2 200)
1606938044258990275541962092341162602522202993782792835301376
Compare with the built in exponentiation:
CL-USER > (= (expt 2 200) (power 2 200))
T
WITH-OPEN-FILE is a macro that opens a file and provides a stream. When the form is returning, the file is automatically closed. FUNCALL calls a function object. The LOOP collects all lines that match the predicate.
The function AVAILABLE-SHELLS calls above function LIST-MATCHING-LINES with a pathname and an anonymous function as the predicate. The predicate returns the pathname of a shell or NIL (if the string is not the filename of a shell).
Example results (on Mac OS X 10.6):
CL-USER > (available-shells)
Common Lisp is most frequently compared with, and contrasted to, Scheme—if only because they are the two most popular Lisp dialects. Scheme predates CL, and comes not only from the same Lisp tradition but from some of the same engineers—Guy L. Steele, with whom Gerald Jay Sussman designed Scheme, chaired the standards committee for Common Lisp.
Common Lisp is a general-purpose programming language, in contrast to Lisp variants such as Emacs Lisp and AutoLISP which are extension languages embedded in particular products (GNU Emacs and AutoCAD, respectively). Unlike many earlier Lisps, Common Lisp (like Scheme) uses lexical variable scope by default for both interpreted and compiled code.
Most of the Lisp systems whose designs contributed to Common Lisp—such as ZetaLisp and Franz Lisp—used dynamically scoped variables in their interpreters and lexically scoped variables in their compilers. Scheme introduced the sole use of lexically scoped variables to Lisp; an inspiration from ALGOL 68. CL supports dynamically scoped variables as well, but they must be explicitly declared as "special". There are no differences in scoping between ANSI CL interpreters and compilers.
Common Lisp is sometimes termed a "Lisp-2" and Scheme a "Lisp-1", referring to CL's use of separate namespaces for functions and variables. (In fact, CL has "many" namespaces, such as those for go tags, block names, and codice_72 keywords). There is a long-standing controversy between CL and Scheme advocates over the tradeoffs involved in multiple namespaces. In Scheme, it is (broadly) necessary to avoid giving variables names which clash with functions; Scheme functions frequently have arguments named codice_92, codice_93, or codice_94 so as not to conflict with the system function codice_95. However, in CL it is necessary to explicitly refer to the function namespace when passing a function as an argument—which is also a common occurrence, as in the codice_11 example above.
CL also differs from Scheme in its handling of boolean values. Scheme uses the special values #t and #f to represent truth and falsity. CL follows the older Lisp convention of using the symbols T and NIL, with NIL standing also for the empty list. In CL, "any" non-NIL value is treated as true by conditionals, such as codice_68, whereas in Scheme all non-#f values are treated as true. These conventions allow some operators in both languages to serve both as predicates (answering a boolean-valued question) and as returning a useful value for further computation, but in Scheme the value '() which is equivalent to NIL in Common Lisp evaluates to true in a boolean expression.
Lastly, the Scheme standards documents require tail-call optimization, which the CL standard does not. Most CL implementations do offer tail-call optimization, although often only when the programmer uses an optimization directive. Nonetheless, common CL coding style does not favor the ubiquitous use of recursion that Scheme style prefers—what a Scheme programmer would express with tail recursion, a CL user would usually express with an iterative expression in codice_75, codice_99, codice_72, or (more recently) with the codice_101 package.
See the Category .
Common Lisp is defined by a specification (like Ada and C) rather than by one implementation (like Perl). There are many implementations, and the standard details areas in which they may validly differ.
In addition, implementations tend to come with extensions, which provide functionality not covered in the standard:
Free and open-source software libraries have been created to support extensions to Common Lisp in a portable way, and are most notably found in the repositories of the Common-Lisp.net and CLOCC (Common Lisp Open Code Collection) projects.
Common Lisp implementations may use any mix of native code compilation, byte code compilation or interpretation. Common Lisp has been designed to support incremental compilers, file compilers and block compilers. Standard declarations to optimize compilation (such as function inlining or type specialization) are proposed in the language specification. Most Common Lisp implementations compile source code to native machine code. Some implementations can create (optimized) stand-alone applications. Others compile to interpreted bytecode, which is less efficient than native code, but eases binary-code portability. Some compilers compile Common Lisp code to C code. The misconception that Lisp is a purely interpreted language is most likely because Lisp environments provide an interactive prompt and that code is compiled one-by-one, in an incremental way. With Common Lisp incremental compilation is widely used.
Some Unix-based implementations (CLISP, SBCL) can be used as a scripting language; that is, invoked by the system transparently in the way that a Perl or Unix shell interpreter is.
See the Category .
Common Lisp is used to develop research applications (often in Artificial Intelligence), for rapid development of prototypes or for deployed applications.
Common Lisp is used in many commercial applications, including the Yahoo! Store web-commerce site, which originally involved Paul Graham and was later rewritten in C++ and Perl. Other notable examples include:
There also exist open-source applications written in Common Lisp, such as:
Since 2011, Zach Beane, with support of the Common Lisp Foundation, has maintained the Quicklisp library manager. It allows automatic download, installing, and loading of over 3600 libraries, all of which are required to work on more than just one implementation of Common Lisp and to have a license that allows their redistribution.
A chronological list of books published (or about to be published) about Common Lisp (the language) or about programming with Common Lisp (especially AI programming). | https://en.wikipedia.org/wiki?curid=6068 |
Color code
A color code or colour code is a system for displaying information by using different colors.
The earliest examples of color codes in use are for long distance communication by use of flags, as in semaphore communication. The United Kingdom adopted a color code scheme for such communication wherein red signified danger and white signified safety, with other colors having similar assignments of meaning.
As chemistry and other technologies advanced, it became expedient to use coloration as a signal for telling apart things that would otherwise be confusingly similar, such as wiring in electrical and electronic devices, and pharmaceutical pills.
The use of color codes has been extended to abstractions, such as the Homeland Security Advisory System color code in the United States. Similarly, hospital emergency codes often incorporate colors (such as the widely used "Code Blue" indicating a cardiac arrest), although they may also include numbers, and may not conform to a uniform standard.
Color codes do present some potential problems. On forms and signage, the use of color can distract from black and white text. They are often difficult for color blind and blind people to interpret, and even for those with normal color vision, use of many colors to code many variables can lead to use of confusingly similar colors.
Systems incorporating color-coding include: | https://en.wikipedia.org/wiki?curid=6069 |
Cauchy sequence
In mathematics, a Cauchy sequence (; ), named after Augustin-Louis Cauchy, is a sequence whose elements become arbitrarily close to each other as the sequence progresses. More precisely, given any small positive distance, all but a finite number of elements of the sequence are less than that given distance from each other.
It is not sufficient for each term to become arbitrarily close to the term. For instance, in the sequence of square roots of natural numbers:
the consecutive terms become arbitrarily close to each other:
However, with growing values of the index , the terms become arbitrarily large, so for any index and distance , there exists an index big enough such that . (Actually, any suffices.) As a result, despite how far one goes, the remaining terms of the sequence never get close to , hence the sequence is not Cauchy.
The utility of Cauchy sequences lies in the fact that in a complete metric space (one where all such sequences are known to converge to a limit), the criterion for convergence depends only on the terms of the sequence itself, as opposed to the definition of convergence, which uses the limit value as well as the terms. This is often exploited in algorithms, both theoretical and applied, where an iterative process can be shown relatively easily to produce a Cauchy sequence, consisting of the iterates, thus fulfilling a logical condition, such as termination.
The notions above are not as unfamiliar as they might at first appear. The customary acceptance of the fact that any real number "x" has a decimal expansion is an implicit acknowledgment that a particular Cauchy sequence of rational numbers (whose terms are the successive truncations of the decimal expansion of "x") has the real limit "x". In some cases it may be difficult to describe "x" independently of such a limiting process involving rational numbers.
Generalizations of Cauchy sequences in more abstract uniform spaces exist in the form of Cauchy filters and Cauchy nets.
A sequence
of real numbers is called a Cauchy sequence if for every positive real number "ε", there is a positive integer "N" such that for all natural numbers "m", "n" > "N"
where the vertical bars denote the absolute value. In a similar way one can define Cauchy sequences of rational or complex numbers. Cauchy formulated such a condition by requiring formula_5 to be infinitesimal for every pair of infinite "m", "n".
If formula_6 is a sequence in the set formula_7, then a "modulus of Cauchy convergence" for the sequence is a function formula_8 from the set of natural numbers to itself, such that formula_9.
Any sequence with a modulus of Cauchy convergence is a Cauchy sequence. The existence of a modulus for a Cauchy sequence follows from the well-ordering property of the natural numbers (let formula_10 be the smallest possible formula_11 in the definition of Cauchy sequence, taking formula_12 to be formula_13). The existence of a modulus also follows from the principle of dependent choice, which is a weak form of the axiom of choice, and it also follows from an even weaker condition called AC00. "Regular Cauchy sequences" are sequences with a given modulus of Cauchy convergence (usually formula_14 or formula_15). Any Cauchy sequence with a modulus of Cauchy convergence is equivalent to a regular Cauchy sequence; this can be proved without using any form of the axiom of choice.
Moduli of Cauchy convergence are used by constructive mathematicians who do not wish to use any form of choice. Using a modulus of Cauchy convergence can simplify both definitions and theorems in constructive analysis. Regular Cauchy sequences were used by Errett Bishop in his Foundations of Constructive Analysis, and by Douglas Bridges in a non-constructive textbook ().
Since the definition of a Cauchy sequence only involves metric concepts, it is straightforward to generalize it to any metric space "X". To do so, the absolute value |"x""m" - "x""n"| is replaced by the distance "d"("x""m", "x""n") (where "d" denotes a metric) between "x""m" and "x""n".
Formally, given a metric space ("X", "d"), a sequence
is Cauchy, if for every positive real number "ε" > 0 there is a positive integer "N" such that for all positive integers "m", "n" > "N", the distance
Roughly speaking, the terms of the sequence are getting closer and closer together in a way that suggests that the sequence ought to have a limit in "X". Nonetheless, such a limit does not always exist within "X": the property of a space that every Cauchy sequence converges in the space is called "completeness", and is detailed below.
A metric space ("X", "d") in which every Cauchy sequence converges to an element of "X" is called complete.
The real numbers are complete under the metric induced by the usual absolute value, and one of the standard constructions of the real numbers involves Cauchy sequences of rational numbers. In this construction, each equivalence class of Cauchy sequences of rational numbers with a certain tail behavior—that is, each class of sequences that get arbitrarily close to one another— is a real number.
A rather different type of example is afforded by a metric space "X" which has the discrete metric (where any two distinct points are at distance 1 from each other). Any Cauchy sequence of elements of "X" must be constant beyond some fixed point, and converges to the eventually repeating term.
The rational numbers Q are not complete (for the usual distance):
There are sequences of rationals that converge (in R) to irrational numbers; these are Cauchy sequences having no limit in Q. In fact, if a real number "x" is irrational, then the sequence ("x""n"), whose "n"-th term is the truncation to "n" decimal places of the decimal expansion of "x", gives a Cauchy sequence of rational numbers with irrational limit "x". Irrational numbers certainly exist in R, for example:
The open interval formula_21 in the set of real numbers with an ordinary distance in R is not a complete space: there is a sequence formula_22 in it, which is Cauchy (for arbitrarily small distance bound formula_23 all terms formula_24 of formula_25 fit in the formula_26 interval), however does not converge in formula_27 — its 'limit', number formula_28, does not belong to the space formula_27.
These last two properties, together with the Bolzano–Weierstrass theorem, yield one standard proof of the completeness of the real numbers, closely related to both the Bolzano–Weierstrass theorem and the Heine–Borel theorem. Every Cauchy sequence of real numbers is bounded, hence by Bolzano-Weierstrass has a convergent subsequence, hence is itself convergent. This proof of the completeness of the real numbers implicitly makes use of the least upper bound axiom. The alternative approach, mentioned above, of the real numbers as the completion of the rational numbers, makes the completeness of the real numbers tautological.
One of the standard illustrations of the advantage of being able to work with Cauchy sequences and make use of completeness is provided by consideration of the summation of an infinite series of real numbers
(or, more generally, of elements of any complete normed linear space, or Banach space). Such a series
formula_30 is considered to be convergent if and only if the sequence of partial sums formula_31 is convergent, where
formula_32. It is a routine matter
to determine whether the sequence of partial sums is Cauchy or not,
since for positive integers "p" > "q",
If formula_34 is a uniformly continuous map between the metric spaces "M" and "N" and ("x""n") is a Cauchy sequence in "M", then formula_35 is a Cauchy sequence in "N". If formula_36 and formula_37 are two Cauchy sequences in the rational, real or complex numbers, then the sum formula_38 and the product formula_39 are also Cauchy sequences.
There is also a concept of Cauchy sequence for a topological vector space formula_7: Pick a local base formula_41 for formula_7 about 0; then (formula_43) is a Cauchy sequence if for each member formula_44, there is some number formula_11 such that whenever
formula_46 is an element of formula_47. If the topology of formula_7 is compatible with a translation-invariant metric formula_49, the two definitions agree.
Since the topological vector space definition of Cauchy sequence requires only that there be a continuous "subtraction" operation, it can just as well be stated in the context of a topological group: A sequence formula_50 in a topological group formula_51 is a Cauchy sequence if for every open neighbourhood formula_52 of the identity in formula_51 there exists some number formula_11 such that whenever formula_55 it follows that formula_56. As above, it is sufficient to check this for the neighbourhoods in any local base of the identity in formula_51.
As in the construction of the completion of a metric space, one can furthermore define the binary relation on Cauchy sequences in formula_51 that formula_50 and formula_60 are equivalent if for every open neighbourhood formula_52 of the identity in formula_51 there exists some number formula_11 such that whenever formula_55 it follows that formula_65. This relation is an equivalence relation: It is reflexive since the sequences are Cauchy sequences. It is symmetric since formula_66 which by continuity of the inverse is another open neighbourhood of the identity. It is transitive since formula_67 where formula_68 and formula_69 are open neighbourhoods of the identity such that formula_70; such pairs exist by the continuity of the group operation.
There is also a concept of Cauchy sequence in a group formula_51:
Let formula_72 be a decreasing sequence of normal subgroups of formula_51 of finite index.
Then a sequence formula_36 in formula_51 is said to be Cauchy (w.r.t. formula_76) if and only if for any formula_12 there is formula_11 such that formula_79.
Technically, this is the same thing as a topological group Cauchy sequence for a particular choice of topology on formula_51, namely that for which formula_76 is a local base.
The set formula_82 of such Cauchy sequences forms a group (for the componentwise product), and the set formula_83 of null sequences (s.th. formula_84) is a normal subgroup of formula_82. The factor group formula_86 is called the completion of formula_51 with respect to formula_76.
One can then show that this completion is isomorphic to the inverse limit of the sequence formula_89.
An example of this construction, familiar in number theory and algebraic geometry is the construction of the "p"-adic completion of the integers with respect to a prime "p". In this case, "G" is the integers under addition, and "H""r" is the additive subgroup consisting of integer multiples of "p""r".
If formula_76 is a cofinal sequence (i.e., any normal subgroup of finite index contains some formula_91), then this completion is canonical in the sense that it is isomorphic to the inverse limit of formula_92, where formula_76 varies over normal subgroups of finite index. For further details, see ch. I.10 in Lang's "Algebra".
A real sequence formula_94 has a natural hyperreal extension, defined for hypernatural values "H" of the index "n" in addition to the usual natural "n". The sequence is Cauchy if and only if for every infinite "H" and "K", the values formula_95 and formula_96 are infinitely close, or adequal, i.e.
where "st" is the standard part function.
introduced a notion of Cauchy completion of a category. Applied to Q (the category whose objects are rational numbers, and there is a morphism from "x" to "y" if and only if "x" ≤ "y"), this Cauchy completion yields R (again interpreted as a category using its natural ordering). | https://en.wikipedia.org/wiki?curid=6085 |
Common Era
Common Era (CE) is one of the notation systems for the world's most widely used calendar era. BCE (Before the Common Era or Before the Current Era) is the era before CE. BCE and CE are alternatives to the Dionysian BC and AD system respectively. The Dionysian era distinguishes eras using BC ("before Christ") and AD (", "in [the] year of [the] Lord"). The two notation systems are numerically equivalent: " CE" and "AD " each describe the current year; "400 BCE" and "400 BC" are each the same year. Both notations refer to the Gregorian calendar (and its predecessor, the Julian calendar). The year-numbering system used by the Gregorian calendar is used throughout the world today, and is an international standard for civil calendars.
The expression has been traced back to 1615, when it first appeared in a book by Johannes Kepler as the Latin ", and to 1635 in English as "Vulgar Era". The term "Common Era" can be found in English as early as 1708, and became more widely used in the mid-19th century by Jewish religious scholars. In the later 20th century, the use of CE and BCE was popularized in academic and scientific publications as a culturally neutral term. It is also used by some authors and publishers who wish to emphasize sensitivity to non-Christians by not explicitly referencing Jesus as "Christ" and "Dominus" ("Lord") through use of the abbreviation "AD".
The year numbering system used with Common Era notation was devised by the Christian monk Dionysius Exiguus in the year 525 to replace the Era of Martyrs system, because he did not wish to continue the memory of a tyrant who persecuted Christians. He attempted to number years from an initial reference date ("epoch"), an event he referred to as the Incarnation of Jesus. Dionysius labeled the column of the table in which he introduced the new era as ""Anni Domini Nostri Jesu Christi"".
This way of numbering years became more widespread in Europe with its use by Bede in England in 731. Bede also introduced the practice of dating years before what he supposed was the year of birth of Jesus, and the practice of not using a year zero. In 1422, Portugal became the last Western European country to switch to the system begun by Dionysius.
The term "Common Era" is traced back in English to its appearance as "Vulgar Era" to distinguish dates on the Ecclesiastic calendar in popular use from dates of the regnal year, the year of reign of a sovereign, typically used in national law. (The word 'vulgar' originally meant 'of the ordinary people', with no derogatory associations.)
The first use of the Latin term "anno aerae nostrae vulgaris" discovered so far was in a 1615 book by Johannes Kepler. Kepler uses it again, as "ab Anno vulgaris aerae", in a 1616 table of ephemerides, and again, as "ab anno vulgaris aerae", in 1617. A 1635 English edition of that book has the title page in English – so far, the earliest-found use of "Vulgar Era" in English. A 1701 book edited by John LeClerc includes "Before Christ according to the Vulgar Æra, 6". A 1716 book in English by Dean Humphrey Prideaux says, "before the beginning of the vulgar æra, by which we now compute the years from his incarnation." A 1796 book uses the term "vulgar era of the nativity".
The first known use of "Christian Era" is as the Latin phrase "annus aerae christianae" on the title page of a 1584 theology book. In 1649, the Latin phrase "annus æræ Christianæ" appeared in the title of an English almanac. A 1652 ephemeris is the first instance found so far of the English use of "Christian Era".
The English phrase "common Era" appears at least as early as 1708, and in a 1715 book on astronomy it is used interchangeably with "Christian Era" and "Vulgar Era". A 1759 history book uses "common æra" in a generic sense, to refer to the common era of the Jews. The first use found so far of the phrase "before the common era" is in a 1770 work that also uses "common era" and "vulgar era" as synonyms, in a translation of a book originally written in German. The 1797 edition of the Encyclopædia Britannica uses the terms "vulgar era" and "common era" synonymously. In 1835, in his book "Living Oracles", Alexander Campbell, wrote: "The vulgar Era, or Anno Domini; the fourth year of Jesus Christ, the first of which was but eight days", and also refers to the "common era" as a synonym for "vulgar era" with "the fact that our Lord was born on the 4th year before the vulgar era, called Anno Domini, thus making (for example) the 42d year from his birth to correspond with the 38th of the common era..." The "Catholic Encyclopedia" (1909) in at least one article reports all three terms (Christian, Vulgar, Common Era) being commonly understood by the early 20th century.
The phrase "common era", in lower case, also appeared in the 19th century in a "generic" sense, not necessarily to refer to the Christian Era, but to any system of dates in common use throughout a civilization. Thus, "the common era of the Jews", "the common era of the Mahometans", "common era of the world", "the common era of the foundation of Rome". When it did refer to the Christian Era, it was sometimes qualified, e.g., "common era of the Incarnation", "common era of the Nativity", or "common era of the birth of Christ".
An adapted translation of "Common Era" into Latin as "Era Vulgaris" ("era"—or, with a macron, "ēra"—being an alternative form of "aera"; "aera" is the usual form) was adopted in the 20th century by some followers of Aleister Crowley, and thus the abbreviation "e.v." or "EV" may sometimes be seen as a replacement for AD.
Although Jews have their own Hebrew calendar, they often use the Gregorian calendar, without the AD prefix. As early as 1825, the abbreviation VE (for Vulgar Era) was in use among Jews to denote years in the Western calendar. As of 2005, Common Era notation has also been in use for Hebrew lessons for more than a century. In 1856, Rabbi and historian Morris Jacob Raphall used the abbreviations CE and BCE in his book "Post-Biblical History of The Jews". Jews have also used the term Current Era.
Some academics in the fields of theology, education, archaeology and history have adopted CE and BCE notation, although there is some disagreement. Several style guides now prefer or mandate its use.
The style guide for the Episcopal Diocese "Maryland Church News" says that BCE and CE should be used.
In the United States, the use of the BCE/CE notation in textbooks was reported in 2005 to be growing. Some publications have moved over to using it exclusively. For example, the 2007 World Almanac was the first edition to switch over to BCE/CE, ending a period of 138 years in which the traditional BC/AD dating notation was used. BCE/CE is used by the College Board in its history tests, and by the Norton Anthology of English Literature. Others have taken a different approach. The US-based History Channel uses BCE/CE notation in articles on non-Christian religious topics such as Jerusalem and Judaism.
In 2002, an advisory panel for the religious education syllabus for England and Wales recommended introducing BCE/CE dates to schools, and by 2018 some local education authorities were using them. In 2018, the National Trust said it would continue to use BC/AD as its house style. English Heritage explains its era policy thus: "It might seem strange to use a Christian calendar system when referring to British prehistory, but the BC/AD labels are widely used and understood." Some parts of the BBC use BCE/CE, but some presenters have said they will not. As at October 2019, the BBC News style guide has entries for AD and BC, but not for CE or BCE.
In June 2006, in the United States, the Kentucky State School Board reversed its decision to use BCE and CE in the state's new Program of Studies, leaving education of students about these concepts a matter of discretion at the local level.
Also in 2011, media reports suggested that the BC/AD notation in Australian school textbooks would be replaced by BCE/CE notation. The story became national news and drew opposition from some politicians and church leaders. Weeks after the story broke, the Australian Curriculum, Assessment and Reporting Authority denied the rumour and stated that the BC/AD notation would remain, with CE and BCE as an optional suggested learning activity.
In 2013 the Canadian Museum of Civilization (now the Canadian Museum of History) in Ottawa, which had previously switched to BCE/CE, decided to change back to BC/AD in material intended for the public, while retaining BCE/CE in academic content.
The style guide for The Guardian says, under the entry for CE/BCE: "some people prefer CE (common era, current era, or Christian era) and BCE (before common era, etc) to AD and BC, which, however, remain our style".
The use of CE in Jewish scholarship was historically motivated by the desire to avoid the implicit "Our Lord" in the abbreviation "AD". Although other aspects of dating systems are based in Christian origins, AD is a direct reference to Jesus as Lord.
Proponents of the Common Era notation assert that the use of BCE/CE shows sensitivity to those who use the same year numbering system as the one that originated with and is currently used by Christians, but who are not themselves Christian.
Former United Nations Secretary-General Kofi Annan has argued:
[T]he Christian calendar no longer belongs exclusively to Christians. People of all faiths have taken to using it simply as a matter of convenience. There is so much interaction between people of different faiths and cultures – different civilizations, if you like – that some shared way of reckoning time is a necessity. And so the Christian Era has become the Common Era.
Adena K. Berkowitz, when arguing at the Supreme Court, opted to use BCE and CE because "Given the multicultural society that we live in, the traditional Jewish designations – B.C.E. and C.E. – cast a wider net of inclusion".
Some oppose the Common Era notation for explicitly religious reasons. Because the BC/AD notation is based on the traditional year of the conception or birth of Jesus, some Christians are offended by the removal of the reference to him in era notation. The Southern Baptist Convention supports retaining the BC/AD abbreviations. Roman Catholic priest and writer on interfaith issues Raimon Panikkar argued that the BCE/CE usage is the less inclusive option as, in his view, using the designation BCE/CE is a "return... to the most bigoted Christian colonialism" towards non-Christians, who do not necessarily consider the time period following the beginning of the calendar to be a "common era".
There are also secular concerns. In 1993 the English language expert Kenneth G. Wilson speculated in his style guide that "if we do end by casting aside the AD/BC convention, almost certainly some will argue that we ought to cast aside as well the conventional numbering system [that is, the method of numbering years] itself, given its Christian basis." The short lived French Republican Calendar, for example, began with the first year of the French First Republic and rejected the seven-day week (with its connections to the Book of Genesis) for a ten-day week.
According to a "Los Angeles Times" report, it was a student's use of BCE/CE notation, inspired by its use within Wikipedia, which prompted the teacher and politician Andrew Schlafly to found Conservapedia, a cultural conservative wiki. One of its "Conservapedia Commandments" is that users must always apply BC/AD notation, since its sponsors perceive BCE/CE notation to "deny the historical basis" of the dating system.
The abbreviation BCE, just as with BC, always follows the year number. Unlike AD, which traditionally precedes the year number, CE always follows the year number (if context requires that it be written at all). Thus, the current year is written as in both notations (or, if further clarity is needed, as CE, or as AD ), and the year that Socrates died is represented as 399 BCE (the same year that is represented by 399 BC in the BC/AD notation). The abbreviations are sometimes written with small capital letters, or with periods (e.g., "B.C.E." or "C.E."). The US-based Society of Biblical Literature style guide for academic texts on religion prefers BCE/CE to BC/AD. | https://en.wikipedia.org/wiki?curid=6088 |
Charles Robert Malden
Charles Robert Malden (9 August 1797 – 23 May 1855), was a nineteenth-century British naval officer, surveyor and educator. He is the discoverer of Malden Island in the central Pacific, which is named in his honour. He also founded Windlesham House School at Brighton, England.
Malden was born in Putney, Surrey, son of Jonas Malden, a surgeon. He entered British naval service at the age of 11 on 22 June 1809. He served nine years as a volunteer 1st class, midshipman, and shipmate, including one year in the English Channel and Bay of Biscay (1809), four years at the Cape of Good Hope and in the East Indies (1809–14), two and a half years on the North American and West Indian stations (1814–16), and a year and a half in the Mediterranean (1817–18). He was present at the capture of Mauritius and Java, and at the battles of Baltimore and New Orleans.
He passed the examination in the elements of mathematics and the theory of navigation at the Royal Naval Academy on 2–4 September 1816, and became a 1st Lieutenant on 1 September 1818. In eight years of active service as an officer, he served two and a half years in a surveying ship in the Mediterranean (1818–21), one and a half years in a surveying sloop in the English Channel and off the coast of Ireland (1823–24), and one and a half years as Surveyor of the frigate during a voyage (1824–26) to and from the Hawaiian Islands (then known as the "Sandwich islands").
In Hawaii he surveyed harbours which, he noted, were "said not to exist by Captains Cook and Vancouver." On the return voyage he discovered and explored uninhabited Malden Island in the central Pacific on 30 July 1825. After his return he left active service but remained at half pay. He served for several years as hydrographer to King William IV.
He married Frances Cole, daughter of Rev. William Hodgson Cole, rector of West Clandon and Vicar of Wonersh, near Guildford, Surrey, on 8 April 1828. Malden became the father of seven sons and a daughter.
From 1830-36 he took pupils for the Royal Navy at Ryde, Isle of Wight. He purchased the school of Henry Worsley at Newport, Isle of Wight, in December 1836, reopened it as a preparatory school on 20 February 1837, and moved it to Montpelier Road in Brighton in December 1837. He built the Windlesham House School at Brighton in 1844, and conducted the school until his death there in 1855. | https://en.wikipedia.org/wiki?curid=6091 |
Chechnya
Chechnya (; ; , ), officially the Chechen Republic (; ; , ), is a federal subject (a republic) of the Russian Federation.
It is located in the North Caucasus, and within of the Caspian Sea. The capital of the republic is the city of Grozny. , the republic was reported to have a population of 1,268,989 people.
After the dissolution of the Soviet Union in 1991, the Chechen-Ingush ASSR was split into two parts: the Republic of Ingushetia and the Chechen Republic. The latter proclaimed the Chechen Republic of Ichkeria, which sought independence. Following the First Chechen War with Russia, Chechnya gained "de facto" independence as the Chechen Republic of Ichkeria. Russian federal control was restored during the Second Chechen War. Since then there has been a systematic reconstruction and rebuilding process, though sporadic fighting continues to take place in the mountains and southern regions into 2019.
According to Leonti Mroveli, the 11th-century Georgian chronicler, the word Caucasian is derived from the Vainakh ancestor Kavkas.
According to George Anchabadze of Ilia State University American linguist Johanna Nichols "has used language to connect the modern people of the Caucasus region to the ancient farmers of the Fertile Crescent" and her research suggests that "farmers of the region were proto-Nakh-Daghestanians." Nichols stated: "The Nakh–Dagestanian languages are the closest thing we have to a direct continuation of the cultural and linguistic community that gave rise to Western civilization." Henry Harpending of the University of Utah supports her claims.
Traces of human settlement that date back to 40,000 BC were found near Lake Kezanoi. Cave paintings, artifacts, and other archaeological evidence indicates continuous habitation for some 8,000 years. People living in these settlements used tools, fire, animal skins as clothing.
The Caucasian Epipaleolithic and early Caucasian Neolithic era, saw the introduction of agriculture, irrigation, and the domestication of animals in the region. Settlements near Ali-Yurt and Magas, discovered in modern times, revealed tools made out of stone: stone axes, polished stones, stone knives, stones with holes drilled in them, clay dishes etc. Settlements made out of clay bricks were discovered in the plains. In the mountains there were settlements made out of stone and surrounded by walls; some of them dated back to 8000 BC. This period also saw the invention of the wheel (3000 BC), horseback riding, metal works (copper, gold, silver, iron), dishes, armor, daggers, knives, arrow tips. The artifacts were found near Nasare-Cort, Muzhichi, Ja-E-Bortz (also known as Surkha-khi), Abbey-Gove (also known as Nazran or Nasare)
From 900–1200, the kingdom in the centre of the Caucasus was split into Alania and Noble Alania (Russian Царственные Аланы). The German scientist Peter Simon Pallas believed that Ingush people (Kist) were the direct descendants from Alania. In 1239, the Alania capital of Maghas was destroyed and Alan confederacy of the Northern Caucasian highlanders, nations, and tribes by Batu Khan (a Mongol leader and a grandson of Genghis Khan) "Magas was destroyed in the beginning of 1239 by the hordes of Batu Khan. Historically Magas was located at approximately the same place on which the new capital of Ingushetia is now built" – D.V.Zayats
In 1300–1400, there was frequent warfare between the Alans, Tamerlan, Tokhtamysh, culminating in the Battle of the Terek River. The Alan tribes built fortresses, castles, and defensive walls protecting the mountains from the invaders. Part of the lowland tribes were occupied by Mongols. The insurgency against Mongols began.
In 1991 the Jordanian historian Abdul-Ghani Khassan presented a photocopy from old Arabic scripts claiming that Alania was in Chechnya and Ingushetia, and the document from Alanian historian Azdin Vazzar (1395–1460) who claimed to be from Nokhcho (Chechen) tribe of Alania.
The 16th century saw the first Russian involvement in the Caucasus. 1558 Temryuk of Kabarda sent his emissaries to Moscow requesting help from Ivan the Terrible against the Vainakh tribes. Ivan the Terrible married Temryuk's daughter Maria Temryukovna. An alliance was formed to gain the ground in the central Caucasus for the expanding Tsardom of Russia against stubborn Vainakh defenders. Chechnya was a nation in the Northern Caucasus that fought against foreign rule continually since the 15th century. The Chechens converted over the next few centuries to Sunni Islam, as Islam was associated with resistance to Russian encroachment.
Peter I first sought to increase Russia's political influence in the Caucasus and the Caspian Sea at the expense of Safavid Persia when he launched the Russo-Persian War (1722–1723). Notable in Chechen history, this particular Russo-Persian War marked the first military encounter between Imperial Russia and the Vainakh. Russian forces succeeded in taking much of the Caucasian territories from Iran for several years.
As the Russians took control of the Caspian corridor and moved into Persian-ruled Dagestan, Peter's forces ran into mountain tribes. Peter sent a cavalry force to subdue them, but the Chechens routed them. In 1732, after Russia already ceded back most of the Caucasus to Persia, now led by Nader Shah, following the Treaty of Resht, Russian troops clashed again with Chechens in a village called Chechen-aul along the Argun River. The Russians were defeated again and withdrew, but this battle is responsible for the apocryphal story about how the Nokchi came to be known as "Chechens"-the people ostensibly named for the place the battle had taken place. The name Chechen was however already used since as early as 1692.
Under intermittent Persian rule since 1555, in 1783 the eastern Georgians of Kartl-Kakheti led by Erekle II and Russia signed the Treaty of Georgievsk. According to this treaty, Kartl-Kakheti received protection from Russia, and Georgia abjured any dependence on Iran. In order to increase its influence in the Caucasus and to secure communications with Kartli and other minority Christian regions of the Transcaucasia which it considered useful in its wars against Persia and Turkey, the Russian Empire began conquering the Northern Caucasus mountains. The Russian Empire used Christianity to justify its conquests, allowing Islam to spread widely because it positioned itself as the religion of liberation from tsardom, which viewed Nakh tribes as "bandits". The rebellion was led by Mansur Ushurma, a Chechen Naqshbandi (Sufi) sheikh—with wavering military support from other North Caucasian tribes. Mansur hoped to establish a Transcaucasus Islamic state under shari'a law. He was unable to fully achieve this because in the course of the war he was betrayed by the Ottomans, handed over to Russians, and executed in 1794.
Following the forced ceding of the current territories of Dagestan, most of Azerbaijan, and Georgia by Persia to Russia, following the Russo-Persian War (1804–1813) and its outcoming Treaty of Gulistan, Russia significantly widened its foothold in the Caucasus at the expense of Persia. Another successful Caucasus war against Persia several years later, starting in 1826 and ending in 1828 with the Treaty of Turkmenchay, and a successful war against Ottoman Turkey in 1828, enabled Russia to use a much larger portion of its army in subduing the natives of the North Caucasus.
The resistance of the Nakh tribes never ended and was a fertile ground for a new Muslim-Avar commander, Imam Shamil, who fought against the Russians from 1834 to 1859 (see Murid War). In 1859, Shamil was captured by Russians at aul Gunib. Shamil left Boysangur Benoiski, a Chechen with one arm, one eye, and one leg, in charge of command at Gunib. Benoiski broke through the siege and continued to fight Russia for another two years until he was captured and killed by Russians. The Russian tsar hoped that by sparing the life of Shamil, the resistance in the North Caucasus would stop, but it did not. Russia began to use a colonisation tactic by destroying Nakh settlements and building Cossack defense lines in the lowlands. The Cossacks suffered defeat after defeat and were constantly attacked by mountaineers, who were robbing them of food and weaponry.
The tsarists' regime used a different approach at the end of the 1860s. They offered Chechens and Ingush to leave the Caucasus for the Ottoman Empire (see Muhajir (Caucasus)). It is estimated that about 80% of Chechens and Ingush left the Caucasus during the deportation. It weakened the resistance which went from open warfare to insurgent warfare. One of the notable Chechen resistance fighters at the end of the 19th century was a Chechen abrek Zelimkhan Gushmazukaev and his comrade-in-arms Ingush abrek Sulom-Beck Sagopshinski. Together they built up small units which constantly harassed Russian military convoys, government mints, and government post-service, mainly in Ingushetia and Chechnya. Ingush aul Kek was completely burned when the Ingush refused to hand over Zelimkhan. Zelimkhan was killed at the beginning of the 20th century. The war between Nakh tribes and Russia resurfaced during the times of the Russian Revolution, which saw the Nakh struggle against Anton Denikin and later against the Soviet Union.
On December 21, 1917, Ingushetia, Chechnya, and Dagestan declared independence from Russia and formed a single state: "United Mountain Dwellers of the North Caucasus" (also known as the Mountainous Republic of the Northern Caucasus) which was recognised by major world powers. The capital of the new state was moved to Temir-Khan-Shura (Dagestan). Tapa Tchermoeff, a prominent Chechen statesman, was elected the first prime minister of the state. The second prime minister elected was Vassan-Girey Dzhabagiev, an Ingush statesman, who also was the author of the constitution of the republic in 1917, and in 1920 he was re-elected for the third term. In 1921 the Russians attacked and occupied the country and forcefully absorbed it into the Soviet state. The Caucasian war for independence restarted, and the government went into exile.
During Soviet rule, Chechnya and Ingushetia were combined to form Checheno-Ingush Autonomous Soviet Socialist Republic. In the 1930s Chechnya was flooded with many Ukrainians fleeing the Holodomor. As a result, many of the Ukrainians settled in Chechen-Ingush ASSR permanently and survived the famine.
Although over 50,000 Chechens and over 12,000 Ingush were fighting against Nazi Germany on the front line (including heroes of the USSR: Abukhadzhi Idrisov, Khanpasha Nuradilov, Movlid Visaitov), and although Nazi German troops were fought to a complete stop at two Chechen-Ingush ASSR cities Malgobek and Ordzhonikidze (renamed to Vladikavkaz) after capturing half of the Caucasus in less than a month; Chechens and Ingush were falsely accused as Nazi supporters and entire nations were deported during Operation Lentil to the Kazakh SSR (later Kazakhstan) in 1944 near the end of World War II where over 60% of Chechen and Ingush populations perished. American historian Norman Naimark writes: The deportation was supposedly justified by the materials prepared by notorious NKVD officer Bogdan Kobulov accusing Chechens and Ingush in a mass conspiracy preparing rebellion and providing assistance to the German forces. Many of the materials were later proved to be fabricated. Even distinguished Red Army officers who fought bravely against Germans (e.g. the commander of 255th Separate Chechen-Ingush regiment Movlid Visaitov, the first to contact American forces at Elbe river) were deported. There is a theory that the real reason why Chechens and Ingush were deported is the desire of Russia to attack Turkey, a non-communist country, as Chechens and Ingush could impede such plans. In 2004, the European Parliament recognised the deportation of Chechens and Ingush as an act of genocide.
The territory of the Chechen-Ingush Autonomous Soviet Socialist Republic was divided between Stavropol Krai (where Grozny Okrug was formed), the Dagestan ASSR, the North Ossetian ASSR, and the Georgian SSR.
The Chechens and Ingush were allowed to return to their land after 1956 during de-Stalinisation under Nikita Khrushchev when Chechen-Ingush Autonomous Soviet Socialist Republic was restored but both boundaries and ethnic composition of the territory significantly changed. There were many (predominantly Russian) migrants from other parts of the Soviet Union, who often settled in the abandoned family homes of Chechens and Ingushes. The republic lost its Prigorodny District which transferred to North Ossetian ASSR but gained predominantly Russian Naursky District and Shelkovskoy District that is considered the homeland for Terek Cossacks.
The Russification policies towards Chechens continued after 1956, with Russian language proficiency required in many aspects of life to provide Chechens better opportunities for advancement in the Soviet system.
On November 26, 1990, the Supreme Council of Chechen-Ingush ASSR adopted the "Declaration of State Sovereignty of the Chechen-Ingush Republic". This declaration was part of the reorganisation of the Soviet Union. This new treaty would have been signed August 22, 1991, which would have transformed 15 republic states into more than 80. The August 19–21, 1991 Soviet coup d'état attempt led to the abandonment of this reorganisation.
With the impending dissolution of the Soviet Union in 1991, an independence movement, the Chechen National Congress, was formed, led by ex-Soviet Air Force general and new Chechen President Dzhokhar Dudayev. It campaigned for the recognition of Chechnya as a separate nation. This movement was opposed by Boris Yeltsin's Russian Federation, which argued that Chechnya had not been an independent entity within the Soviet Union—as the Baltic, Central Asian, and other Caucasian States had—but was part of the Russian Soviet Federative Socialist Republic and hence did not have a right under the Soviet constitution to secede. It also argued that other republics of Russia, such as Tatarstan, would consider seceding from the Russian Federation if Chechnya were granted that right. Finally, it argued that Chechnya was a major hub in the oil infrastructure of Russia and hence its secession would hurt the country's economy and energy access.
In the ensuing decade, the territory was locked in an ongoing struggle between various factions, usually fighting unconventionally and forgoing the position held by the several successive Russian governments through the current administration.
The First Chechen War took place from 1994 to 1996, when Russian forces attempted to regain control over Chechnya, which had declared independence in November 1991. Despite overwhelming numerical superiority in men, weaponry, and air support, the Russian forces were unable to establish effective permanent control over the mountainous area due to numerous successful full-scale battles and insurgency raids. In three months, Russia lost more tanks (over 1,997 tanks) in Grozny than during the Battle of Berlin in 1945.
The Budyonnovsk hospital hostage crisis in 1995 shocked the Russian public and led to international condemnation of the Chechen rebels.
In April 1996 the first democratically elected president of Chechnya, Dzhokhar Dudayev, was killed by Russian forces using a booby trap bomb and a missile fired from a warplane after he was located by triangulating the position of a satellite phone he was using.
The widespread demoralisation of the Russian forces in the area and a successful offensive to re-take Grozny by Chechen rebel's forces led by Aslan Maskhadov prompted Russian President Boris Yeltsin to declare a ceasefire in 1996, and sign a peace treaty a year later that saw a withdrawal of Russian forces.
After the war, parliamentary and presidential elections took place in January 1997 in Chechnya and brought to power new President Aslan Maskhadov, chief of staff and prime minister in the Chechen coalition government, for a five-year term. Maskhadov sought to maintain Chechen sovereignty while pressing the Russian government to help rebuild the republic, whose formal economy and infrastructure were virtually destroyed. Russia continued to send money for the rehabilitation of the republic; it also provided pensions and funds for schools and hospitals. Most of these funds were taken by Chechen authorities and divided between favoured warlords. Nearly half a million people (40% of Chechnya's prewar population) had been internally displaced and lived in refugee camps or overcrowded villages. There was an economic downturn. Two Russian brigades were permanently stationed in Chechnya.
In lieu of the devastated economic structure, kidnapping emerged as the principal source of income countrywide, procuring over US$200 million during the three-year independence of the chaotic fledgling state, although victims were rarely killed. In 1998, 176 people were kidnapped, 90 of whom were released, according to official accounts. President Maskhadov started a major campaign against hostage-takers, and on October 25, 1998, Shadid Bargishev, Chechnya's top anti-kidnapping official, was killed in a remote-controlled car bombing. Bargishev's colleagues then insisted they would not be intimidated by the attack and would go ahead with their offensive. Political violence and religious extremism, blamed on "Wahhabism", was rife. In 1998, Grozny authorities declared a state of emergency. Tensions led to open clashes between the Chechen National Guard and Islamist militants, such as the July 1998 confrontation in Gudermes.
The War of Dagestan began on August 7, 1999, during which the Islamic International Brigade (IIPB) began an unsuccessful incursion into the neighboring Russian republic of Dagestan in favour of the Shura of Dagestan which sought independence from Russia. In September, a series of apartment bombs that killed around 300 people in several Russian cities, including Moscow, were blamed on the Chechen separatists. Some journalists contested the official explanation, instead blaming the Russian Secret Service for blowing up the buildings to initiate a new military campaign against Chechnya. In response to the bombings, a prolonged air campaign of retaliatory strikes against the Ichkerian regime and a ground offensive that began in October 1999 marked the beginning of the Second Chechen War. Much better organized and planned than the first Chechen War, the Russian military took control over most regions. The Russian forces used brutal force, killing 60 Chechen civilians during a mop-up operation in Aldy, Chechnya on February 5, 2000. After the re-capture of Grozny in February 2000, the Ichkerian regime fell apart.
Chechen rebels continued to fight Russian troops and conduct terrorist attacks. In October 2002, 40–50 Chechen rebels seized a Moscow theatre and took about 900 civilians hostage. The crisis ended with 117 hostages and up to 50 rebels dead, mostly due to an unknown aerosol pumped into the building by Russian special forces to incapacitate the people inside.
In response to the increasing terrorism, Russia tightened its grip on Chechnya and expanded its anti-terrorist operations throughout the region. Russia installed a pro-Russian Chechen regime. In 2003, a referendum was held on a constitution that reintegrated Chechnya within Russia but provided limited autonomy. According to the Chechen government, the referendum passed with 95.5% of the votes and almost 80% turnout. The Economist was skeptical of the results, arguing that "few outside the Kremlin regard the referendum as fair".
In September 2004, separatist rebels occupied a school in the town of Beslan, North Ossetia, demanding recognition of the independence of Chechnya and a Russian withdrawal. 1,100 people (including 777 children) were taken hostage. The attack lasted three days, resulting in the deaths of over 331 people, including 186 children. After the 2004 school siege, Russian president Vladimir Putin announced sweeping security and political reforms, sealing borders in the Caucasus region and revealing plans to give the central government more power. He also vowed to take tougher action against domestic terrorism, including preemptive strikes against Chechen separatists. In 2005 and 2006, separatist leaders Aslan Maskhadov and Shamil Basayev were killed.
Since 2007, Chechnya has been governed by Ramzan Kadyrov. Kadyrov's rule has been characterised by high-level corruption, a poor human rights record, widespread use of torture, and a growing cult of personality. Allegations of Anti-gay purges in Chechnya were initially reported on 1 April 2017.
In April 2009, Russia ended its counter-terrorism operation and pulled out the bulk of its army. The insurgency in the North Caucasus continued even after this date. The Caucasus Emirate has fully adopted the tenets of a Salafist jihadist group through its strict adherence to the Sunni Hanbali obedience to the literal interpretation of the Quran and the Sunnah.
Situated in the eastern part of the North Caucasus, partially in Eastern Europe, Chechnya is surrounded on nearly all sides by Russian Federal territory. In the west, it borders North Ossetia and Ingushetia, in the north, Stavropol Krai, in the east, Dagestan, and to the south, Georgia. Its capital is Grozny.
Rivers:
The Chechen Republic is divided into 15 districts and 3 cities of republican significance.
There are no true districts of Chechnya, but many believe that the different dialects of the Chechen language define different districts. The main dialects are:
Grozny, also known as the Dzhokhar dialect, is the dialect of people who live in and in some towns around Grozny.
Naskhish, a dialect spoken to the northeast of Chechnya. The most notable difference in this dialect is the addition of the letters "ȯ", "ј" and "є"
Day, pronounced like the word 'die' is spoken in a small section of the south, around and in the town of Day.
There are other dialects which are believed to define districts, but because these areas are so isolated, not much research has been done on these areas.
According to the 2010 Census, the population of the republic is 1,268,989, up from 1,103,686 recorded in the 2002 Census. That number has been questioned by demographers, who think such population growth after two deadly wars is highly implausible.
As of the 2010 Census, Chechens at 1,206,551 make up 95.3% of the republic's population. Other groups include Russians (24,382, or 1.9%), Kumyks (12,221, or 1%), Ingush (1,296 or 0.1%) and a host of smaller groups, each accounting for less than 0.5% of the total population. The Armenian community, which used to number around 15,000 in Grozny alone, has dwindled to a few families. The Armenian church of Grozny was demolished in 1930. The birth rate was 25.41 in 2004. (25.7 in Achkhoi Martan, 19.8 in Groznyy, 17.5 in Kurchaloi, 28.3 in Urus Martan and 11.1 in Vedeno). According to the Chechen State Statistical Committee, Chechnya's population had grown to 1.205 million in January 2006.
At the end of the Soviet era, ethnic Russians (including Cossacks) comprised about 23% of the population (269,000 in 1989).
According to some Russian sources, from 1991 to 1994 tens of thousands of people of non-Chechen ethnicity (mostly Russians, Ukrainians, and Armenians) left the republic amidst reports of violence and discrimination against the non-Chechen population, as well as widespread lawlessness and ethnic cleansing under the government of Dzhokhar Dudayev. However, Russian economists Boris Lvin and Andrei Iliaronov believe emigration of the Russian-speaking population from Chechnya was no more intense than that from Kalmykia, Tuva and Sakha-Yakutia.
The languages used in the Republic are Chechen and Russian. Chechen belongs to the Vaynakh or North-central Caucasian language family, which also includes Ingush and Batsb. Some scholars place it in a wider North Caucasian languages.
Note: TFR 2009–12 source.
(in the territory of modern Chechnya)
Islam
Islam is the predominant religion in Chechnya, practiced by 95% of those polled in Grozny in 2010. Chechens are overwhelmingly adherents to the Shafi'i Madhhab of Sunni Islam, the republic having converted to Islam between the 16th and the 19th centuries. Due to historical importance, many Chechens are Sufis, of either the Qadiri or Naqshbandi orders. Most of the population follows either the Shafi'i or the Hanafi, schools of jurisprudence, fiqh. The Shafi'i school of jurisprudence has a long tradition among the Chechens, and thus it remains the most practiced.
Following the end of the Soviet Union, there has been an Islamic revival in Chechnya, and in 2011 it was estimated that there were 465 mosques, including the Akhmad Kadyrov Mosque in Grozny accommodating 10,000 worshipers, as well 31 madrasas, including an Islamic university named Kunta-haji and a Center of Islamic Medicine in Grozny which is the largest such institution in Europe.
On 19 January 2015, 12 days after the Charlie Hebdo shooting, a march took place in Grozny against the publication of caricatures of the prophet Mohammed. The Chechen Ministry of Interior reported that more than a million people participated, while according to the sources of Caucasian Knot the number was between 350,000 and 500,000.
Christianity
The once-strong Russian minority in Chechnya, mostly Terek Cossacks and estimated as numbering approximately 25,000 in 2012, are predominantly Russian Orthodox, although presently only one church exists in Grozny. In August 2011, Archbishop Zosima of Vladikavkaz and Makhachkala performed the first mass baptism ceremony in the history of the Chechen Republic in the Terek River of Naursky District in which 35 citizens of Naursky and Shelkovsky districts were converted to Orthodoxy.
Since 1990, the Chechen Republic has had many legal, military, and civil conflicts involving separatist movements and pro-Russian authorities. Today, Chechnya is a relatively stable federal republic, although there is still some separatist movement activity. Its regional constitution entered into effect on April 2, 2003, after an all-Chechen referendum was held on March 23, 2003. Some Chechens were controlled by regional teips, or clans, despite the existence of pro- and anti-Russian political structures.
The former separatist religious leader (mufti) Akhmad Kadyrov, looked upon as a traitor by many separatists, was elected president with 83% of the vote in an internationally monitored election on October 5, 2003. Incidents of ballot stuffing and voter intimidation by Russian soldiers and the exclusion of separatist parties from the polls were subsequently reported by the Organization for Security and Co-operation in Europe (OSCE) monitors. On May 9, 2004, Kadyrov was assassinated in Grozny football stadium by a landmine explosion that was planted beneath a VIP stage and detonated during a parade, and Sergey Abramov was appointed to the position of acting prime minister after the incident. However, since 2005 Ramzan Kadyrov (son of Akhmad Kadyrov) has been the caretaker prime minister, and in 2007 was appointed as the new president. Many allege he is the wealthiest and most powerful man in the republic, with control over a large private militia referred to as the "Kadyrovtsy". The militia, which began as his father's security force, has been accused of killings and kidnappings by human rights organizations such as Human Rights Watch.
In 2009, the US government financed American organization Freedom House included Chechnya in the "Worst of the Worst" list of most repressive societies in the world, together with Burma, North Korea, Tibet, and others.
In addition to the Russian regional government, there was a separatist Ichkeria government that was not recognized by any state (although members have been given political asylum in European and Arab countries, as well as the United States).
Ichkeria was a member of the Unrepresented Nations and Peoples Organization between 1991 and 2010. Former president of Georgia Zviad Gamsakhurdia deposed in a military coup of 1991 and a participant of the Georgian Civil War, recognized the independence of the Chechen Republic of Ichkeria in 1993. Diplomatic relations with Ichkeria were also established by the partially recognized Islamic Emirate of Afghanistan under the Taliban government on January 16, 2000. This recognition ceased with the fall of the Taliban in 2001. However, despite Taliban recognition, there were no friendly relations between the Taliban and Ichkeria—Maskhadov rejected their recognition, stating that the Taliban were illegitimate. Ichkeria also received vocal support from the Baltic countries, a group of Ukrainian nationalists and Poland; Estonia once voted to recognize, but the act never was followed through due to pressure applied by both Russia and the EU.
The president of this government was Aslan Maskhadov, the Foreign Minister was Ilyas Akhmadov, who was the spokesman for Maskhadov. Aslan Maskhadov had been elected in an internationally monitored election in 1997 for 4 years, which took place after signing a peace agreement with Russia. In 2001 he issued a decree prolonging his office for one additional year; he was unable to participate in the 2003 presidential election since separatist parties were barred by the Russian government, and Maskhadov faced accusations of terrorist offenses in Russia. Maskhadov left Grozny and moved to the separatist-controlled areas of the south at the onset of the Second Chechen War. Maskhadov was unable to influence a number of warlords who retain effective control over Chechen territory, and his power was diminished as a result. Russian forces killed Maskhadov on March 8, 2005, and the assassination of Maskhadov was widely criticized since it left no legitimate Chechen separatist leader with whom to conduct peace talks. Akhmed Zakayev, Deputy Prime Minister and a Foreign Minister under Maskhadov, was appointed shortly after the 1997 election and is currently living under asylum in England. He and others chose Abdul Khalim Saidullayev, a relatively unknown Islamic judge who was previously the host of an Islamic program on Chechen television, to replace Maskhadov following his death. On June 17, 2006, it was reported that Russian special forces killed Abdul Khalim Saidullayev in a raid in a Chechen town Argun.
The successor of Saidullayev became Doku Umarov. On October 31, 2007, Umarov abolished the Chechen Republic of Ichkeria and its presidency and in its place proclaimed the Caucasus Emirate with himself as its Emir. This change of status has been rejected by many Chechen politicians and military leaders who continue to support the existence of the republic.
The Internal Displacement Monitoring Centre reports that after hundreds of thousands of ethnic Russians and Chechens fled their homes following inter-ethnic and separatist conflicts in Chechnya in 1994 and 1999, more than 150,000 people still remain displaced in Russia today.
On September 1, 1997, Criminal Code reportedly being implemented in the Chechen Republic-Ichkeriya, Article 148 punishes "anal sexual intercourse between a man and a woman or a man and a man". For first- and second-time offenders, the punishment is caning. A third conviction leads to the death penalty, which can be carried out in a number of ways including stoning or beheading.
Human rights groups criticized the conduct of the 2005 parliamentary elections as unfairly influenced by the central Russian government and military.
In 2006 Human Rights Watch reported that pro-Russian Chechen forces under the command of Ramzan Kadyrov, as well as federal police personnel, used torture to get information about separatist forces. "If you are detained in Chechnya, you face a real and immediate risk of torture. And there is little chance that your torturer will be held accountable", said Holly Cartner, Director of the Europe and Central Asia division of the Human Rights Watch.
On February 1, 2009, "The New York Times" released extensive evidence to support allegations of consistent torture and executions under the Kadyrov government. The accusations were sparked by the assassination in Austria of a former Chechen rebel who had gained access to Kadyrov's inner circle, 27-year-old Umar Israilov.
On July 1, 2009, Amnesty International released a detailed report covering the human rights violations committed by the Russian Federation against Chechen citizens. Among the most prominent features was that those abused had no method of redress against assaults, ranging from kidnapping to torture, while those responsible were never held accountable. This led to the conclusion that Chechnya was being ruled without law, being run into further devastating destabilization.
On March 10, 2011, Human Rights Watch reported that since Chechenization, the government has pushed for enforced Islamic dress code. The president Ramzan Kadyrov is quoted as saying "I have the right to criticize my wife. She doesn't. With us [in Chechen society], a wife is a housewife. A woman should know her place. A woman should give her love to us [men]... She would be [man's] property. And the man is the owner. Here, if a woman does not behave properly, her husband, father, and brother are responsible. According to our tradition, if a woman fools around, her family members kill her... That's how it happens, a brother kills his sister or a husband kills his wife... As a president, I cannot allow for them to kill. So, let women not wear shorts...". He has also openly defended honor killings on several occasions.
On July 9, 2017, Russian newspaper Novaya Gazeta reported that a number of people were subject to an extrajudicial execution on the night of January 26, 2017. It published 27 names of the people known to be dead, but stressed that the list is "not all [of those killed]"; the newspaper asserted that 50 people may have been killed in the execution. Some of the dead were gay, but not all; the deaths appeared to have been triggered by the death of a policeman, and according to the author of the report, Elena Milashina, were executed for terrorism.
In 2017, it was reported by Novaya Gazeta and human rights groups that Chechen authorities had allegedly set up concentration camps, one of which is in Argun, where gay men are interrogated and subjected to physical violence.
On 11 January 2019, it was reported that another 'gay purge' had begun in the country in December 2018, with several gay men and women being detained. The Russian LGBT Network believes that around 40 people were detained and two killed.
During the war, the Chechen economy fell apart. In 1994, the separatists planned to introduce a new currency, but the change did not occur due to the re-taking of Chechnya by Russian troops in the Second Chechen War.
The economic situation in Chechnya has improved considerably since 2000. According to the "New York Times", major efforts to rebuild Grozny have been made, and improvements in the political situation have led some officials to consider setting up a tourism industry, though there are claims that construction workers are being irregularly paid and that poor people have been displaced.
Chechnya's unemployment was 67% in 2006 and fell to 21.5% in 2014.
Total revenue of the budget of Chechnya for 2017 was 59.2 billion rubles. Of these, 48.5 billion rubles were so-called "gratuitous receipts" from the federal budget of the Russian Federation. | https://en.wikipedia.org/wiki?curid=6095 |
Carboxylic acid
A carboxylic acid is an organic compound that contains a carboxyl group (C(=O)OH) attached to an R-group. The general formula of a carboxylic acid is R–COOH, with R referring to the alkyl group. Carboxylic acids occur widely. Important examples include the amino acids and fatty acids. Deprotonation of a carboxylic acid gives a carboxylate anion.
Carboxylic acids are commonly identified by their trivial names. They often have the suffix "-ic acid". IUPAC-recommended names also exist; in this system, carboxylic acids have an "-oic acid" suffix. For example, butyric acid (C3H7CO2H) is butanoic acid by IUPAC guidelines. For nomenclature of complex molecules containing a carboxylic acid, the carboxyl can be considered position one of the parent chain even if there are other substituents, such as 3-chloropropanoic acid. Alternately, it can be named as a "carboxy" or "carboxylic acid" substituent on another parent structure, such as 2-carboxyfuran.
The carboxylate anion (R–COO− or RCO2−) of a carboxylic acid is usually named with the suffix "-ate", in keeping with the general pattern of "-ic acid" and "-ate" for a conjugate acid and its conjugate base, respectively. For example, the conjugate base of acetic acid is acetate.
Carbonic acid, which occurs in bicarbonate buffer systems in nature, is not generally classed as one of the carboxylic acids, despite that it has a moiety that looks like a COOH group.
Carboxylic acids are polar. Because they are both hydrogen-bond acceptors (the carbonyl –C=O) and hydrogen-bond donors (the hydroxyl –OH), they also participate in hydrogen bonding. Together, the hydroxyl and carbonyl group form the functional group carboxyl. Carboxylic acids usually exist as dimers in nonpolar media due to their tendency to "self-associate". Smaller carboxylic acids (1 to 5 carbons) are soluble in water, whereas higher carboxylic acids have limited solubility due to the increasing hydrophobic nature of the alkyl chain. These longer chain acids tend to be soluble in less-polar solvents such as ethers and alcohols. Aqueous sodium hydroxide and carboxylic acids, even hydrophobic ones, react to yield water-soluble sodium salts. For example, enathic acid has a low solubility in water (0.2 g/L), but its sodium salt is very soluble in water.
Carboxylic acids tend to have higher boiling points than water, because of their greater surface areas and their tendency to form stabilised dimers through hydrogen bonds. For boiling to occur, either the dimer bonds must be broken or the entire dimer arrangement must be vaporised, increasing the enthalpy of vaporization requirements significantly.
Carboxylic acids are Brønsted–Lowry acids because they are proton (H+) donors. They are the most common type of organic acid.
Carboxylic acids are typically weak acids, meaning that they only partially dissociate into H3O+ cations and RCOO− anions in neutral aqueous solution. For example, at room temperature, in a 1-molar solution of acetic acid, only 0.4% of the acid are dissociated. Electron-withdrawing substituents, such as -CF3 group, give stronger acids (the pKa of formic acid is 3.75 whereas trifluoroacetic acid, with a trifluoromethyl substituent, has a pKa of 0.23). Electron-donating substituents give weaker acids (the pKa of formic acid is 3.75 whereas acetic acid, with a methyl substituent, has a pKa of 4.76)
Deprotonation of carboxylic acids gives carboxylate anions; these are resonance stabilized, because the negative charge is delocalized over the two oxygen atoms, increasing the stability of the anion. Each of the carbon–oxygen bonds in the carboxylate anion has a partial double-bond character. The carbonyl carbon's partial positive charge is also weakened by the -1/2 negative charges on the 2 oxygen atoms.
Carboxylic acids often have strong sour odours. Esters of carboxylic acids tend to have pleasant odours, and many are used in perfume.
Carboxylic acids are readily identified as such by infrared spectroscopy. They exhibit a sharp band associated with vibration of the C–O vibration bond ("ν"C=O) between 1680 and 1725 cm−1. A characteristic "ν"O–H band appears as a broad peak in the 2500 to 3000 cm−1 region. By 1H NMR spectrometry, the hydroxyl hydrogen appears in the 10–13 ppm region, although it is often either broadened or not observed owing to exchange with traces of water.
Many carboxylic acids are produced industrially on a large scale. They are also frequently found in nature. Esters of fatty acids are the main components of lipids and polyamides of aminocarboxylic acids are the main components of proteins.
Carboxylic acids are used in the production of polymers, pharmaceuticals, solvents, and food additives. Industrially important carboxylic acids include acetic acid (component of vinegar, precursor to solvents and coatings), acrylic and methacrylic acids (precursors to polymers, adhesives), adipic acid (polymers), citric acid (a flavor and preservative in food and beverages), ethylenediaminetetraacetic acid (chelating agent), fatty acids (coatings), maleic acid (polymers), propionic acid (food preservative), terephthalic acid (polymers). Important carboxylate salts are soaps.
In general, industrial routes to carboxylic acids differ from those used on a smaller scale because they require specialized equipment.
Preparative methods for small scale reactions for research or for production of fine chemicals often employ expensive consumable reagents.
Many reactions produce carboxylic acids but are used only in specific cases or are mainly of academic interest.
The most widely practiced reactions convert carboxylic acids into esters, amides, carboxylate salts, acid chlorides, and alcohols. Carboxylic acids react with bases to form carboxylate salts, in which the hydrogen of the hydroxyl (–OH) group is replaced with a metal cation. For example, acetic acid found in vinegar reacts with sodium bicarbonate (baking soda) to form sodium acetate, carbon dioxide, and water:
Carboxylic acids also react with alcohols to give esters. This process is widely used, e.g. in the production of polyesters. Likewise, carboxylic acids are converted into amides, but this conversion typically does not occur by direct reaction of the carboxylic acid and the amine. Instead esters are typical precursors to amides. The conversion of amino acids into peptides is a significant biochemical process that requires ATP.
The hydroxyl group on carboxylic acids may be replaced with a chlorine atom using thionyl chloride to give acyl chlorides. In nature, carboxylic acids are converted to thioesters.
Like esters, most of carboxylic acid can be reduced to alcohols by hydrogenation or using hydride or alkyl transferring agents (since they will deprotonate the acids instead without transfer) such as lithium aluminium hydride or Grignard reagents (organolithium compounds).
"N","N"-Dimethyl(chloromethylene)ammonium chloride (ClHC=N+(CH3)2Cl−) is a highly chemoselective agent for carboxylic acid reduction. It selectively activates the carboxylic acid to give the carboxymethyleneammonium salt, which can be reduced by a mild reductant like lithium tris("t"-butoxy)aluminum hydride to afford an aldehyde in a one pot procedure. This procedure is known to tolerate reactive carbonyl functionalities such as ketone as well as moderately reactive ester, olefin, nitrile, and halide moieties.
The carboxyl radical, •COOH, only exists briefly. The acid dissociation constant of •COOH has been measured using electron paramagnetic resonance spectroscopy. The carboxyl group tends to dimerise to form oxalic acid. | https://en.wikipedia.org/wiki?curid=6099 |
Chernobyl
Chernobyl (, ), also called Chornobyl (), is a partially abandoned city in the Chernobyl Exclusion Zone, situated in the Ivankiv Raion of northern Kiev Oblast, Ukraine. Chernobyl is about north of Kiev, and southwest of the Belarusian city of Gomel. Before its evacuation, the city had about 14,000 residents, while around 1,000 people live in the city now.
The city was evacuated on 5 May 1986, 9 days after the nuclear accident at the nearby Chernobyl Nuclear Power Plant, which was the most disastrous nuclear accident in history, to that point in time. The plant is north of the Chernobyl city. Pripyat is a city much closer to the plant; it was built in the 1970s as a home for the power plant's workers, and its population of 50,000 was evacuated 30 hours after the accident. The city of Slavutych, built for those evacuated from Pripyat, also received the population evacuated from Chernobyl.
The city was the administrative center of Chernobyl Raion (district) from 1923. After the accident, in 1988, the raion was disestablished and administration was transferred to the neighboring Ivankiv Raion.
Although Chernobyl is mostly a ghost town today, a small number of animals and people still live there, in houses marked with signs that read, "Owner of this house lives here". Workers on watch and administrative personnel of the Chernobyl Exclusion Zone are also stationed in the city. There are two general stores, and a tourist hotel.
The city's name is the same as a Ukrainian name for "Artemisia vulgaris", mugwort or common wormwood, which is (or more commonly , 'common artemisia'). The name is inherited from or , a compound of + , the parts related to and , 'stalk', so named in distinction to the lighter-stemmed wormwood "A. absinthium".
The name in languages used nearby is:
The name in languages formerly used in the area is:
On the Ptolemy's world map there is the city of Azagarium which, according to "Dictionary of Ancient Geography" of Alexander Macbean (published in 1773 in London), is a town of Sarmatia Europaea on the Borysthenes, 36° East longitude and 50°40' latitude. Now supposed to be Czernobol, a town of Poland, in Red Russia, Palatinate of Kiow (see Kiev Voivodeship), not far from the Borystenes. On the map Azagarium is located on the right bank of Borystenes. It also was located just north of the city of Amadoca which was a capital of Amadocium situated between the lake (palus) of Amadoca (in the west) and the mountains (montes) of Amadoca (in the east).
Chernobyl was originally part of the land of Kievan Rus′. The first known mention of Chernobyl is from an 1193 charter, which describes it as a hunting-lodge of Knyaz Rurik Rostislavich. In the 13th century, it was a crown village of the Grand Duchy of Lithuania. The village was granted to Filon Kmita, a captain of the royal cavalry, as a fiefdom in 1566. The province where Chernobyl is located was transferred to the Kingdom of Poland in 1569, and later annexed by the Russian Empire in 1793. Prior to the 20th century, Chernobyl was inhabited by Ukrainian peasants, some Polish people and a relatively large number of Jews.
Jews were brought to Chernobyl by Filon Kmita, during the Polish campaign of colonization. After 1596, the traditionally Eastern Orthodox Ukrainian peasantry of the district were forcibly converted, by Poland, to the Greek Catholic Uniate religion. Many of these converts returned to Eastern Orthodoxy after the Partitions of Poland.
In 1626, during the Counter-reformation, the Dominican church and monastery were founded by Lukasz Sapieha. A group of Old Catholics opposed the decrees of the Council of Trent. In 1832, following the failed Polish November Uprising, the Dominican monastery was sequestrated. The church of the Old Catholics was disbanded in 1852.
Until the end of the 19th century, Chernobyl was a privately owned city that belonged to the Chodkiewicz family. In 1896 they sold the city to the state, but until 1910 they owned a castle and a house in the city.
In the second half of the 18th century, Chernobyl became a major center of Hasidic Judaism. The Chernobyl Hasidic dynasty had been founded by Rabbi Menachem Nachum Twersky. The Jewish population suffered greatly from pogroms in October 1905 and in March–April 1919; many Jews were killed or robbed at the instigation of the Russian nationalist Black Hundreds. When the Twersky Dynasty left Chernobyl in 1920, it ceased to exist as a center of Hasidism.
Chernobyl had a population of 10,800 in 1898, including 7,200 Jews. Chernobyl was occupied in World War I by German forces; Ukrainians and Bolsheviks fought over the city in the ensuing Civil War. In the Polish–Soviet War of 1919–20, Chernobyl was taken first by the Polish Army and then by cavalry of the Red Army. From 1921 onwards, it was officially incorporated into the Ukrainian SSR.
Between 1929 and 1933, Chernobyl suffered from killings during Stalin's collectivization campaign. It was also affected by the famine that resulted from Stalin's policies. The Polish and German community of Chernobyl was deported to Kazakhstan in 1936, during the Frontier Clearances.
During World War II, Chernobyl was occupied by the German Army from 25 August 1941 to 17 November 1943. The Jewish community was murdered during the Nazi occupation of 1941–44.
Twenty years later, the area was chosen as the site of the first nuclear power station to be built on Ukrainian soil. The Duga over-the-horizon radar array, several miles outside of Chernobyl, was the origin of the Russian Woodpecker; it was designed as part of an anti-ballistic missile early warning radar network. With the dissolution of the Soviet Union in 1991, Chernobyl remained part of Ukraine.
On 26 April 1986, one of the reactors at the Chernobyl Nuclear Power Plant exploded after unsanctioned experiments on the reactor by plant operators were done improperly. The resulting loss of control was due to design flaws of the RBMK reactor, which made it unstable when operated at low power, and prone to thermal runaway where increases in temperature increase reactor power output.
Chernobyl city was evacuated 9 days after the disaster. The level of contamination with caesium-137 was around 555 kBq/m2 (surface ground deposition in 1986).
Later analyses concluded that, even with very conservative estimates, relocation of the city (or of any area below 1500 kBq/m2) could not be justified on the grounds of radiological health.
This however does not account for the uncertainty in the first few days of the accident about further depositions and weather patterns.
Moreover, an earlier short-term evacuation could have averted more significant doses from short-lived isotope radiation (specifically iodine-131, which has a half-life of about eight days).
Estimates of health effects are a subject of some controversy, see Effects of the Chernobyl disaster.
In 1998, average caesium-137 doses from the accident (estimated at 1-2mSv per year) did not exceed those from other sources of exposure. Current effective caesium-137 dose rates as of 2019 are 200-250 nSv/h, or roughly 1.7-2.2 mSv per year,
which is comparable to the worldwide average background radiation from natural sources.
The base of operations for the administration and monitoring of the Chernobyl Exclusion Zone was moved from Pripyat to Chernobyl. Chernobyl currently contains offices for the State Agency of Ukraine on the Exclusion Zone Management and accommodations for visitors. Apartment blocks have been repurposed as accommodations for employees of the State Agency. The length of time that workers may spend within the Chernobyl Exclusion Zone is restricted by regulations that have been implemented to limit radiation exposure. Today, visits are allowed to Chernobyl but limited by strict rules.
In 2003, the United Nations Development Programme launched a project, called the Chernobyl Recovery and Development Programme (CRDP), for the recovery of the affected areas. The main goal of the CRDP's activities is supporting the efforts of the Government of Ukraine to mitigate the long-term social, economic, and ecological consequences of the Chernobyl disaster.
The city has become overgrown and many types of animals live there. According to census information collected over an extended period of time, it is estimated that more mammals live there now than before the disaster. | https://en.wikipedia.org/wiki?curid=6100 |
Baháʼí Faith
The Baháʼí Faith (; "") is a religion teaching the essential worth of all religions, and the unity of all people. Established by Baháʼu'lláh in 1863, it initially grew in Persia and parts of the Middle East, where it has faced ongoing persecution since its inception. It is estimated to have between 5 and 8 million adherents, known as Baháʼís, spread throughout most of the world's countries and territories.
The religion has three central figures: the Báb (1819–1850) is considered a herald who taught that God would soon send a prophet in the same way of Jesus or Muhammad and was executed by Iranian authorities in 1850; Baháʼu'lláh (1817-1892), who claimed to be that prophet in 1863 and faced exile and imprisonment for most of his life; and his son, ʻAbdu'l-Bahá (1844-1921) who was released from confinement in 1908 and made teaching trips to Europe and America. Following ʻAbdu'l-Bahá's death in 1921, leadership of the religion fell to his grandson Shoghi Effendi (1897–1957). Baháʼís around the world annually elect local, regional, and national Spiritual Assemblies that govern the affairs of the religion, and every five years the members of all National Spiritual Assemblies elect the Universal House of Justice, the nine-member supreme governing institution of the worldwide Baháʼí community, which sits in Haifa, Israel, near the Shrine of the Báb.
Baháʼí teachings are in some ways similar to other monotheistic faiths: God is considered single and all-powerful. However, Baháʼu'lláh taught that religion is orderly and progressively revealed by one God through Manifestations of God who are the founders of major world religions throughout history; Buddha, Jesus, and Muhammad being the most recent before the Báb and Baháʼu'lláh. Baháʼís regard the major religions as fundamentally unified in purpose, though varied in social practices and interpretations. There is a similar emphasis on the unity of all people, openly rejecting notions of racism and nationalism. At the heart of Baháʼí teachings is the goal of a unified world order that ensures the prosperity of all nations, races, creeds, and classes.
Letters written by Baháʼu'lláh to various individuals, including some heads of state, have been collected and assembled into a canon of Baháʼí scripture, including works by his son ʻAbdu'l-Bahá, and the Báb, who is regarded as Baháʼu'lláh's forerunner. Prominent among Baháʼí literature are the "Kitáb-i-Aqdas", "Kitáb-i-Íqán", "Some Answered Questions", and "The Dawn-Breakers".
In English-language use, the word "Baháʼí " is used either as an adjective to refer to the Baháʼí Faith or as a term for a follower of Baháʼu'lláh. It is derived from the Arabic (), meaning "glory" or "splendor".
The older term "Bahaʼism" (or "Bahaism") is still used, for example as a variant of "Bahai Faith" by the US Library of Congress, though it is now less common and the Baháʼí community prefers "Baháʼí Faith".
The teachings of Baháʼu'lláh, the founder of the Baháʼí Faith, form the foundation for Baháʼí belief. Three principles are central to these teachings: the unity of God, the unity of religion, and the unity of humanity. Baha'is believe that God periodically reveals his will through divine messengers, whose purpose is to transform the character of humankind and to develop, within those who respond, moral and spiritual qualities. Religion is thus seen as orderly, unified, and progressive from age to age.
The Baháʼí writings describe a single, personal, inaccessible, omniscient, omnipresent, imperishable, and almighty God who is the creator of all things in the universe. The existence of God and the universe is thought to be eternal, without a beginning or end. Though inaccessible directly, God is nevertheless seen as conscious of creation, with a will and purpose that is expressed through messengers termed Manifestations of God.
Baháʼí teachings state that God is too great for humans to fully comprehend, or to create a complete and accurate image of by themselves. Therefore, human understanding of God is achieved through his revelations via his Manifestations. In the Baháʼí religion, God is often referred to by titles and attributes (for example, the All-Powerful, or the All-Loving), and there is a substantial emphasis on monotheism. The Baháʼí teachings state that the attributes which are applied to God are used to translate Godliness into human terms and also to help individuals concentrate on their own attributes in worshipping God to develop their potentialities on their spiritual path. According to the Baháʼí teachings the human purpose is to learn to know and love God through such methods as prayer, reflection, and being of service to others.
Baháʼí notions of progressive religious revelation result in their accepting the validity of the well known religions of the world, whose founders and central figures are seen as Manifestations of God. Religious history is interpreted as a series of dispensations, where each "manifestation" brings a somewhat broader and more advanced revelation that is rendered as a text of scripture and passed on through history with greater or lesser reliability but at least true in substance, suited for the time and place in which it was expressed. Specific religious social teachings (for example, the direction of prayer, or dietary restrictions) may be revoked by a subsequent manifestation so that a more appropriate requirement for the time and place may be established. Conversely, certain general principles (for example, neighbourliness, or charity) are seen to be universal and consistent. In Baháʼí belief, this process of progressive revelation will not end; it is, however, believed to be cyclical. Baháʼís do not expect a new manifestation of God to appear within 1000 years of Baháʼu'lláh's revelation.
Baháʼí beliefs are sometimes described as syncretic combinations of earlier religious beliefs. Baháʼís, however, assert that their religion is a distinct tradition with its own scriptures, teachings, laws, and history. While the religion was initially seen as a sect of Islam, most religious specialists now see it as an independent religion, with its religious background in Shiʻa Islam being seen as analogous to the Jewish context in which Christianity was established. Muslim institutions and clergy, both Sunni and Shia, consider Baháʼís to be deserters or apostates from Islam, which has led to Baháʼís being persecuted. Baháʼís describe their faith as an independent world religion, differing from the other traditions in its relative age and in the appropriateness of Baháʼu'lláh's teachings to the modern context. Baháʼu'lláh is believed to have fulfilled the messianic expectations of these precursor faiths.
The Baháʼí writings state that human beings have a "rational soul", and that this provides the species with a unique capacity to recognize God's status and humanity's relationship with its creator. Every human is seen to have a duty to recognize God through his messengers, and to conform to their teachings. Through recognition and obedience, service to humanity and regular prayer and spiritual practice, the Baháʼí writings state that the soul becomes closer to God, the spiritual ideal in Baháʼí belief. According to Baháʼí belief when a human dies the soul is permanently separated from the body and carries on in the next world where it is judged based on the person's actions in the physical world. Heaven and Hell are taught to be spiritual states of nearness or distance from God that describe relationships in this world and the next, and not physical places of reward and punishment achieved after death.
The Baháʼí writings emphasize the essential equality of human beings, and the abolition of prejudice. Humanity is seen as essentially one, though highly varied; its diversity of race and culture are seen as worthy of appreciation and acceptance. Doctrines of racism, nationalism, caste, social class, and gender-based hierarchy are seen as artificial impediments to unity. The Baháʼí teachings state that the unification of humanity is the paramount issue in the religious and political conditions of the present world.
Shoghi Effendi, the head of the religion from 1921 to 1957, wrote the following summary of what he considered to be the distinguishing principles of Baháʼu'lláh's teachings, which, he said, together with the laws and ordinances of the "Kitáb-i-Aqdas" constitute the bedrock of the Baháʼí Faith:
The following principles are frequently listed as a quick summary of the Baháʼí teachings. They are derived from transcripts of speeches given by ʻAbdu'l-Bahá during his tour of Europe and North America in 1912. The list is not authoritative and a variety of such lists circulate.
With specific regard to the pursuit of world peace, Baháʼu'lláh prescribed a world-embracing collective security arrangement for the establishment of a temporary era of peace referred to in the Baha'i teachings as the Lesser Peace. For the establishment of a lasting peace (The Most Great Peace) and the purging of the "overwhelming Corruptions" it is necessary that all the people of the world universally unite under a universal Faith.
The Baháʼí teachings speak of a "Greater Covenant", being universal and endless, and a "Lesser Covenant", being unique to each religious dispensation. The Lesser Covenant is viewed as an agreement between a Messenger of God and his followers and includes social practices and the continuation of authority in the religion. At this time Baháʼís view Baháʼu'lláh's revelation as a binding lesser covenant for his followers; in the Baháʼí writings being firm in the covenant is considered a virtue to work toward. The Greater Covenant is viewed as a more enduring agreement between God and humanity, where a Manifestation of God is expected to come to humanity about every thousand years, at times of turmoil and uncertainty.
With unity as an essential teaching of the religion, Baháʼís follow an administration they believe is divinely ordained, and therefore see attempts to create schisms and divisions as efforts that are contrary to the teachings of Baháʼu'lláh. Schisms have occurred over the succession of authority, but any Baháʼí divisions have had relatively little success and have failed to attract a sizeable following. The followers of such divisions are regarded as Covenant-breakers and shunned, essentially excommunicated.
The "canonical texts" are the writings of the Báb, Baháʼu'lláh, ʻAbdu'l-Bahá, Shoghi Effendi and the Universal House of Justice, and the authenticated talks of ʻAbdu'l-Bahá. The writings of the Báb and Baháʼu'lláh are considered as divine revelation, the writings and talks of ʻAbdu'l-Bahá and the writings of Shoghi Effendi as authoritative interpretation, and those of the Universal House of Justice as authoritative legislation and elucidation. Some measure of divine guidance is assumed for all of these texts. Some of Baháʼu'lláh's most important writings include the Kitáb-i-Aqdas, literally the "Most Holy Book", which defines many laws and practices for individuals and society, the Kitáb-i-Íqán, literally the "Book of Certitude", which became the foundation of much of Baháʼí belief, the Gems of Divine Mysteries, which includes further doctrinal foundations, and the Seven Valleys and the Four Valleys which are mystical treatises.
Although the Baháʼí teachings have a strong emphasis on social and ethical issues, a number of foundational texts have been described as mystical. "The Seven Valleys" is considered Baháʼu'lláh's "greatest mystical composition." It was written to a follower of Sufism, in the style of ʻAttar, the Persian Muslim poet, and sets forth the stages of the soul's journey towards God. It was first translated into English in 1906, becoming one of the earliest available books of Baháʼu'lláh to the West. "The Hidden Words" is another book written by Baháʼu'lláh during the same period, containing 153 short passages in which Baháʼu'lláh claims to have taken the basic essence of certain spiritual truths and written them in brief form.
The Baháʼí Faith formed from the Iranian religion of the Báb, a merchant who began preaching a new interpretation of Shia Islam in 1844. The Báb's claim to divine revelation was rejected by the generality of Islamic clergy in Iran, ending in his public execution by authorities in 1850. The Báb taught that God would soon send a new messenger, and Baháʼís consider Baháʼu'lláh to be that person. Although they are distinct movements, the Báb is so interwoven into Baháʼí theology and history that Baháʼís celebrate his birth, death, and declaration as holy days, consider him one of their three central figures (along with Baháʼu'lláh and ʻAbdu'l-Bahá), and a historical account of the Bábí movement ("The Dawn-Breakers") is considered one of three books that every Baháʼí should "master" and read "over and over again".
The Baháʼí community was mostly confined to the Iranian and Ottoman empires until after the death of Baháʼu'lláh in 1892, at which time he had followers in 13 countries of Asia and Africa. Under the leadership of his son, ʻAbdu'l-Bahá, the religion gained a footing in Europe and America, and was consolidated in Iran, where it still suffers intense persecution. After the death of ʻAbdu'l-Bahá in 1921, the leadership of the Baháʼí community entered a new phase, evolving from a single individual to an administrative order with both elected bodies and appointed individuals.
On the evening of 22 May 1844, Siyyid ʻAlí-Muhammad of Shiraz gained his first convert and took on the title of "the Báb" ( "the Gate"), referring to his later claim to the status of Mahdi of Shiʻa Islam. His followers were therefore known as Bábís. As the Báb's teachings spread, which the Islamic clergy saw as blasphemous, his followers came under increased persecution and torture. The conflicts escalated in several places to military sieges by the Shah's army. The Báb himself was imprisoned and eventually executed in 1850.
Baháʼís see the Báb as the forerunner of the Baháʼí Faith, because the Báb's writings introduced the concept of "He whom God shall make manifest", a Messianic figure whose coming, according to Baháʼís, was announced in the scriptures of all of the world's great religions, and whom Baháʼu'lláh, the founder of the Baháʼí Faith, claimed to be in 1863. The Báb's tomb, located in Haifa, Israel, is an important place of pilgrimage for Baháʼís. The remains of the Báb were brought secretly from Iran to the Holy Land and eventually interred in the tomb built for them in a spot specifically designated by Baháʼu'lláh. The main written works translated into English of the Báb's are collected in Selections from the Writings of the Báb out of the estimated 135 works.
Mírzá Husayn ʻAlí Núrí was one of the early followers of the Báb, and later took the title of Baháʼu'lláh. Bábís faced a period of persecution that peaked in 1852–53 after a few individuals made a failed attempt to assassinate the Shah. Although they acted alone, the government responded with collective punishment, killing many Bábís. Baháʼu'lláh was put in prison. He claimed that in 1853, while incarcerated in the dungeon of the Síyáh-Chál in Tehran, he received the first intimations that he was the one anticipated by the Báb when he received a visit from the Maid of Heaven.
Shortly thereafter he was expelled from Tehran to Baghdad, in the Ottoman Empire; then to Constantinople (now Istanbul); and then to Adrianople (now Edirne). In 1863, at the time of his banishment from Baghdad to Constantinople, Baháʼu'lláh declared his claim to a divine mission to his family and followers. Tensions then grew between him and Subh-i-Azal, the appointed leader of the Bábís who did not recognize Baháʼu'lláh's claim. Throughout the rest of his life Baháʼu'lláh gained the allegiance of most of the Bábís, who came to be known as Baháʼís. Beginning in 1866, he began declaring his mission as a Messenger of God in letters to the world's religious and secular rulers, including Pope Pius IX, Napoleon III, and Queen Victoria.
He produced over 18,000 works in his lifetime, in both Arabic and Persian, of which only 8% have been translated into English.
Baháʼu'lláh was banished by Sultan Abdülaziz a final time in 1868 to the Ottoman penal colony of ʻAkká, in present-day Israel. Towards the end of his life, the strict and harsh confinement was gradually relaxed, and he was allowed to live in a home near ʻAkká, while still officially a prisoner of that city. He died there in 1892. Baháʼís regard his resting place at Bahjí as the Qiblih to which they turn in prayer each day.
ʻAbbás Effendi was Baháʼu'lláh's eldest son, known by the title of ʻAbdu'l-Bahá (Servant of Bahá). His father left a will that appointed ʻAbdu'l-Bahá as the leader of the Baháʼí community, and designated him as the "Centre of the Covenant", "Head of the Faith", and the sole authoritative interpreter of Baháʼu'lláh's writings. ʻAbdu'l-Bahá had shared his father's long exile and imprisonment, which continued until ʻAbdu'l-Bahá's own release as a result of the Young Turk Revolution in 1908. Following his release he led a life of travelling, speaking, teaching, and maintaining correspondence with communities of believers and individuals, expounding the principles of the Baháʼí Faith.
There are over 27,000 extant documents by ʻAbdu'l-Bahá, mostly letters, of which only a fraction have been translated into English. Among the more well known are "The Secret of Divine Civilization", the "Tablet to Auguste-Henri Forel", and "Some Answered Questions". Additionally notes taken of a number of his talks were published in various volumes like "Paris Talks" during his journeys to the West.
Baháʼu'lláh's "Kitáb-i-Aqdas" and "The Will and Testament of ʻAbdu'l-Bahá" are foundational documents of the Baháʼí administrative order. Baháʼu'lláh established the elected Universal House of Justice, and ʻAbdu'l-Bahá established the appointed hereditary Guardianship and clarified the relationship between the two institutions. In his Will, ʻAbdu'l-Bahá appointed Shoghi Effendi, his eldest grandson, as the Guardian of the Baháʼí Faith. Shoghi Effendi served for 36 years as the head of the religion until his death.
Shoghi Effendi throughout his lifetime translated Baháʼí texts; developed global plans for the expansion of the Baháʼí community; developed the Baháʼí World Centre; carried on a voluminous correspondence with communities and individuals around the world; and built the administrative structure of the religion, preparing the community for the election of the Universal House of Justice. He unexpectedly died after a brief illness on 4 November 1957, in London, England, under conditions that did not allow for a successor to be appointed.
In 1937, Shoghi Effendi launched a seven-year plan for the Baháʼís of North America, followed by another in 1946. In 1953, he launched the first international plan, the Ten Year World Crusade. This plan included extremely ambitious goals for the expansion of Baháʼí communities and institutions, the translation of Baháʼí texts into several new languages, and the sending of Baháʼí pioneers into previously unreached nations. He announced in letters during the Ten Year Crusade that it would be followed by other plans under the direction of the Universal House of Justice, which was elected in 1963 at the culmination of the Crusade. The House of Justice then launched a nine-year plan in 1964, and a series of subsequent multi-year plans of varying length and goals followed, guiding the direction of the international Baháʼí community.
Since 1963, the Universal House of Justice has been the elected head of the Baháʼí Faith. The general functions of this body are defined through the writings of Baháʼu'lláh and clarified in the writings of Abdu'l-Bahá and Shoghi Effendi. These functions include teaching and education, implementing Baháʼí laws, addressing social issues, and caring for the weak and the poor.
The House of Justice directs the work of the Baháʼí community through a series of multi-year international plans that began with a nine-year plan in 1964. In the current plan, the House of Justice encourages the Baháʼís around the world to focus on capacity building through children's classes, junior youth groups, devotional gatherings, and study circles. Additional lines of action include social action and participation in the prevalent discourses of society. The years from 2001 until 2021 represent four successive five-year plans, culminating in the centennial anniversary of the passing of ʻAbdu'l-Bahá. Annually, on 21 April, the Universal House of Justice sends a 'Ridván' message to the worldwide Baháʼí community, that updates Baháʼís on current developments and provides further guidance for the year to come.
At local, regional, and national levels, Baháʼís elect members to nine-person Spiritual Assemblies, which run the affairs of the religion. There are also appointed individuals working at various levels, including locally and internationally, which perform the function of propagating the teachings and protecting the community. The latter do not serve as clergy, which the Baháʼí Faith does not have. The Universal House of Justice remains the supreme governing body of the Baháʼí Faith, and its 9 members are elected every five years by the members of all National Spiritual Assemblies. Any male Baháʼí, 21 years or older, is eligible to be elected to the Universal House of Justice; all other positions are open to male and female Baháʼís.
A Baháʼí-published document reported 4.74 million Baháʼís in 1986 growing at a rate of 4.4%. Baháʼí sources since 1991 usually estimate the worldwide Baháʼí population to be above 5 million. The "World Christian Encyclopedia" estimated 7.1 million Baháʼís in the world in 2000, representing 218 countries, and 7.3 million in 2010 with the same source. They further state: "The Baha'i Faith is the only religion to have grown faster in every United Nations region over the past 100 years than the general population; Bahaʼi was thus the fastest-growing religion between 1910 and 2010, growing at least twice as fast as the population of almost every UN region." This source's only systematic flaw was to consistently have a higher estimate of Christians than other cross-national data sets.
The Baháʼí Faith is currently the largest religious minority in Iran, Panama, Belize, and South Carolina; the second largest international religion in Bolivia, Zambia, and Papua New Guinea; and the third largest international religion in Chad and Kenya.
According to "The World Almanac and Book of Facts 2004":
The Baháʼí Faith is a medium-sized religion and was listed in "The Britannica Book of the Year" (1992–present) as the second most widespread of the world's independent religions in terms of the number of countries represented. According to "Britannica", the Baháʼí Faith (as of 2010) is established in 221 countries and territories and has an estimated seven million adherents worldwide. Additionally, Baháʼís have self-organized in most of the nations of the world.
The Baháʼí religion was ranked by the Foreign Policy magazine as the world's second fastest growing religion by percentage (1.7%) in 2007.
The following are a few examples from Baháʼu'lláh's teachings on personal conduct that are required or encouraged of his followers:
The following are a few examples from Baháʼu'lláh's teachings on personal conduct that are prohibited or discouraged.
While some of the laws from the Kitáb-i-Aqdas are applicable at the present time, others are dependent upon the existence of a predominantly Baháʼí society, such as the punishments for arson or murder. The laws, when not in direct conflict with the civil laws of the country of residence, are binding on every Baháʼí, and the observance of personal laws, such as prayer or fasting, is the sole responsibility of the individual.
The purpose of marriage in the Baháʼí faith is mainly to foster spiritual harmony, fellowship and unity between a man and a woman and to provide a stable and loving environment for the rearing of children. The Baháʼí teachings on marriage call it a "fortress for well-being and salvation" and place marriage and the family as the foundation of the structure of human society. Baháʼu'lláh highly praised marriage, discouraged divorce, and required chastity outside of marriage; Baháʼu'lláh taught that a husband and wife should strive to improve the spiritual life of each other. Interracial marriage is also highly praised throughout Baháʼí scripture.
Baháʼís intending to marry are asked to obtain a thorough understanding of the other's character before deciding to marry. Although parents should not choose partners for their children, once two individuals decide to marry, they must receive the consent of all living biological parents, whether they are Baháʼí or not. The Baháʼí marriage ceremony is simple; the only compulsory part of the wedding is the reading of the wedding vows prescribed by Baháʼu'lláh which both the groom and the bride read, in the presence of two witnesses. The vows are "We will all, verily, abide by the Will of God."
Baháʼu'lláh prohibited a mendicant and ascetic lifestyle. Monasticism is forbidden, and Baháʼís are taught to practice spirituality while engaging in useful work. The importance of self-exertion and service to humanity in one's spiritual life is emphasised further in Baháʼu'lláh's writings, where he states that work done in the spirit of service to humanity enjoys a rank equal to that of prayer and worship in the sight of God.
Most Baháʼí meetings occur in individuals' homes, local Baháʼí centers, or rented facilities. Worldwide, as of 2018, ten Baháʼí Houses of Worship, including eight Mother Temples and two local Houses of Worship have been built and a further five are planned for construction. Two of these houses of worship are national while the other three are going to be local temples. Baháʼí writings refer to an institution called a "Mas͟hriqu'l-Ad͟hkár" (Dawning-place of the Mention of God), which is to form the center of a complex of institutions including a hospital, university, and so on. The first ever Mas͟hriqu'l-Ad͟hkár in ʻIshqábád, Turkmenistan, has been the most complete House of Worship.
The Baháʼí calendar is based upon the calendar established by the Báb. The year consists of 19 months, each having 19 days, with four or five intercalary days, to make a full solar year. The Baháʼí New Year corresponds to the traditional Iranian New Year, called Naw Rúz, and occurs on the vernal equinox, near 21 March, at the end of the month of fasting. Baháʼí communities gather at the beginning of each month at a meeting called a Feast for worship, consultation and socializing.
Each of the 19 months is given a name which is an attribute of God; some examples include Baháʼ (Splendour), ʻIlm (Knowledge), and Jamál (Beauty). The Baháʼí week is familiar in that it consists of seven days, with each day of the week also named after an attribute of God. Baháʼís observe 11 Holy Days throughout the year, with work suspended on 9 of these. These days commemorate important anniversaries in the history of the religion.
The symbols of the religion are derived from the Arabic word Baháʼ ( "splendor" or "glory"), with a numerical value of 9, which is why the most common symbol is the nine-pointed star. The ringstone symbol and calligraphy of the Greatest Name are also often encountered. The former consists of two five-pointed stars interspersed with a stylized Baháʼ whose shape is meant to recall the three onenesses, while the latter is a calligraphic rendering of the phrase Yá Baháʼu'l-Abhá ( "O Glory of the Most Glorious!").The five-pointed star is the official symbol of the Baháʼí Faith, known as the "Haykal" ("temple"). It was initiated and established by the Báb and various works were written in calligraphy shaped into a five-pointed star.
Since its inception the Baháʼí Faith has had involvement in socio-economic development beginning by giving greater freedom to women, promulgating the promotion of female education as a priority concern, and that involvement was given practical expression by creating schools, agricultural co-ops, and clinics.
The religion entered a new phase of activity when a message of the Universal House of Justice dated 20 October 1983 was released. Baháʼís were urged to seek out ways, compatible with the Baháʼí teachings, in which they could become involved in the social and economic development of the communities in which they lived. Worldwide in 1979 there were 129 officially recognized Baháʼí socio-economic development projects. By 1987, the number of officially recognized development projects had increased to 1482.
Current initiatives of social action include activities in areas like health, sanitation, education, gender equality, arts and media, agriculture, and the environment. Educational projects include schools, which range from village tutorial schools to large secondary schools, and some universities. By 2017 there were an estimated 40,000 small scale projects, 1,400 sustained projects, and 135 Baháʼí inspired organizations.
Baháʼu'lláh wrote of the need for world government in this age of humanity's collective life. Because of this emphasis the international Baháʼí community has chosen to support efforts of improving international relations through organizations such as the League of Nations and the United Nations, with some reservations about the present structure and constitution of the UN. The Baháʼí International Community is an agency under the direction of the Universal House of Justice in Haifa, and has consultative status with the following organizations:
The Baháʼí International Community has offices at the United Nations in New York and Geneva and representations to United Nations regional commissions and other offices in Addis Ababa, Bangkok, Nairobi, Rome, Santiago, and Vienna. In recent years, an Office of the Environment and an Office for the Advancement of Women were established as part of its United Nations Office. The Baháʼí Faith has also undertaken joint development programs with various other United Nations agencies. In the 2000 Millennium Forum of the United Nations a Baháʼí was invited as the only non-governmental speaker during the summit.
Baháʼís continue to be persecuted in Islamic countries, as Islamic leaders do not recognize the Baháʼí Faith as an independent religion, but rather as apostasy from Islam. The most severe persecutions have occurred in Iran, where more than 200 Baháʼís were executed between 1978 and 1998, and in Egypt. The rights of Baháʼís have been restricted to greater or lesser extents in numerous other countries, including Afghanistan, Indonesia, Iraq, Morocco, Yemen, and several countries in sub-Saharan Africa.
The marginalization of the Iranian Baháʼís by current governments is rooted in historical efforts by Muslim clergy to persecute the religious minority. When the Báb started attracting a large following, the clergy hoped to stop the movement from spreading by stating that its followers were enemies of God. These clerical directives led to mob attacks and public executions. Starting in the twentieth century, in addition to repression aimed at individual Baháʼís, centrally directed campaigns that targeted the entire Baháʼí community and its institutions were initiated. In one case in Yazd in 1903 more than 100 Baháʼís were killed. Baháʼí schools, such as the Tarbiyat boys' and girls' schools in Tehran, were closed in the 1930s and 1940s, Baháʼí marriages were not recognized and Baháʼí texts were censored.
During the reign of Mohammad Reza Pahlavi, to divert attention from economic difficulties in Iran and from a growing nationalist movement, a campaign of persecution against the Baháʼís was instituted. An approved and coordinated anti-Baháʼí campaign (to incite public passion against the Baháʼís) started in 1955 and it included the spreading of anti-Baháʼí propaganda on national radio stations and in official newspapers. During that campaign, initiated by Mulla Muhammad Taghi Falsafi, the Baha'i center in Tehran was demolished at the orders of Tehran military governor, General Timor Bakhtiar. In the late 1970s the Shah's regime consistently lost legitimacy due to criticism that it was pro-Western. As the anti-Shah movement gained ground and support, revolutionary propaganda was spread which alleged that some of the Shah's advisors were Baháʼís. Baháʼís were portrayed as economic threats, and as supporters of Israel and the West, and societal hostility against the Baháʼís increased.
Since the Islamic Revolution of 1979 Iranian Baháʼís have regularly had their homes ransacked or have been banned from attending university or from holding government jobs, and several hundred have received prison sentences for their religious beliefs, most recently for participating in study circles. Baháʼí cemeteries have been desecrated and property has been seized and occasionally demolished, including the House of Mírzá Buzurg, Baháʼu'lláh's father. The House of the Báb in Shiraz, one of three sites to which Baháʼís perform pilgrimage, has been destroyed twice.
In May 2018, the Iranian authorities expelled a young woman student from university of Isfahan because she was Baháʼí. In March 2018, two more Baháʼí students were expelled from universities in the cities of Zanjan and Gilan because of their religion.
According to a US panel, attacks on Baháʼís in Iran increased under Mahmoud Ahmadinejad's presidency. The United Nations Commission on Human Rights revealed an October 2005 confidential letter from Command Headquarters of the Armed Forces of Iran ordering its members to identify Baháʼís and to monitor their activities. Due to these actions, the Special Rapporteur of the United Nations Commission on Human Rights stated on 20 March 2006, that she "also expresses concern that the information gained as a result of such monitoring will be used as a basis for the increased persecution of, and discrimination against, members of the Baháʼí faith, in violation of international standards. The Special Rapporteur is concerned that this latest development indicates that the situation with regard to religious minorities in Iran is, in fact, deteriorating."
On 14 May 2008, members of an informal body known as the "Friends" that oversaw the needs of the Baháʼí community in Iran were arrested and taken to Evin prison. The Friends court case has been postponed several times, but was finally underway on 12 January 2010. Other observers were not allowed in the court. Even the defence lawyers, who for two years have had minimal access to the defendants, had difficulty entering the courtroom. The chairman of the U.S. Commission on International Religious Freedom said that it seems that the government has already predetermined the outcome of the case and is violating international human rights law. Further sessions were held on 7 February 2010, 12 April 2010 and 12 June 2010. On 11 August 2010 it became known that the court sentence was 20 years imprisonment for each of the seven prisoners which was later reduced to ten years. After the sentence, they were transferred to Gohardasht prison. In March 2011 the sentences were reinstated to the original 20 years. On 3 January 2010, Iranian authorities detained ten more members of the Baha'i minority, reportedly including Leva Khanjani, granddaughter of Jamaloddin Khanjani, one of seven Baha'i leaders jailed since 2008 and in February, they arrested his son, Niki Khanjani.
The Iranian government claims that the Baháʼí Faith is not a religion, but is instead a political organization, and hence refuses to recognize it as a minority religion. However, the government has never produced convincing evidence supporting its characterization of the Baháʼí community. Also, the government's statements that Baháʼís who recanted their religion would have their rights restored, attest to the fact that Baháʼís are persecuted solely for their religious affiliation. The Iranian government also accuses the Baháʼí Faith of being associated with Zionism. These accusations against the Baháʼís have no basis in historical fact, and the accusations are used by the Iranian government to use the Baháʼís as "scapegoats". In fact it was the Iranian leader Naser al-Din Shah Qajar who banished Baháʼu'lláh from Iran to the Ottoman Empire and Baháʼu'lláh was later exiled by the Ottoman Sultan, at the behest of the Iranian Shah, to territories further away from Iran and finally to Acre in Syria, which only a century later was incorporated into the state of Israel.
During the 1920s Egypt's religious Tribunal recognized the Baha'i Faith as a new, independent religion, totally separate from Islam, due to the nature of the 'laws, principles and beliefs' of the Baha'is. At the same time the Tribunal condemned "in most unequivocal and emphatic language the followers of Baha'u'llah as the believers in heresy, offensive and injurious to Islam, and wholly incompatible with the accepted doctrines and practice of its orthodox adherents."
Baháʼí institutions and community activities have been illegal under Egyptian law since 1960. All Baháʼí community properties, including Baháʼí centers, libraries, and cemeteries, have been confiscated by the government and fatwas have been issued charging Baháʼís with apostasy.
The Egyptian identification card controversy began in the 1990s when the government modernized the electronic processing of identity documents, which introduced a de facto requirement that documents must list the person's religion as Muslim, Christian, or Jewish (the only three religions officially recognized by the government). Consequently, Baháʼís were unable to obtain government identification documents (such as national identification cards, birth certificates, death certificates, marriage or divorce certificates, or passports) necessary to exercise their rights in their country unless they lied about their religion, which conflicts with Baháʼí religious principle. Without documents, they could not be employed, educated, treated in hospitals, travel outside of the country, or vote, among other hardships. Following a protracted legal process culminating in a court ruling favorable to the Baháʼís, the interior minister of Egypt released a decree on 14 April 2009, amending the law to allow Egyptians who are not Muslim, Christian, or Jewish to obtain identification documents that list a dash in place of one of the three recognized religions. The first identification cards were issued to two Baháʼís under the new decree on 8 August 2009. | https://en.wikipedia.org/wiki?curid=4251 |
Dots and Boxes
Dots and Boxes is a pencil-and-paper game for two players (sometimes more). It was first published in the 19th century by French mathematician Édouard Lucas, who called it . It has gone by many other names, including the game of dots, dot to dot grid, boxes, and pigs in a pen.
The game starts with an empty grid of dots. Usually two players take turns adding a single horizontal or vertical line between two "unjoined" adjacent dots. A player who completes the fourth side of a 1×1 box earns one point and takes another turn. (A point is typically recorded by placing a mark that identifies the player in the box, such as an initial.) The game ends when no more lines can be placed. The winner is the player with the most points. The board may be of any size grid. When short on time, or to learn the game, a 2×2 board (3×3 dots) is suitable. A 5×5 board, on the other hand, is good for experts.
The diagram on the right shows a game being played on a 2×2 board (3×3 dots). The second player ("B") plays a rotated mirror image of the first player's moves, hoping to divide the board into two pieces and tie the game. But the first player ("A") makes a "sacrifice" at move 7 and B accepts the sacrifice, getting one box. However, B "must" now add another line, and so B connects the center dot to the center-right dot, causing the remaining unscored boxes to be joined together in a "chain" (shown at the end of move 8). With A's next move, A gets all three of them and ends the game, winning 3–1.
For most novice players, the game begins with a phase of more-or-less randomly connecting dots, where the only strategy is to avoid adding the third side to any box. This continues until all the remaining (potential) boxes are joined together into "chains" – groups of one or more adjacent boxes in which any move gives all the boxes in the chain to the opponent. At this point, players typically take all available boxes, then "open" the smallest available chain to their opponent. For example, a novice player faced with a situation like position 1 in the diagram on the right, in which some boxes can be captured, may take all the boxes in the chain, resulting in position 2. But, with their last move, they have to open the next, larger chain, and the novice loses the game.
A more "experienced" player faced with position 1 will instead play the "double-cross strategy", taking all but 2 of the boxes in the chain and leaving position 3. The opponent will take these two boxes and then be forced to open the next chain. By achieving position 3, player A wins. The same double-cross strategy applies no matter how many long chains there are: a player using this strategy will take all but two boxes in each chain and take all the boxes in the last chain. If the chains are long enough, then this player will win.
The next level of strategic complexity, between "experts" who would both use the double-cross strategy (if they were allowed to), is a battle for "control": An expert player tries to force their opponent "to open the first long chain", because the player who first opens a long chain usually loses. Against a player who does not understand the concept of a sacrifice, the expert simply has to make the correct number of sacrifices to encourage the opponent to hand him the first chain long enough to ensure a win. "If" the other player also sacrifices, the expert has to additionally manipulate the number of available sacrifices through earlier play.
In combinatorial game theory, dots and boxes is an impartial game and many positions can be analyzed using Sprague–Grundy theory. However, Dots and Boxes lacks the normal play convention of most impartial games (where the last player to move wins), which complicates the analysis considerably.
Dots and Boxes need not be played on a rectangular grid – it can be played on a triangular grid or a hexagonal grid. There is also a variant in Bolivia where it is played in a chakana or Inca cross grid, which adds more complications to the game.
Dots and Boxes has a dual graph form called "Strings-and-Coins". This game is played on a network of coins (vertices) joined by strings (edges). Players take turns cutting a string. When a cut leaves a coin with no strings, the player "pockets" the coin and takes another turn. The winner is the player who pockets the most coins. Strings-and-Coins can be played on an arbitrary graph.
A variant Kropki played in Poland allows a player to claim a region of several squares as soon as its boundary is completed. In analyses of Dots and Boxes, starting with outer lines is called a "Swedish board" while the standard version is called an "American board". An intermediate version with the outer left and bottom sides starting with lines is called an "Icelandic board".
A game called Trxilt, combines some elements of Dots and Boxes with some elements of Chess. | https://en.wikipedia.org/wiki?curid=4260 |
Big Brother (Nineteen Eighty-Four)
Big Brother is a fictional character and symbol in George Orwell's dystopian 1949 novel "Nineteen Eighty-Four".
He is ostensibly the leader of Oceania, a totalitarian state wherein the ruling party Ingsoc wields total power "for its own sake" over the inhabitants. In the society that Orwell describes, every citizen is under constant surveillance by the authorities, mainly by telescreens (with the exception of the Proles). The people are constantly reminded of this by the slogan "Big Brother is watching you": a maxim that is ubiquitously on display.
In modern culture, the term "Big Brother" has entered the lexicon as a synonym for abuse of government power, particularly in respect to civil liberties, often specifically related to mass surveillance.
In the essay section of his novel "1985", Anthony Burgess states that Orwell got the idea for the name of Big Brother from advertising billboards for educational correspondence courses from a company called Bennett's during World War II. The original posters showed J. M. Bennett himself, a kindly-looking old man offering guidance and support to would-be students with the phrase "Let me be your father." According to Burgess, after Bennett's death, his son took over the company and the posters were replaced with pictures of the son (who looked imposing and stern in contrast to his father's kindly demeanor) with the text "Let me be your big brother".
Additional speculation from Douglas Kellner of the University of California, Los Angeles argued that Big Brother represents Joseph Stalin. Another theory is that the inspiration for Big Brother was Brendan Bracken, the Minister of Information until 1945. Orwell worked under Bracken on the BBC's Indian Service. Bracken was customarily referred to by his employees by his initials, B.B., the same initials as the character Big Brother. Orwell also resented the wartime censorship and need to manipulate information which he felt came from the highest levels of the Minister of Information and from Bracken's office in particular.
In the novel, it is never explicitly indicated if Big Brother is or had been a real person, or is a fictional personification of the Party, similar to Britannia and Uncle Sam. Big Brother is described as appearing on posters and telescreens as a man in his mid-forties. In Party propaganda, Big Brother is presented as one of the founders of the Party.
At one point, Winston Smith, the protagonist of Orwell's novel, tries "to remember in what year he had first heard mention of Big Brother. He thought it must have been at some time in the sixties, but it was impossible to be certain. In the Party histories, Big Brother figured as the leader and guardian of the Revolution since its very earliest days. His exploits had been gradually pushed backwards in time until already they extended into the fabulous world of the forties and the thirties, when the capitalists in their strange cylindrical hats still rode through the streets of London".
In the book "The Theory and Practice of Oligarchical Collectivism", read by Winston Smith and purportedly written by Goldstein, Big Brother is referred to as infallible and all-powerful. No one has ever seen him and there is a reasonable certainty that he will never die. He is simply "the guise in which the Party chooses to exhibit itself to the world" since the emotions of love, fear and reverence are more easily focused on an individual (if only a face on the hoardings and a voice on the telescreens) than an organisation. When Winston Smith is later arrested, O'Brien repeats that Big Brother will never die. When Smith asks if Big Brother exists, O'Brien describes him as "the embodiment of the Party" and says that he will exist as long as the Party exists. When Winston asks "Does Big Brother exist the same way I do?" (meaning is Big Brother an actual human being), O'Brien replies "You do not exist" (meaning that Smith is now an unperson; an example of doublethink).
Big Brother is the subject of a cult of personality. A spontaneous ritual of devotion to "BB" is illustrated at the end of the compulsory Two Minutes Hate:
Though Oceania's Ministry of Truth, Ministry of Plenty and Ministry of Peace each have names with meanings deliberately opposite to their real purpose, the Ministry of Love is perhaps the most straightforward as "rehabilitated thought criminals" leave the Ministry as loyal subjects who have been brainwashed into adoring (loving) Big Brother, hence its name.
The term "ministry" implies that each of these ministries is headed by a minister. If so, however, these ministers seem to be shadowy figures, whose names, words and acts are not publicised—public attention being focused solely on Big Brother.
The character, as represented solely by a single still photograph, was played in the 1954 BBC adaptation by production designer Roy Oxley. In the 1956 film adaptation, Big Brother was represented by an illustration of a stern-looking disembodied head.
In the film starring John Hurt released in 1984, the Big Brother photograph was of actor Bob Flag. Both Oxley and Flag sported small moustaches.
Since the publication of "Nineteen Eighty-Four", the phrase "Big Brother" has come into common use to describe any prying or overly-controlling authority figure and attempts by government to increase surveillance. Big Brother and other Orwellian imagery are often referenced in the joke known as the Russian reversal.
Iain Moncreiffe and Don Pottinger jokingly mentioned in their 1956 book "Blood Royal" the sentence: "Without Little Father need for Big Brother", referring to the Russian Revolution and the Soviet Union.
The worldwide reality television show "Big Brother" is based on the novel's concept of people being under constant surveillance. In 2000, after the United States version of the CBS program "Big Brother" premiered, the Estate of George Orwell sued CBS and its production company Orwell Productions, Inc. in federal court in Chicago for copyright and trademark infringement. The case was "Estate of Orwell v. CBS", 00-c-5034 (ND Ill). On the eve of trial, the case settled worldwide to the parties' "mutual satisfaction", but the amount that CBS paid to the Orwell Estate was not disclosed. CBS had not asked the Estate for permission. Under current laws, the novel will remain under copyright protection until 2020 in the European Union and until 2044 in the United States.
The December 2002 issue of "Gear" magazine featured a story about technologies and trends that could violate personal privacy moving society closer to a "Big Brother" state and utilised a recreation of the movie poster from the film version of "1984". The magazine "Book" ranked Big Brother no. 59 on its "100 best characters in fiction since 1900" list. "Wizard" magazine rated him the 75th-greatest villain of all time.
The iconic image of Big Brother (played by David Graham) played a key role in Apple's "1984" television commercial introducing the Macintosh. The Orwell Estate viewed the Apple commercial as a copyright infringement and sent a cease-and-desist letter to Apple and its advertising agency. The commercial was never televised again, though the date mentioned in the ad (24 January) was but two days later, making it unlikely that it would have been re-aired. Subsequent (now posthumous) ads featuring Steve Jobs (for a variety of products including audio books) have mimicked the format and appearance of that original ad campaign, with the appearance of Jobs nearly identical to that of Big Brother. In 2008, "The Simpsons" spoofed Apple's Big Brother commercial in an episode entitled "Mypods and Boomsticks".
Computer company Microsoft patented in 2011 a product distribution system with a camera or capture device that monitors the viewers that consume the product, allowing the provider to take "remedial action" if the actual viewers do not match the distribution license. The system has been compared with "1984" telescreen surveillance system.
A series of laws intended to implement the European Union Data Retention Directive in Romania were nicknamed "the Big Brother laws" by Romanian media and critics as they would have led to blanket storage of citizens' telecommunications data for six months. All of these laws were struck down as unconstitutional by the Constitutional Court of Romania and the Directive itself was ultimately invalidated by the Court of Justice of the European Union.
In the 2010 video game "BioShock 2", there is an enemy named the Big Sister. The phrase "Big Sister is watching" is a reference to the 1984 phrase "Big Brother is watching".
China's Social Credit System has been described as akin to "Big Brother" by detractors, where citizens and businesses are given or deducted good behavior points depending on their choices. | https://en.wikipedia.org/wiki?curid=4261 |
Binary search algorithm
In computer science, binary search, also known as half-interval search, logarithmic search, or binary chop, is a search algorithm that finds the position of a target value within a sorted array. Binary search compares the target value to the middle element of the array. If they are not equal, the half in which the target cannot lie is eliminated and the search continues on the remaining half, again taking the middle element to compare to the target value, and repeating this until the target value is found. If the search ends with the remaining half being empty, the target is not in the array.
Binary search runs in logarithmic time in the worst case, making formula_1 comparisons, where formula_2 is the number of elements in the array, the formula_3 is Big O notation, and formula_4 is the logarithm. Binary search is faster than linear search except for small arrays. However, the array must be sorted first to be able to apply binary search. There are specialized data structures designed for fast searching, such as hash tables, that can be searched more efficiently than binary search. However, binary search can be used to solve a wider range of problems, such as finding the next-smallest or next-largest element in the array relative to the target even if it is absent from the array.
There are numerous variations of binary search. In particular, fractional cascading speeds up binary searches for the same value in multiple arrays. Fractional cascading efficiently solves a number of search problems in computational geometry and in numerous other fields. Exponential search extends binary search to unbounded lists. The binary search tree and B-tree data structures are based on binary search.
Binary search works on sorted arrays. Binary search begins by comparing an element in the middle of the array with the target value. If the target value matches the element, its position in the array is returned. If the target value is less than the element, the search continues in the lower half of the array. If the target value is greater than the element, the search continues in the upper half of the array. By doing this, the algorithm eliminates the half in which the target value cannot lie in each iteration.
Given an array formula_5 of formula_2 elements with values or records formula_7sorted such that formula_8, and target value formula_9, the following subroutine uses binary search to find the index of formula_9 in formula_5.
This iterative procedure keeps track of the search boundaries with the two variables formula_12 and formula_14. The procedure may be expressed in pseudocode as follows, where the variable names and types remain the same as above, codice_1 is the floor function, and codice_2 refers to a specific value that conveys the failure of the search.
Alternatively, the algorithm may take the ceiling of formula_18. This may change the result if the target value appears more than once in the array.
In the above procedure, the algorithm checks whether the middle element (formula_17) is equal to the target (formula_9) in every iteration. Some implementations leave out this check during each iteration. The algorithm would perform this check only when one element is left (when formula_33). This results in a faster comparison loop, as one comparison is eliminated per iteration. However, it requires one more iteration on average.
Hermann Bottenbruch published the first implementation to leave out this check in 1962.
Where codice_3 is the ceiling function, the pseudocode for this version is:
The procedure may return any index whose element is equal to the target value, even if there are duplicate elements in the array. For example, if the array to be searched was formula_51 and the target was formula_52, then it would be correct for the algorithm to either return the 4th (index 3) or 5th (index 4) element. The regular procedure would return the 4th element (index 3) in this case. It does not always return the first duplicate (consider formula_53 which still returns the 4th element). However, it is sometimes necessary to find the leftmost element or the rightmost element for a target value that is duplicated in the array. In the above example, the 4th element is the leftmost element of the value 4, while the 5th element is the rightmost element of the value 4. The alternative procedure above will always return the index of the rightmost element if such an element exists.
To find the leftmost element, the following procedure can be used:
If formula_69 and formula_70, then formula_71 is the leftmost element that equals formula_9. Even if formula_9 is not in the array, formula_12 is the rank of formula_9 in the array, or the number of elements in the array that are less than formula_9.
Where codice_1 is the floor function, the pseudocode for this version is:
To find the rightmost element, the following procedure can be used:
If formula_92 and formula_93, then formula_94 is the rightmost element that equals formula_9. Even if "formula_9" is not in the array, formula_97 is the number of elements in the array that are greater than "formula_9".
Where codice_1 is the floor function, the pseudocode for this version is:
The above procedure only performs "exact" matches, finding the position of a target value. However, it is trivial to extend binary search to perform approximate matches because binary search operates on sorted arrays. For example, binary search can be used to compute, for a given value, its rank (the number of smaller elements), predecessor (next-smallest element), successor (next-largest element), and nearest neighbor. Range queries seeking the number of elements between two values can be performed with two rank queries.
In terms of the number of comparisons, the performance of binary search can be analyzed by viewing the run of the procedure on a binary tree. The root node of the tree is the middle element of the array. The middle element of the lower half is the left child node of the root, and the middle element of the upper half is the right child node of the root. The rest of the tree is built in a similar fashion. Starting from the root node, the left or right subtrees are traversed depending on whether the target value is less or more than the node under consideration.
In the worst case, binary search makes formula_103 iterations of the comparison loop, where the formula_104 notation denotes the floor function that yields the greatest integer less than or equal to the argument, and formula_105 is the binary logarithm. This is because the worst case is reached when the search reaches the deepest level of the tree, and there are always formula_103 levels in the tree for any binary search.
The worst case may also be reached when the target element is not in the array. If formula_107 is one less than a power of two, then this is always the case. Otherwise, the search may perform formula_103iterations if the search reaches the deepest level of the tree. However, it may make formula_109 iterations, which is one less than the worst case, if the search ends at the second-deepest level of the tree.
On average, assuming that each element is equally likely to be searched, binary search makes formula_110 iterations when the target element is in the array. This is approximately equal to formula_111 iterations. When the target element is not in the array, binary search makes formula_112 iterations on average, assuming that the range between and outside elements is equally likely to be searched.
In the best case, where the target value is the middle element of the array, its position is returned after one iteration.
In terms of iterations, no search algorithm that works only by comparing elements can exhibit better average and worst-case performance than binary search. The comparison tree representing binary search has the fewest levels possible as every level above the lowest level of the tree is filled completely. Otherwise, the search algorithm can eliminate few elements in an iteration, increasing the number of iterations required in the average and worst case. This is the case for other search algorithms based on comparisons, as while they may work faster on some target values, the average performance over "all" elements is worse than binary search. By dividing the array in half, binary search ensures that the size of both subarrays are as similar as possible.
Binary search requires three pointers to elements, which may be array indices or pointers to memory locations, regardless of the size of the array. However, it requires at least formula_113 bits to encode a pointer to an element of an array with formula_2 elements. Therefore, the space complexity of binary search is formula_1. In addition, it takes formula_116 space to store the array.
The average number of iterations performed by binary search depends on the probability of each element being searched. The average case is different for successful searches and unsuccessful searches. It will be assumed that each element is equally likely to be searched for successful searches. For unsuccessful searches, it will be assumed that the intervals between and outside elements are equally likely to be searched. The average case for successful searches is the number of iterations required to search every element exactly once, divided by formula_2, the number of elements. The average case for unsuccessful searches is the number of iterations required to search an element within every interval exactly once, divided by the formula_118 intervals.
In the binary tree representation, a successful search can be represented by a path from the root to the target node, called an "internal path". The length of a path is the number of edges (connections between nodes) that the path passes through. The number of iterations performed by a search, given that the corresponding path has length formula_119, is formula_120 counting the initial iteration. The "internal path length" is the sum of the lengths of all unique internal paths. Since there is only one path from the root to any single node, each internal path represents a search for a specific element. If there are formula_2 elements, which is a positive integer, and the internal path length is formula_122, then the average number of iterations for a successful search formula_123, with the one iteration added to count the initial iteration.
Since binary search is the optimal algorithm for searching with comparisons, this problem is reduced to calculating the minimum internal path length of all binary trees with formula_2 nodes, which is equal to:
formula_125
For example, in a 7-element array, the root requires one iteration, the two elements below the root require two iterations, and the four elements below require three iterations. In this case, the internal path length is:
formula_126
The average number of iterations would be formula_127 based on the equation for the average case. The sum for formula_122 can be simplified to:
formula_129
Substituting the equation for formula_122 into the equation for formula_131:
formula_132
For integer formula_2, this is equivalent to the equation for the average case on a successful search specified above.
Unsuccessful searches can be represented by augmenting the tree with "external nodes", which forms an "extended binary tree". If an internal node, or a node present in the tree, has fewer than two child nodes, then additional child nodes, called external nodes, are added so that each internal node has two children. By doing so, an unsuccessful search can be represented as a path to an external node, whose parent is the single element that remains during the last iteration. An "external path" is a path from the root to an external node. The "external path length" is the sum of the lengths of all unique external paths. If there are formula_2 elements, which is a positive integer, and the external path length is formula_135, then the average number of iterations for an unsuccessful search formula_136, with the one iteration added to count the initial iteration. The external path length is divided by formula_137 instead of formula_2 because there are formula_137 external paths, representing the intervals between and outside the elements of the array.
This problem can similarly be reduced to determining the minimum external path length of all binary trees with formula_2 nodes. For all binary trees, the external path length is equal to the internal path length plus formula_141. Substituting the equation for formula_122:
formula_143
Substituting the equation for formula_135 into the equation for formula_145, the average case for unsuccessful searches can be determined:
formula_146
Each iteration of the binary search procedure defined above makes one or two comparisons, checking if the middle element is equal to the target in each iteration. Assuming that each element is equally likely to be searched, each iteration makes 1.5 comparisons on average. A variation of the algorithm checks whether the middle element is equal to the target at the end of the search. On average, this eliminates half a comparison from each iteration. This slightly cuts the time taken per iteration on most computers. However, it guarantees that the search takes the maximum number of iterations, on average adding one iteration to the search. Because the comparison loop is performed only formula_103 times in the worst case, the slight increase in efficiency per iteration does not compensate for the extra iteration for all but very large formula_107.
In analyzing the performance of binary search, another consideration is the time required to compare two elements. For integers and strings, the time required increases linearly as the encoding length (usually the number of bits) of the elements increase. For example, comparing a pair of 64-bit unsigned integers would require comparing up to double the bits as comparing a pair of 32-bit unsigned integers. The worst case is achieved when the integers are equal. This can be significant when the encoding lengths of the elements are large, such as with large integer types or long strings, which makes comparing elements expensive. Furthermore, comparing floating-point values (the most common digital representation of real numbers) is often more expensive than comparing integers or short strings.
On most computer architectures, the processor has a hardware cache separate from RAM. Since they are located within the processor itself, caches are much faster to access but usually store much less data than RAM. Therefore, most processors store memory locations that have been accessed recently, along with memory locations close to it. For example, when an array element is accessed, the element itself may be stored along with the elements that are stored close to it in RAM, making it faster to sequentially access array elements that are close in index to each other (locality of reference). On a sorted array, binary search can jump to distant memory locations if the array is large, unlike algorithms (such as linear search and linear probing in hash tables) which access elements in sequence. This adds slightly to the running time of binary search for large arrays on most systems.
Sorted arrays with binary search are a very inefficient solution when insertion and deletion operations are interleaved with retrieval, taking formula_149 time for each such operation. In addition, sorted arrays can complicate memory use especially when elements are often inserted into the array. There are other data structures that support much more efficient insertion and deletion. Binary search can be used to perform exact matching and set membership (determining whether a target value is in a collection of values). There are data structures that support faster exact matching and set membership. However, unlike many other searching schemes, binary search can be used for efficient approximate matching, usually performing such matches in formula_150 time regardless of the type or structure of the values themselves. In addition, there are some operations, like finding the smallest and largest element, that can be performed efficiently on a sorted array.
Linear search is a simple search algorithm that checks every record until it finds the target value. Linear search can be done on a linked list, which allows for faster insertion and deletion than an array. Binary search is faster than linear search for sorted arrays except if the array is short, although the array needs to be sorted beforehand. All sorting algorithms based on comparing elements, such as quicksort and merge sort, require at least formula_151 comparisons in the worst case. Unlike linear search, binary search can be used for efficient approximate matching. There are operations such as finding the smallest and largest element that can be done efficiently on a sorted array but not on an unsorted array.
A binary search tree is a binary tree data structure that works based on the principle of binary search. The records of the tree are arranged in sorted order, and each record in the tree can be searched using an algorithm similar to binary search, taking on average logarithmic time. Insertion and deletion also require on average logarithmic time in binary search trees. This can be faster than the linear time insertion and deletion of sorted arrays, and binary trees retain the ability to perform all the operations possible on a sorted array, including range and approximate queries.
However, binary search is usually more efficient for searching as binary search trees will most likely be imperfectly balanced, resulting in slightly worse performance than binary search. This even applies to balanced binary search trees, binary search trees that balance their own nodes, because they rarely produce the tree with the fewest possible levels. Except for balanced binary search trees, the tree may be severely imbalanced with few internal nodes with two children, resulting in the average and worst-case search time approaching formula_107 comparisons. Binary search trees take more space than sorted arrays.
Binary search trees lend themselves to fast searching in external memory stored in hard disks, as binary search trees can be efficiently structured in filesystems. The B-tree generalizes this method of tree organization. B-trees are frequently used to organize long-term storage such as databases and filesystems.
For implementing associative arrays, hash tables, a data structure that maps keys to records using a hash function, are generally faster than binary search on a sorted array of records. Most hash table implementations require only amortized constant time on average. However, hashing is not useful for approximate matches, such as computing the next-smallest, next-largest, and nearest key, as the only information given on a failed search is that the target is not present in any record. Binary search is ideal for such matches, performing them in logarithmic time. Binary search also supports approximate matches. Some operations, like finding the smallest and largest element, can be done efficiently on sorted arrays but not on hash tables.
A related problem to search is set membership. Any algorithm that does lookup, like binary search, can also be used for set membership. There are other algorithms that are more specifically suited for set membership. A bit array is the simplest, useful when the range of keys is limited. It compactly stores a collection of bits, with each bit representing a single key within the range of keys. Bit arrays are very fast, requiring only formula_153 time. The Judy1 type of Judy array handles 64-bit keys efficiently.
For approximate results, Bloom filters, another probabilistic data structure based on hashing, store a set of keys by encoding the keys using a bit array and multiple hash functions. Bloom filters are much more space-efficient than bit arrays in most cases and not much slower: with formula_154 hash functions, membership queries require only formula_155 time. However, Bloom filters suffer from false positives.
There exist data structures that may improve on binary search in some cases for both searching and other operations available for sorted arrays. For example, searches, approximate matches, and the operations available to sorted arrays can be performed more efficiently than binary search on specialized data structures such as van Emde Boas trees, fusion trees, tries, and bit arrays. These specialized data structures are usually only faster because they take advantage of the properties of keys with a certain attribute (usually keys that are small integers), and thus will be time or space consuming for keys that lack that attribute. As long as the keys can be ordered, these operations can always be done at least efficiently on a sorted array regardless of the keys. Some structures, such as Judy arrays, use a combination of approaches to mitigate this while retaining efficiency and the ability to perform approximate matching.
Uniform binary search stores, instead of the lower and upper bounds, the difference in the index of the middle element from the current iteration to the next iteration. A lookup table containing the differences is computed beforehand. For example, if the array to be searched is , the middle element (formula_17) would be . In this case, the middle element of the left subarray () is and the middle element of the right subarray () is . Uniform binary search would store the value of as both indices differ from by this same amount. To reduce the search space, the algorithm either adds or subtracts this change from the index of the middle element. Uniform binary search may be faster on systems where it is inefficient to calculate the midpoint, such as on decimal computers.
Exponential search extends binary search to unbounded lists. It starts by finding the first element with an index that is both a power of two and greater than the target value. Afterwards, it sets that index as the upper bound, and switches to binary search. A search takes formula_157 iterations before binary search is started and at most formula_158 iterations of the binary search, where formula_159 is the position of the target value. Exponential search works on bounded lists, but becomes an improvement over binary search only if the target value lies near the beginning of the array.
Instead of calculating the midpoint, interpolation search estimates the position of the target value, taking into account the lowest and highest elements in the array as well as length of the array. It works on the basis that the midpoint is not the best guess in many cases. For example, if the target value is close to the highest element in the array, it is likely to be located near the end of the array.
A common interpolation function is linear interpolation. If formula_5 is the array, formula_161 are the lower and upper bounds respectively, and formula_9 is the target, then the target is estimated to be about formula_163 of the way between formula_12 and formula_14. When linear interpolation is used, and the distribution of the array elements is uniform or near uniform, interpolation search makes formula_166 comparisons.
In practice, interpolation search is slower than binary search for small arrays, as interpolation search requires extra computation. Its time complexity grows more slowly than binary search, but this only compensates for the extra computation for large arrays.
Fractional cascading is a technique that speeds up binary searches for the same element in multiple sorted arrays. Searching each array separately requires formula_167 time, where formula_154 is the number of arrays. Fractional cascading reduces this to formula_169 by storing specific information in each array about each element and its position in the other arrays.
Fractional cascading was originally developed to efficiently solve various computational geometry problems. Fractional cascading has been applied elsewhere, such as in data mining and Internet Protocol routing.
Binary search has been generalized to work on certain types of graphs, where the target value is stored in a vertex instead of an array element. Binary search trees are one such generalization—when a vertex (node) in the tree is queried, the algorithm either learns that the vertex is the target, or otherwise which subtree the target would be located in. However, this can be further generalized as follows: given an undirected, positively weighted graph and a target vertex, the algorithm learns upon querying a vertex that it is equal to the target, or it is given an incident edge that is on the shortest path from the queried vertex to the target. The standard binary search algorithm is simply the case where the graph is a path. Similarly, binary search trees are the case where the edges to the left or right subtrees are given when the queried vertex is unequal to the target. For all undirected, positively weighted graphs, there is an algorithm that finds the target vertex in formula_1 queries in the worst case.
Noisy binary search algorithms solve the case where the algorithm cannot reliably compare elements of the array. For each pair of elements, there is a certain probability that the algorithm makes the wrong comparison. Noisy binary search can find the correct position of the target with a given probability that controls the reliability of the yielded position. Every noisy binary search procedure must make at least formula_171 comparisons on average, where formula_172 is the binary entropy function and formula_173 is the probability that the procedure yields the wrong position. The noisy binary search problem can be considered as a case of the Rényi-Ulam game, a variant of Twenty Questions where the answers may be wrong.
Classical computers are bounded to the worst case of exactly formula_174 iterations when performing binary search. Quantum algorithms for binary search are still bounded to a proportion of formula_175 queries (representing iterations of the classical procedure), but the constant factor is less than one, providing for a lower time complexity on quantum computers. Any "exact" quantum binary search procedure—that is, a procedure that always yields the correct result—requires at least formula_176 queries in the worst case, where formula_177 is the natural logarithm. There is an exact quantum binary search procedure that runs in formula_178 queries in the worst case. In comparison, Grover's algorithm is the optimal quantum algorithm for searching an unordered list of elements, and it requires formula_179 queries.
The idea of sorting a list of items to allow for faster searching dates back to antiquity. The earliest known example was the Inakibit-Anu tablet from Babylon dating back to . The tablet contained about 500 Sexagesimal numbers and their reciprocals sorted in Lexicographical order, which made searching for a specific entry easier. In addition, several lists of names that were sorted by their first letter were discovered on the Aegean Islands. "Catholicon", a Latin dictionary finished in 1286 CE, was the first work to describe rules for sorting words into alphabetical order, as opposed to just the first few letters.
In 1946, John Mauchly made the first mention of binary search as part of the Moore School Lectures, a seminal and foundational college course in computing. In 1957, William Wesley Peterson published the first method for interpolation search. Every published binary search algorithm worked only for arrays whose length is one less than a power of two until 1960, when Derrick Henry Lehmer published a binary search algorithm that worked on all arrays. In 1962, Hermann Bottenbruch presented an ALGOL 60 implementation of binary search that placed the comparison for equality at the end, increasing the average number of iterations by one, but reducing to one the number of comparisons per iteration. The uniform binary search was developed by A. K. Chandra of Stanford University in 1971. In 1986, Bernard Chazelle and Leonidas J. Guibas introduced fractional cascading as a method to solve numerous search problems in computational geometry.
When Jon Bentley assigned binary search as a problem in a course for professional programmers, he found that ninety percent failed to provide a correct solution after several hours of working on it, mainly because the incorrect implementations failed to run or returned a wrong answer in rare edge cases. A study published in 1988 shows that accurate code for it is only found in five out of twenty textbooks. Furthermore, Bentley's own implementation of binary search, published in his 1986 book "Programming Pearls", contained an overflow error that remained undetected for over twenty years. The Java programming language library implementation of binary search had the same overflow bug for more than nine years.
In a practical implementation, the variables used to represent the indices will often be of fixed size, and this can result in an arithmetic overflow for very large arrays. If the midpoint of the span is calculated as formula_18, then the value of formula_181 may exceed the range of integers of the data type used to store the midpoint, even if formula_12 and formula_14 are within the range. If "formula_12" and "formula_14" are nonnegative, this can be avoided by calculating the midpoint as formula_186.
An infinite loop may occur if the exit conditions for the loop are not defined correctly. Once "formula_12" exceeds "formula_14", the search has failed and must convey the failure of the search. In addition, the loop must be exited when the target element is found, or in the case of an implementation where this check is moved to the end, checks for whether the search was successful or failed at the end must be in place. Bentley found that most of the programmers who incorrectly implemented binary search made an error in defining the exit conditions.
Many languages' standard libraries include binary search routines: | https://en.wikipedia.org/wiki?curid=4266 |
Belle and Sebastian
Belle and Sebastian are a Scottish rock band formed in Glasgow in 1994. Led by Stuart Murdoch, the band has released nine albums. Much of their work had been released on Jeepster Records, but they are now signed to Matador Records worldwide.
Belle and Sebastian were formed in Glasgow, Scotland in 1994 by Stuart Murdoch and Stuart David, both of whom had enrolled at Stow College's Beatbox programme for unemployed musicians. Together, with music professor Alan Rankine (formerly of The Associates), they recorded some demos, which were picked up by the college's Music Business course that produces and releases one single each year on the college's label, Electric Honey. As the band had a number of songs already and the label was extremely impressed with the demos, Belle and Sebastian were allowed to record a full-length album, which was recorded mostly live over three days, entitled "Tigermilk". Murdoch once described the band as a "product of botched capitalism". The band took their name from a short story Murdoch had written inspired by the television adaptation of the French novel "Belle et Sébastien" about a six-year-old boy and his dog.
"Tigermilk" was recorded in three days and originally only one thousand copies were pressed in vinyl. As of 2007, these original copies were valued up to £400. The warm reception the album received inspired Murdoch and David to turn the band into a full-time project, recruiting Stevie Jackson (guitar and vocals), Isobel Campbell (cello/vocals), Chris Geddes (keys) and Richard Colburn (drums) to fill out the group.
After the success of the debut album, Belle and Sebastian were signed to Jeepster Records in August 1996 and "If You're Feeling Sinister", their second album, was released on 18 November. The album was named by "Spin" as one of the 100 greatest albums between 1985 and 2005, and is widely considered the band's masterpiece. Just before the recording of "Sinister", Sarah Martin (violin/vocals) joined the band. Following this a series of EPs were released in 1997. The first of these was "Dog on Wheels", which contained four demo tracks recorded before the real formation of the band. In fact, the only long-term band members to play on the songs were Murdoch, David, and Mick Cooke, who played trumpet on the EP but would not officially join the band until a few years later. It charted at No. 59 in the UK singles chart.
The "Lazy Line Painter Jane" EP followed in July. The track was recorded in the church where Murdoch lived and features vocals from Monica Queen. The EP narrowly missed out on the UK top 40, peaking at No. 41. The last of the 1997 EPs was October's "3.. 6.. 9 Seconds of Light". The EP was made Single of the Week in both the "NME" and "Melody Maker" and reached No. 32 in the charts, thus becoming the band's first top 40 single.
The band released their third LP, "The Boy with the Arab Strap" in 1998, and it reached No. 12 in the UK charts. "Arab Strap" garnered an NPR interview and positive reviews from "Rolling Stone" and the "Village Voice," among others; however, the album has its detractors, including "Pitchfork", who gave the album a particularly poor review, calling it a "parody" of their earlier work (Pitchfork has since removed the review from their website, but re-reviewed the album positively in 2018). During the recording of the album, long-time studio trumpet-player Mick Cooke was asked to join the band as a full member. The "This Is Just a Modern Rock Song" EP followed later that year.
In 1999, the band was awarded with Best Newcomer (for their third album) at the BRIT Awards, upsetting better-known acts such as Steps and 5ive. That same year, the band hosted their own festival, the Bowlie Weekender. "Tigermilk" was also given a full release by Jeepster before the band started work on their next LP. The result was "Fold Your Hands Child, You Walk Like a Peasant", which became the band's first top 10 album in the UK. A stand-alone single, "Legal Man", reached No. 15 and gave them their first appearance on Top of the Pops.
As the band's popularity and recognition was growing worldwide, their music began appearing in films and on television. The 2000 film "High Fidelity" mentions the band and features a clip from the song "Seymour Stein" from "The Boy with the Arab Strap". Also, the title track from "Arab Strap" was played over the end credits of the UK television series "Teachers," and the lyric "Colour my life with the chaos of trouble" from the song was quoted by one of the characters in the 2009 film "(500) Days of Summer".
Stuart David soon left the band to concentrate on his side project, Looper, and his book writing, which included his "The Idle Thoughts of a Daydreamer". He was replaced by Bobby Kildea of V-Twin. The "Jonathan David" single, sung by Stevie Jackson, was released in June 2001 and was followed by "I'm Waking Up to Us" in November. "I'm Waking Up to Us" saw the band use an outside producer (Mike Hurst) for the first time. Most of 2002 was spent touring and recording a soundtrack album, "Storytelling" (for "Storytelling" by Todd Solondz). Campbell left the band in the spring of 2002, in the middle of the band's North American tour to pursue a solo career, first as The Gentle Waves, and later under her own name. She later collaborated with singer Mark Lanegan on three albums.
The band left Jeepster in 2002, signing a four-album deal with Rough Trade Records. Their first album for Rough Trade, "Dear Catastrophe Waitress", was released in 2003, and was produced by Trevor Horn. The album showed a markedly more "produced" sound compared to their first four LPs, as the band was making a concerted effort to produce more "radio-friendly" music. The album was warmly received and is credited with restoring the band's "indie cred". The album also marked the return of Murdoch as the group's primary songwriter, following the poorly received "Fold Your Hands Child, You Walk Like a Peasant" and "Storytelling", both of which were more collaborative than the band's early work. A documentary DVD, "Fans Only", was released by Jeepster in October 2003, featuring promotional videos, live clips and unreleased footage. A single from the album, "Step into My Office, Baby" followed in November 2003; it would be their first single to be taken from an album, and included a track recorded with Divine Comedy producer Darren Allison entitled "Love on the March".
The Thin Lizzy-inspired "I'm a Cuckoo" was the second single from the album. It achieved their highest chart position yet, reaching No. 14 in the UK. The "Books" EP followed, a double A-side single led by "Wrapped Up in Books" from "Dear Catastrophe Waitress" and the new "Your Cover's Blown". This EP became the band's third top 20 UK release, and the band was nominated for both the Mercury Music Prize and an Ivor Novello Award. In January 2005, B&S was voted Scotland's greatest band in a poll by The List, beating Simple Minds, Idlewild, Travis, Franz Ferdinand, and The Proclaimers, among others.
In April 2005, members of the band visited Israel and the Palestinian territories with the UK charity War on Want; the group subsequently recorded a song inspired by the trip titled "The Eighth Station of the Cross Kebab House", which would first appear on the digital-download version of the charity album and would later have a physical release as a B-side on 2006's "Funny Little Frog" single. "Push Barman to Open Old Wounds", a compilation of the Jeepster singles and EPs, was released in May 2005 while the band were recording their seventh album in California. The result of the sessions was "The Life Pursuit", produced by Tony Hoffer. The album, originally intended to be a double album, became the band's highest-charting album upon its release in February 2006, peaking at No. 8 in the UK and No. 65 on the US "Billboard" 200. "Funny Little Frog", which preceded it, also proved to be their highest-charting single, debuting at No. 13.
On 6 July 2006, the band played a historic show with the Los Angeles Philharmonic at the Hollywood Bowl. The opening act at the 18,000 seat sell-out concert was The Shins. The members of the band see this as a landmark event, with Stevie Jackson saying, "This is the biggest thrill of my entire life". In October 2006, members of the band helped put together a CD collection of new songs for children titled "Colours Are Brighter", with the involvement of major bands such as Franz Ferdinand and The Flaming Lips.
On 18 November 2008 the band released "The BBC Sessions", which features songs from the period of 1996–2001 (including the last recordings featuring Isobel Campbell before she left the band), along with a second disc featuring a recording of a live performance in Belfast from Christmas 2001.
On 17 July 2010, the band performed their first UK gig in almost four years to a crowd of around 30,000 at Latitude Festival in Henham Park, Southwold. They performed two new songs, "I Didn't See It Coming" and "I'm Not Living in the Real World".
Their eighth studio album, released in the UK and internationally on 25 September 2010, was titled "Belle and Sebastian Write About Love". The first single from the album, as well as the record's title track "Write About Love", was released in the US on 7 September 2010. "Write About Love" entered the UK albums chart in its first week of release, peaking at No. 8 as of 19 October 2010. Norah Jones is featured on the track "Little Lou, Ugly Jack, Prophet John".
In December 2010 Belle and Sebastian curated the sequel to the "Bowlie Weekender" in the form of "Bowlie 2" presented by All Tomorrow's Parties.
In 2013, Pitchfork TV released an hour-long documentary in February, directed by RJ Bentler which focused on the band's 1996 album If You're Feeling Sinister, as well as the formation and early releases of the band. The documentary featured interviews with every member that was present on the album, as well as several archival photos and videos from the band's early days. The band compiled a second compilation album "The Third Eye Centre" which included the b-sides and rarities released after "Push Barman to Open Old Wounds", from the albums "Dear Catastrophe Waitress", "The Life Pursuit", and "Write About Love". In an interview at the end of 2013, Mick Cooke confirmed he had left the band on good terms.
The band received an 'Outstanding Contribution To Music Award' at the NME Awards 2014.
In 2014, the band returned to the studio, recording in Atlanta, Georgia for their ninth studio album, along with announcing tour dates for various festivals and concerts across the world during 2014. Their ninth album "Girls in Peacetime Want to Dance" was released on 19 January 2015. It was their first album with Dave McGowan, who had been their touring bassist since 2011.
The Belle and Sebastian song "There's Too Much Love" forms much of the soundtrack for the Brazilian film "The Way He Looks", about a blind, gay teenage boy and his friends, released in 2014.
Belle and Sebastian performed at the world-famous Glastonbury Festival on Sunday 28 June 2015, on 'The Other Stage' and at O2 Academy, Glasgow in March 2017 which was televised in the U.K. as part of the 'BBC 6 MUSIC Presents Festival'.
In mid-2017, the band put out a new single, "We Were Beautiful". During the same year, the band appeared in the news for a comical story that occurred during their US tour, in which they accidentally forgot Colburn in a North Dakota Walmart. In December 2017 and January and February 2018, the band released a trio of EPs under the name "How to Solve Our Human Problems".
On 3 November 2018, the band announced that Dave McGowan had officially become a member.
In August 2019, to celebrate the 20th anniversary of the first Bowlie Weekender festival, Belle & Sebastian held a third festival, dubbed the "Boaty Weekender". Unlike the previous two festivals, the Boaty Weekender was held on a cruise ship in the Mediterranean Sea instead of U.K. holiday parks.
Current members
Former members
Timeline | https://en.wikipedia.org/wiki?curid=4267 |
Broadcast domain
A broadcast domain is a logical division of a computer network, in which all nodes can reach each other by broadcast at the data link layer. A broadcast domain can be within the same LAN segment or it can be bridged to other LAN segments.
In terms of current popular technologies, any computer connected to the same Ethernet repeater or switch is a member of the same broadcast domain. Further, any computer connected to the same set of inter-connected switches/repeaters is a member of the same broadcast domain. Routers and other higher-layer devices form boundaries between broadcast domains.
The notion of broadcast domain should be contrasted with that of collision domain, which would be all nodes on the same set of inter-connected repeaters, divided by switches and learning bridges. Collision domains are generally smaller than, and contained within, broadcast domains.
While some layer two network devices are able to divide the collision domains, broadcast domains are only divided by layer 3 network devices such as routers or layer 3 switches. Separating VLANs divides broadcast domains as well.
The distinction between broadcast and collision domains comes about because simple Ethernet and similar systems use a shared transmission system. In simple Ethernet (without switches or bridges), data frames are transmitted to all other nodes on a network. Each receiving node checks the destination address of each frame, and simply ignores any frame not addressed to its own MAC address or the broadcast address.
Switches act as buffers, receiving and analyzing the frames from each connected network segment. Frames destined for nodes connected to the originating segment are not forwarded by the switch. Frames destined for a specific node on a different segment are sent only to that segment. Only broadcast frames are forwarded to all other segments. This reduces unnecessary traffic and collisions.
In such a switched network, transmitted frames may not be received by all other reachable nodes. Nominally, only broadcast frames will be received by all other nodes. Collisions are localized to the network segment they occur on. Thus, the broadcast domain is the entire inter-connected layer two network, and the segments connected to each switch/bridge port are each a collision domain.
Not all network systems or media feature broadcast/collision domains. For example, PPP links.
With a sufficiently sophisticated switch, it is possible to create a network in which the normal notion of a broadcast domain is strictly controlled. One implementation of this concept is termed a "private VLAN". Another implementation is possible with Linux and iptables. One helpful analogy is that by creating multiple VLANs, the number of broadcast domains increases, but the size of each broadcast domain decreases. This is because a virtual LAN (or VLAN) is technically a broadcast domain.
This is achieved by designating one or more "server" or "provider" nodes, either by MAC address or switch port. Broadcast frames are allowed to originate from these sources, and are sent to all other nodes. Broadcast frames from all other sources are directed only to the server/provider nodes. Traffic from other sources not destined to the server/provider nodes ("peer-to-peer" traffic) is blocked.
The result is a network based on a nominally shared transmission system; like Ethernet, but in which "client" nodes cannot communicate with each other, only with the server/provider. A common application is Internet providers. Allowing direct data link layer communication between customer nodes exposes the network to various security attacks, such as ARP spoofing. Controlling the broadcast domain in this fashion provides many of the advantages of a point-to-point network, using commodity broadcast-based hardware. | https://en.wikipedia.org/wiki?curid=4279 |
Beechcraft
Beechcraft is a brand of Textron Aviation since 2014. Originally, it was a brand of Beech Aircraft Corporation, an American manufacturer of general aviation, commercial, and military aircraft, ranging from light single-engined aircraft to twin-engined turboprop transports, business jets, and military trainers. Beech later became a division of Raytheon and then Hawker Beechcraft before a bankruptcy sale turned its assets over to Textron (parent company of Beech's historical cross-town Wichita rival, Cessna Aircraft Company).
Beech Aircraft Company was founded in Wichita, Kansas, in 1932 by Walter Beech as president, his wife Olive Ann Beech as secretary, Ted A. Wells as vice president of engineering, K. K. Shaul as treasurer, and investor C. G. Yankey as vice president. The company began operations in an idle Cessna factory. With designer Ted Wells, they developed the first aircraft under the Beechcraft name, the classic Beechcraft Model 17 Staggerwing, which first flew in November 1932. Over 750 Staggerwings were built, with 270 manufactured for the United States Army Air Forces during World War II.
Beechcraft was not Beech's first company, as he had previously formed Travel Air in 1924 and the design numbers used at Beechcraft followed the sequence started at Travel Air, and were then continued at Curtiss-Wright, after Travel Air had been absorbed into the much larger company in 1929. Beech became president of the Curtiss-Wright's airplane division and VP of sales, but became dissatisfied with being so far removed from aircraft production and quit to form Beechcraft, using the original Travel Air facilities and employing many of the same people. Model numbers prior to 11/11000 were built under the "Travel Air" name, while Curtiss-Wright built the CW-12, 14, 15, and 16 as well as previous successful Travel Air models (mostly the model 4).
In 1942 Beech won its first Army-Navy "E" Award production award and became one of the elite five percent of war contracting firms in the country to win five straight awards for production efficiency, mostly for the production of the Beechcraft Model 18 which remains in widespread use worldwide. Beechcraft ranked 69th among United States corporations in the value of World War II military production contracts.
After the war, the Staggerwing was replaced by the revolutionary Beechcraft Bonanza with a distinctive V-tail. Perhaps the best known Beech aircraft, the single-engined Bonanza has been manufactured in various models since 1947. The Bonanza has had the longest production run of any airplane, past or present, in the world. Other important Beech aircraft are the King Air and Super King Air line of twin-engined turboprops, in production since 1964, the Baron, a twin-engined variant of the Bonanza, and the Beechcraft Model 18, originally a business transport and commuter airliner from the late 1930s through the 1960s, which remains in active service as a cargo transport.
In 1950, Olive Ann Beech was installed as president and CEO of the company, after the sudden death of her husband from a heart attack on November 29 of that year. She continued as CEO until Beech was purchased by Raytheon Company on February 8, 1980. Ted Wells had been replaced as chief engineer by Herbert Rawdon, who remained at the post until his retirement in the early 1960s.
Throughout much of the 20th century, Beechcraft was considered one of the "Big Three" in the field of general aviation manufacturing, along with Cessna and Piper Aircraft.
In 1994, Raytheon merged Beechcraft with the Hawker product line it had acquired in 1993 from British Aerospace, forming Raytheon Aircraft Company. In 2002, the Beechcraft brand was revived to again designate the Wichita-produced aircraft. In 2006, Raytheon sold Raytheon Aircraft to Goldman Sachs creating Hawker Beechcraft. Since its inception Beechcraft has resided in Wichita, Kansas, also the home of chief competitor Cessna, the birthplace of Learjet and of Stearman, whose trainers were used in large numbers during WW II.
The entry into bankruptcy of Hawker Beechcraft on May 3, 2012 ended with its emergence on February 16, 2013 as a new entity, Beechcraft Corporation, with the Hawker Beechcraft name being retired. The new and much smaller company produce the King Air line of aircraft as well as the T-6 and AT-6 military trainer/attack aircraft, as well as the piston-powered single-engined Bonanza and twin-engined Baron aircraft. The jet line was discontinued, but the new company continues to support the aircraft already produced with parts, plus engineering and airworthiness documentation.
By October 2013, the company, now financially turned around, was up for sale.
On December 26, 2013, Textron agreed to purchase Beechcraft, including the discontinued Hawker jet line, for $1.4 billion. The sale was concluded in the first half of 2014, with government approval. Textron CEO Scott Donnelly indicated that Beechcraft and Cessna would be combined to form a new light aircraft manufacturing concern, Textron Aviation, that would result in US$65M-$85M in annual savings over keeping the companies separate. Textron has kept both the Beechcraft and Cessna names as separate brands.
As of January 2019, Beechcraft is producing the following models:
Beech Factory Airport - Houses Beechcraft's head office, manufacturing facility, and runway for test flights | https://en.wikipedia.org/wiki?curid=4282 |
Battle of Peleliu
The Battle of Peleliu, codenamed Operation Stalemate II by the United States military, was fought between the U.S. and Japan during the Mariana and Palau Campaign of World War II, from September to November 1944, on the island of Peleliu.
U.S. Marines of the 1st Marine Division, and later soldiers of the U.S. Army's 81st Infantry Division, fought to capture an airstrip on the small coral island of Peleliu. This battle was part of a larger offensive campaign known as Operation Forager, which ran from June to November 1944, in the Pacific Theater.
Major General William Rupertus, Commander of the 1st Marine Division, predicted the island would be secured within four days. However, after repeated Imperial Army defeats in previous island campaigns, Japan had developed new island-defense tactics and well-crafted fortifications that allowed stiff resistance, extending the battle through more than two months. The heavily outnumbered Japanese defenders put up such stiff resistance, often fighting to the death in the Emperor's name, that the island became known in Japanese as the "Emperor's Island."
In the United States, this was a controversial battle because of the island's negligible strategic value and the high casualty rate, which exceeded that of all other amphibious operations during the Pacific War. The National Museum of the Marine Corps called it "the bitterest battle of the war for the Marines".
By 1944, American victories in the Southwest and Central Pacific had brought the war closer to Japan, with American bombers able to strike at the Japanese main islands from air bases secured during the Mariana Islands campaign (June–August 1944). There was disagreement among the U.S. Joint Chiefs over two proposed strategies to defeat the Japanese Empire. The strategy proposed by General Douglas MacArthur called for the recapture of the Philippines, followed by the capture of Okinawa, then an attack on the Japanese mainland. Admiral Chester Nimitz favored a more direct strategy of bypassing the Philippines, but seizing Okinawa and Taiwan as staging areas to an attack on the Japanese mainland, followed by the future invasion of Japan's southernmost islands. Both strategies included the invasion of Peleliu, but for different reasons.
The 1st Marine Division had already been chosen to make the assault. President Franklin D. Roosevelt traveled to Pearl Harbor to personally meet both commanders and hear their arguments. MacArthur's strategy was chosen. However, before MacArthur could retake the Philippines, the Palau Islands, specifically Peleliu and Angaur, were to be neutralized and an airfield built to protect MacArthur's right flank.
By 1944, Peleliu Island was occupied by about 11,000 Japanese of the 14th Infantry Division with Korean and Okinawan labourers. Colonel Kunio Nakagawa, commander of the division's 2nd Regiment, led the preparations for the island's defense.
After their losses in the Solomons, Gilberts, Marshalls, and Marianas, the Imperial Army assembled a research team to develop new island-defense tactics. They chose to abandon the old strategy of stopping the enemy at the beach, where they were exposed to naval gunfire. The new tactics would only disrupt the landings at the water's edge and depend on an in-depth defense farther inland. Colonel Nakagawa used the rough terrain to his advantage, by constructing a system of heavily fortified bunkers, caves, and underground positions all interlocked into a "honeycomb" system. The traditional "banzai charge" attack was also discontinued as being both wasteful of men and ineffective. These changes would force the Americans into a war of attrition, requiring increasingly more resources.
Nakagawa's defenses were centred on Peleliu's highest point, Umurbrogol Mountain, a collection of hills and steep ridges located at the center of Peleliu overlooking a large portion of the island, including the crucial airfield. The Umurbrogol contained some 500 limestone caves, interconnected by tunnels. Many of these were former mine shafts that were turned into defensive positions. Engineers added sliding armored steel doors with multiple openings to serve both artillery and machine guns. Cave entrances were opened or altered to be slanted as a defense against grenade and flamethrower attacks. The caves and bunkers were connected to a vast tunnel and trench system throughout central Peleliu, which allowed the Japanese to evacuate or reoccupy positions as needed, and to take advantage of shrinking interior lines.
The Japanese were well armed with and mortars and anti-aircraft cannons, backed by a light tank unit and an anti-aircraft detachment.
The Japanese also used the beach terrain to their advantage. The northern end of the landing beaches faced a coral promontory that overlooked the beaches from a small peninsula, a spot later known to the Marines who assaulted it simply as "The Point". Holes were blasted into the ridge to accommodate a gun, and six 20 mm cannons. The positions were then sealed shut, leaving just a small slit to fire on the beaches. Similar positions were crafted along the stretch of landing beaches.
The beaches were also filled with thousands of obstacles for the landing craft, principally mines and a large number of heavy artillery shells buried with the fuses exposed to explode when they were run over. A battalion was placed along the beach to defend against the landing, but they were meant to merely delay the inevitable American advance inland.
Unlike the Japanese, who drastically altered their tactics for the upcoming battle, the American invasion plan was unchanged from that of previous amphibious landings, even after suffering 3,000 casualties and two months of delaying tactics against the entrenched Japanese defenders at the Battle of Biak. On Peleliu, American planners chose to land on the southwest beaches because of their proximity to the airfield on South Peleliu. The 1st Marine Regiment, commanded by Colonel Lewis B. (Chesty) Puller, was to land on the northern end of the beaches. The 5th Marine Regiment, under Colonel Harold D. Harris, would land in the center, and the 7th Marine Regiment, under Col. Herman H. Hanneken, would land at the southern end.
The division's artillery regiment, the 11th Marines under Col. William H. Harrison, would land after the infantry regiments. The plan was for the 1st and 7th Marines to push inland, guarding the 5th Marines left and right flank, and allowing them to capture the airfield located directly to the center of the landing beaches. The 5th Marines were to push to the eastern shore, cutting the island in half. The 1st Marines would push north into the Umurbrogol, while the 7th Marines would clear the southern end of the island. Only one battalion was left behind in reserve, with the U.S. Army's 81st Infantry Division available for support from Angaur, just south of Peleliu.
On September 4, the Marines shipped off from their station on Pavuvu, just north of Guadalcanal, a trip across the Pacific to Peleliu. A U.S. Navy's Underwater Demolition Team went in first to clear the beaches of obstacles, while Navy warships began their pre-invasion bombardment of Peleliu on September 12.
The battleships , , , and , heavy cruisers , , and , and light cruisers , and , led by the command ship , subjected the tiny island, only in size, to a massive three-day bombardment, pausing only to permit air strikes from the three aircraft carriers, five light aircraft carriers, and eleven escort carriers with the attack force. A total of 519 rounds of shells, 1,845 rounds of shells and 1,793 bombs were dropped on the islands during this period.
The Americans believed the bombardment to be successful, as Rear Admiral Jesse Oldendorf claimed that the Navy had run out of targets. In reality, the majority of the Japanese positions were completely unharmed. Even the battalion left to defend the beaches was virtually unscathed. During the assault, the island's defenders exercised unusual firing discipline to avoid giving away their positions. The bombardment managed only to destroy Japan's aircraft on the island, as well as the buildings surrounding the airfield. The Japanese remained in their fortified positions, ready to attack the American landing troops.
United States Pacific Fleet
Admiral Chester W. Nimitz
US Third Fleet
Admiral William F. Halsey, Jr.
Joint Expeditionary Force (Task Force 31)
Vice Admiral Theodore S. Wilkinson
Expeditionary Troops (Task Force 36)
III Amphibious Corps
Major General Julian C. Smith, USMC
Western Landing Force (TG 36.1)
Major General Roy S. Geiger, USMC
1st Marine Division
Beach assignments
Palau District Group
Lieutenant General Inoue Sadao (HQ on Koror Island)
Vice Admiral Yoshioka Ito
Maj. Gen. Kenjiro Murai
14th Division (Lt. Gen. Sadao)
Peleliu Sector Unit (Lt. Col. Kunio Nakagawa)
U.S. Marines landed on Peleliu at 08:32, on September 15, the 1st Marines to the north on White Beach 1 and 2 and the 5th and 7th Marines to the center and south on Orange Beach 1, 2, and 3. As the other landing craft approached the beaches, the Marines were caught in a crossfire when the Japanese opened the steel doors guarding their positions and fired artillery. The positions on the coral promontories guarding each flank fired on the Marines with 47 mm guns and 20 mm cannons. By 09:30, the Japanese had destroyed 60 LVTs and DUKWs.
The 1st Marines were quickly bogged down by heavy fire from the extreme left flank and a 30-foot-high coral ridge, "The Point". Colonel Chesty Puller narrowly escaped death when a dud high velocity artillery round struck his LVT. His communications section was destroyed on its way to the beach by a hit from a 47 mm round. The 7th Marines faced a cluttered Orange Beach 3, with natural and man-made obstacles, forcing the Amtracs to approach in column.
The 5th Marines made the most progress on the first day, aided by cover provided by coconut groves. They pushed toward the airfield, but were met with Nakagawa's first counterattack. His armored tank company raced across the airfield to push the Marines back, but was soon engaged by tanks, howitzers, naval guns, and dive bombers. Nakagawa's tanks and escorting infantrymen were quickly destroyed.
At the end of the first day, the Americans held their stretch of landing beaches, but little else. Their biggest push in the south moved inland, but the 1st Marines to the north made very little progress because of the extremely thick resistance. The Marines had suffered 200 dead and 900 wounded. Rupertus, still unaware of his enemy's change of tactics, believed the Japanese would quickly crumble since their perimeter had been broken.
On the second day, the 5th Marines moved to capture the airfield and push toward the eastern shore. They ran across the airfield, enduring heavy artillery fire from the highlands to the north, suffering heavy casualties in the process. After capturing the airfield, they rapidly advanced to the eastern end of Peleliu, leaving the island's southern defenders to be destroyed by the 7th Marines.
This area was hotly contested by the Japanese, who still occupied numerous pillboxes. Heat indices were around , and the Marines soon suffered high casualties from heat exhaustion. Further complicating the situation, the Marines' water was distributed in empty oil drums, contaminating the water with the oil residue. Still, by the eighth day the 5th and 7th Marines had accomplished their objectives, holding the airfield and the southern portion of the island, although the airfield remained under threat of sustained Japanese fire from the heights of Umurbrogol Mountain until the end of the battle.
American forces put the airfield to use on the third day. L-2 Grasshoppers from VMO-3 began aerial spotting missions for Marine artillery and naval gunfire support. On September 26 (D+11), Marine F4U Corsairs from VMF-114 landed on the airstrip. The Corsairs began dive-bombing missions across Peleliu, firing rockets into open cave entrances for the infantrymen, and dropping napalm; it was only the second time the latter weapon had been used in the Pacific. Napalm proved useful, burning away the vegetation hiding spider holes and usually killing their occupants.
The time from liftoff to the target area for the Corsairs based on Peleliu Airfield was very short, sometimes only 10 to 15 seconds. Consequently, there was almost no time for pilots to raise their aircraft undercarriage; most pilots did not bother and left them down during the air strike. After the air strike was completed and the payload dropped, the Corsair simply turned back into the landing pattern again.
The fortress at the end of the southern landing beaches (a.k.a. “The Point”) continued to cause heavy Marine casualties due to enfilading fire from Japanese heavy machine guns and anti-tank artillery across the landing beaches. Puller ordered Captain George P. Hunt, commander of K Company, 3rd Battalion, 1st Marines, to capture the position. Hunt's company approached The Point short on supplies, having lost most of its machine guns while approaching the beaches. Hunt's second platoon was pinned down for nearly a day in an anti-tank trench between fortifications. The rest of his company was endangered when the Japanese cut a hole in their line, surrounding his company and leaving his right flank cut off.
However, a rifle platoon began knocking out the Japanese gun positions one by one. Using smoke grenades for concealment, the platoon swept through each hole, destroying the positions with rifle grenades and close-quarters combat. After knocking out the six machine gun positions, the Marines faced the 47 mm gun cave. A lieutenant blinded the 47 mm gunner's visibility with a smoke grenade, allowing Corporal Henry W. Hahn to launch a grenade through the cave's aperture. The grenade detonated the 47 mm's shells, forcing the cave's occupants out with their bodies alight and their ammunition belts exploding around their waists. A Marine fire team was positioned on the flank of the cave where the emerging occupants were shot down.
K Company had captured The Point, but Nakagawa counterattacked. The next 30 hours saw four major counterattacks against a sole company, critically low on supplies, out of water, and surrounded. The Marines soon had to resort to hand-to-hand combat to fend off the Japanese attackers. By the time reinforcements arrived, the company had successfully repulsed all of the Japanese attacks, but had been reduced to 18 men, suffering 157 casualties during the battle for The Point. Hunt and Hahn were both awarded the Navy Cross for their actions.
The 5th Marines—after having secured the airfield—were sent to capture Ngesebus Island, just north of Peleliu. Ngesebus was occupied by many Japanese artillery positions, and was the site of an airfield still under construction. The tiny island was connected to Peleliu by a small causeway, but 5th Marines commander Harris opted instead to make a shore-to-shore amphibious landing, predicting the causeway to be an obvious target for the island's defenders.
Harris coordinated a pre-landing bombardment of the island on September 28, carried out by Army guns, naval guns, howitzers from the 11th Marines, strafing runs from VMF-114's Corsairs, and fire from the approaching LVTs. Unlike the Navy's bombardment of Peleliu, Harris' assault on Ngesebus successfully killed most of the Japanese defenders. The Marines still faced opposition in the ridges and caves, but the island fell quickly, with relatively light casualties for the 5th Marines. They had suffered 15 killed and 33 wounded, and inflicted 470 casualties on the Japanese.
After capturing The Point, the 1st Marines moved north into the Umurbrogol pocket, named "Bloody Nose Ridge" by the Marines. Puller led his men in numerous assaults, but each resulted in severe casualties from Japanese fire. The 1st Marines were trapped in the narrow paths between the ridges, with each ridge fortification supporting the other with deadly crossfire.
The Marines took increasingly high casualties as they slowly advanced through the ridges. The Japanese again showed unusual fire discipline, striking only when they could inflict maximum casualties. As casualties mounted, Japanese snipers began to take aim at stretcher bearers, knowing that if stretcher bearers were injured or killed, more would have to return to replace them, and the snipers could steadily pick off more and more Marines. The Japanese also infiltrated the American lines at night to attack the Marines in their fighting holes. The Marines built two-man fighting holes, so one Marine could sleep while the other kept watch for infiltrators.
One particularly bloody battle on Bloody Nose came when the 1st Battalion, 1st Marines—under the command of Major Raymond Davis—attacked Hill 100. Over six days of fighting, the battalion suffered 71% casualties. Captain Everett Pope and his company penetrated deep into the ridges, leading his remaining 90 men to seize what he thought was Hill 100. It took a day's fighting to reach what he thought was the crest of the hill, which was in fact another ridge occupied by more Japanese defenders.
Trapped at the base of the ridge, Captain Pope set up a small defense perimeter, which was attacked relentlessly by the Japanese throughout the night. The Marines soon ran out of ammunition, and had to fight the attackers with knives and fists, even resorting to throwing coral rock and empty ammunition boxes at the Japanese. Pope and his men managed to hold out until dawn came, which brought on more deadly fire. When they evacuated the position, only nine men remained. Pope later received the Medal of Honor for the action. (Picture of the Peleliu Memorial dedicated on the 50th anniversary of the landing on Peleliu with Captain Pope's name)
The Japanese eventually inflicted 70% casualties on Puller's 1st Marines, or 1,749 men. After six days of fighting in the ridges of Umurbrogol, General Roy Geiger, commander of the III Amphibious Corps, sent elements of U.S. Army's 81st Infantry Division to Peleliu to relieve the regiment. The 321st Regiment Combat Team landed on the western beaches of Peleliu—at the northern end of Umurbrogol mountain—on 23 September. The 321st and the 7th Marines encircled The Pocket by 24 Sept., D+9.
By 15 October, the 7th Marines had suffered 46% casualties and General Geiger relieved them with the 5th Marines. Col. Harris adopted siege tactics, using bulldozers and flame-thrower tanks, pushing from the north. On October 30, the 81st Infantry Division took over command of Peleliu, taking another six weeks, with the same tactics, to reduce The Pocket.
On 24 November, Nakagawa proclaimed "Our sword is broken and we have run out of spears". He then burnt his regimental colors and performed ritual suicide. He was posthumously promoted to lieutenant general for his valor displayed on Peleliu. On 27 November, the island was declared secure, ending the 73-day-long battle.
A Japanese lieutenant with twenty-six 2nd Infantry soldiers and eight 45th Guard Force sailors held out in the caves in Peleliu until April 22, 1947, and surrendered after a Japanese admiral convinced them the war was over.
The reduction of the Japanese pocket around Umurbrogol mountain has been called the most difficult fight that the U.S. military encountered in the entire war. The 1st Marine Division was severely mauled and it remained out of action until the invasion of Okinawa began on April 1, 1945. In total, the 1st Marine Division suffered over 6,500 casualties during their month on Peleliu, over one third of their entire division. The 81st Infantry Division also suffered heavy losses with 3,300 casualties during their tenure on the island.
Postwar statisticians calculated that it took U.S. forces over 1500 rounds of ammunition to kill each Japanese defender and that, during the course of the battle, the Americans expended 13.32 million rounds of .30-calibre, 1.52 million rounds of .45-calibre, 693,657 rounds of .50-calibre bullets, 118,262 hand grenades, and approximately 150,000 mortar rounds.
The battle was controversial in the United States due to the island's lack of strategic value and the high casualty rate. The defenders lacked the means to interfere with potential US operations in the Philippines and the airfield captured on Peleliu did not play a key role in subsequent operations. Instead, the Ulithi Atoll in the Caroline Islands was used as a staging base for the invasion of Okinawa. The high casualty rate exceeded all other amphibious operations during the Pacific War.
In addition, few news reports were published about the battle because Rupertus' prediction of a "three days" victory motivated only six reporters to report from shore. The battle was also overshadowed by MacArthur's return to the Philippines and the Allies' push towards Germany in Europe.
The battles for Angaur and Peleliu showed Americans the pattern of future Japanese island defense but they made few adjustments for the Iwo Jima and Okinawa. Naval bombardment prior to amphibious assault at Iwo Jima was only slightly more effective than at Peleliu, but at Okinawa the preliminary shelling was much improved. Frogmen performing underwater demolition at Iwo Jima confused the enemy by sweeping both coasts, but later alerted Japanese defenders to the exact assault beaches at Okinawa. American ground forces at Peleliu gained experience in assaulting heavily fortified positions such as they would find again at Okinawa.
On the recommendation of Admiral William F. Halsey, Jr., the planned occupation of Yap Island in the Caroline Islands was canceled. Halsey actually recommended that the landings on Peleliu and Angaur be canceled, too, and their Marines and soldiers be thrown into Leyte Island instead, but was overruled by Nimitz.
In the March of Time's 1951 documentary TV series, Crusade in the Pacific, Episode 17 is "The Fight for Bloody Nose Ridge."
In NBC-TV's 1952-53 documentary TV series Victory at Sea, Episode 18, "Two if by Sea" covers the assaults at Peleliu and Angaur.
The Battle of Peleliu is featured in many World War II themed video games, including "". The player takes the role of a US Marine tasked with taking Peleliu Airfield, repelling counter-attacks, destroying machine-gun and mortar positions and eventually securing Japanese artillery emplacements at the point. In flight-simulation game "War Thunder", two teams of players clash to hold the southern and northern airfields. In multi-player shooter "Red Orchestra 2: Rising Storm", a team of American troops attack the defensive Japanese team's control points.
The battle including footage and stills are featured in the fifth episode of "Ken Burns' The War".
The battle features in episodes 5, 6 and 7 of the TV mini-series "The Pacific".
In his book, "With the Old Breed", Eugene Bondurant Sledge described his experiences in the battle for Peleliu.
In 2015, the Japanese magazine "Young Animal" commenced serialization of "" by Masao Hiratsuka and artist Kazuyoshi Takeda, telling the story of the battle in manga form.
One of the final scenes in "Parer's War", a 2014 Australian television film, shows the Battle of Peleliu recorded by Damien Parer with his camera at the time of his death.
The Peleliu Campaign features as one of the campaigns in the 2019 solitaire tactical wargame “Fields of Fire” Volume 2, designed by Ben Hull, published by GMT Games LLC.
For heroism: | https://en.wikipedia.org/wiki?curid=4283 |
Battle of Stalingrad
In the Battle of Stalingrad (23 August 1942 – 2 February 1943), Germany and its allies fought the Soviet Union for control of the city of Stalingrad (now Volgograd) in Southern Russia. Marked by fierce close-quarters combat and direct assaults on civilians in air raids, it is one of the bloodiest battles in the history of warfare, with an estimated 2 million total casualties. After their defeat at Stalingrad, the German High Command had to withdraw considerable military forces from the Western Front to replace their losses.
The German offensive to capture Stalingrad began in August 1942, using the 6th Army and elements of the 4th Panzer Army. The attack was supported by intense "Luftwaffe" bombing that reduced much of the city to rubble. The fighting degenerated into house-to-house fighting, as both sides poured reinforcements into the city. By mid-November, the Germans had pushed the Soviet defenders back at great cost into narrow zones along the west bank of the Volga River.
On 19 November, the Red Army launched Operation Uranus, a two-pronged attack targeting the weaker Romanian and Hungarian armies protecting the 6th Army's flanks. The Axis forces on the flanks were overrun and the 6th Army was cut off and surrounded in the Stalingrad area. Adolf Hitler ordered that the 6th Army make no attempt to break out; instead, attempts were made to supply it by air and to break the encirclement from the outside. Heavy fighting continued for another two months. At the beginning of February 1943, the Axis forces in Stalingrad, having exhausted their ammunition and food, surrendered after five months, one week and three days of fighting.
By the spring of 1942, despite the failure of Operation Barbarossa to decisively defeat the Soviet Union in a single campaign, the Wehrmacht had captured vast expanses of territory, including Ukraine, Belarus, and the Baltic republics. Elsewhere, the war had been progressing well: the U-boat offensive in the Atlantic had been very successful and Erwin Rommel had just captured Tobruk. In the east, they had stabilised their front in a line running from Leningrad in the north to Rostov in the south. There were a number of salients, but these were not particularly threatening. Hitler was confident that he could master the Red Army after the winter of 1942, because even though Army Group Centre ("Heeresgruppe Mitte") had suffered heavy losses west of Moscow the previous winter, 65% of its infantry had not been engaged and had been rested and re-equipped. Neither Army Group North nor Army Group South had been particularly hard pressed over the winter. Stalin was expecting the main thrust of the German summer attacks to be directed against Moscow again.
With the initial operations being very successful, the Germans decided that their summer campaign in 1942 would be directed at the southern parts of the Soviet Union. The initial objectives in the region around Stalingrad were the destruction of the industrial capacity of the city and the deployment of forces to block the Volga River. The river was a key route from the Caucasus and the Caspian Sea to central Russia. Its capture would disrupt commercial river traffic. The Germans cut the pipeline from the oilfields when they captured Rostov on 23 July. The capture of Stalingrad would make the delivery of Lend Lease supplies via the Persian Corridor much more difficult.
On 23 July 1942, Hitler personally rewrote the operational objectives for the 1942 campaign, greatly expanding them to include the occupation of the city of Stalingrad. Both sides began to attach propaganda value to the city, based on it bearing the name of the leader of the Soviet Union. Hitler proclaimed that after Stalingrad's capture, its male citizens were to be killed and all women and children were to be deported because its population was "thoroughly communistic" and "especially dangerous". It was assumed that the fall of the city would also firmly secure the northern and western flanks of the German armies as they advanced on Baku, with the aim of securing these strategic petroleum resources for Germany. The expansion of objectives was a significant factor in Germany's failure at Stalingrad, caused by German overconfidence and an underestimation of Soviet reserves.
The Soviets realised that they were pressed for time and resources. They ordered that anyone strong enough to hold a rifle be sent to fight.
Army Group South was selected for a sprint forward through the southern Russian steppes into the Caucasus to capture the vital Soviet oil fields there. The planned summer offensive, code-named "Fall Blau" (Case Blue), was to include the German 6th, 17th, 4th Panzer and 1st Panzer Armies. Army Group South had overrun the Ukrainian Soviet Socialist Republic in 1941. Poised in Eastern Ukraine, it was to spearhead the offensive.
Hitler intervened, however, ordering the Army Group to split in two. Army Group South (A), under the command of Wilhelm List, was to continue advancing south towards the Caucasus as planned with the 17th Army and First Panzer Army. Army Group South (B), including Friedrich Paulus's 6th Army and Hermann Hoth's 4th Panzer Army, was to move east towards the Volga and Stalingrad. Army Group B was commanded by General Maximilian von Weichs.
The start of "Case Blue" had been planned for late May 1942. However, a number of German and Romanian units that were to take part in "Blau" were besieging Sevastopol on the Crimean Peninsula. Delays in ending the siege pushed back the start date for "Blau" several times, and the city did not fall until early July.
Operation Fridericus I by the Germans against the "Isium bulge", pinched off the Soviet salient in the Second Battle of Kharkov, and resulted in the envelopment of a large Soviet force between 17 May and 29 May. Similarly, Operation Wilhelm attacked Voltshansk on 13 June, and Operation Fridericus attacked Kupiansk on 22 June.
"Blau" finally opened as Army Group South began its attack into southern Russia on 28 June 1942. The German offensive started well. Soviet forces offered little resistance in the vast empty steppes and started streaming eastward. Several attempts to re-establish a defensive line failed when German units outflanked them. Two major pockets were formed and destroyed: the first, northeast of Kharkov, on 2 July, and a second, around Millerovo, Rostov Oblast, a week later. Meanwhile, the Hungarian 2nd Army and the German 4th Panzer Army had launched an assault on Voronezh, capturing the city on 5 July.
The initial advance of the 6th Army was so successful that Hitler intervened and ordered the 4th Panzer Army to join Army Group South (A) to the south. A massive traffic jam resulted when the 4th Panzer and the 1st Panzer both required the few roads in the area. Both armies were stopped dead while they attempted to clear the resulting mess of thousands of vehicles. The delay was long, and it is thought that it cost the advance at least one week. With the advance now slowed, Hitler changed his mind and reassigned the 4th Panzer Army back to the attack on Stalingrad.
By the end of July, the Germans had pushed the Soviets across the Don River. At this point, the Don and Volga Rivers are only apart, and the Germans left their main supply depots west of the Don, which had important implications later in the course of the battle. The Germans began using the armies of their Italian, Hungarian and Romanian allies to guard their left (northern) flank. Occasionally Italian actions were mentioned in official German communiques. Italian forces were generally held in little regard by the Germans, and were accused of having low morale: in reality, the Italian divisions fought comparatively well, with the 3rd Mountain Infantry Division Ravenna and 5th Infantry Division Cosseria proving to have good morale, according to a German liaison officer, and were forced to retreat only after a massive armoured attack in which German reinforcements had failed to arrive in time, according to a German historian. Indeed the Italians distinguished themselves in numerous battles, such as the Battle of Nikolayevka.
On 25 July the Germans faced stiff resistance with a Soviet bridgehead west of Kalach. "We had had to pay a high cost in men and material...left on the Kalach battlefield were numerous burnt-out or shot-up German tanks."
The Germans formed bridgeheads across the Don on 20 August, with the 295th and 76th Infantry Divisions enabling the XIVth Panzer Corps "to thrust to the Volga north of Stalingrad." The German 6th Army was only a few dozen kilometres from Stalingrad. The 4th Panzer Army, ordered south on 13 July to block the Soviet retreat "weakened by the 17th Army and the 1st Panzer Army", had turned northwards to help take the city from the south.
To the south, Army Group A was pushing far into the Caucasus, but their advance slowed as supply lines grew overextended. The two German army groups were not positioned to support one another due to the great distances involved.
After German intentions became clear in July 1942, Stalin appointed General Andrey Yeryomenko as commander of the Southeastern Front on 1 August 1942. Yeryomenko and Commissar Nikita Khrushchev were tasked with planning the defence of Stalingrad. The eastern border of Stalingrad was the wide River Volga, and over the river, additional Soviet units were deployed. These units became the newly formed 62nd Army, which Yeryomenko placed under the command of Lieutenant General Vasiliy Chuikov on 11 September 1942. The situation was extremely dire. When asked how he interpreted his task, he responded, "We will defend the city or die in the attempt." The 62nd Army's mission was to defend Stalingrad at all costs. Chuikov's generalship during the battle earned him one of his two Hero of the Soviet Union awards.
During the defence of Stalingrad, the Red Army deployed five armies (28th, 51st, 57th, 62nd and 64th Armies) in and around the city and an additional nine armies in the encirclement counter offensive. The nine armies amassed for the counter offensive were the 24th Army, 65th Army, 66th Army and 16th Air Army from the north as part of the Don Front offensive and 1st Guards Army, 5th Tank, 21st Army, 2nd Air Army and 17th Air Army from the south as part of the Southwestern Front.
David Glantz indicated that four hard-fought battles – collectively known as the Kotluban Operations – north of Stalingrad, where the Soviets made their greatest stand, decided Germany's fate before the Nazis ever set foot in the city itself, and were a turning point in the war. Beginning in late August, continuing in September and into October, the Soviets committed between two and four armies in hastily coordinated and poorly controlled attacks against the Germans' northern flank. The actions resulted in more than 200,000 Soviet Army casualties but did slow the German assault.
On 23 August the 6th Army reached the outskirts of Stalingrad in pursuit of the 62nd and 64th Armies, which had fallen back into the city. Kleist later said after the war:
The Soviets had enough warning of the German advance to ship grain, cattle, and railway cars across the Volga and out of harm's way, but most civilian residents were not evacuated. This "harvest victory" left the city short of food even before the German attack began. Before the "Heer" reached the city itself, the "Luftwaffe" had rendered the River Volga, vital for bringing supplies into the city, unusable to Soviet shipping. Between 25 and 31 July, 32 Soviet ships were sunk, with another nine crippled.
The battle began with the heavy bombing of the city by "Generaloberst" Wolfram von Richthofen's "Luftflotte 4", which in the summer and autumn of 1942 was the single most powerful air formation in the world. Some 1,000 tons of bombs were dropped in 48 hours, more than in London at the height of the Blitz. Stalin refused to evacuate civilian population from the city, so when bombing began 400,000 civilians were trapped within city boundaries. The exact number of civilians killed during the course of the battle is unknown but was most likely very high. Around 40,000 civilians were moved to Germany as slave workers, some fled the city during battle and a small number were evacuated by the Soviets. In February 1943 only between 10,000 to 60,000 civilians were still alive in Stalingrad. Much of the city was quickly turned to rubble, although some factories continued production while workers joined in the fighting. The Stalingrad Tractor Factory continued to turn out T-34 tanks literally until German troops burst into the plant. The 369th (Croatian) Reinforced Infantry Regiment was the only non-German unit selected by the "Wehrmacht" to enter Stalingrad city during assault operations. It fought as part of the 100th Jäger Division.
Stalin rushed all available troops to the east bank of the Volga, some from as far away as Siberia. All the regular ferries were quickly destroyed by the Luftwaffe, which then targeted troop barges being towed slowly across the river by tugs. It has been said that Stalin prevented civilians from leaving the city in the belief that their presence would encourage greater resistance from the city's defenders. Civilians, including women and children, were put to work building trenchworks and protective fortifications. A massive German air raid on 23 August caused a firestorm, killing hundreds and turning Stalingrad into a vast landscape of rubble and burnt ruins. Ninety percent of the living space in the Voroshilovskiy area was destroyed. Between 23 and 26 August, Soviet reports indicate 955 people were killed and another 1,181 wounded as a result of the bombing. Casualties of 40,000 were greatly exaggerated, and after 25 August the Soviets did not record any civilian and military casualties as a result of air raids.
The Soviet Air Force, the "Voyenno-Vozdushnye Sily" (VVS), was swept aside by the Luftwaffe. The VVS bases in the immediate area lost 201 aircraft between 23 and 31 August, and despite meagre reinforcements of some 100 aircraft in August, it was left with just 192 serviceable aircraft, 57 of which were fighters. The Soviets continued to pour aerial reinforcements into the Stalingrad area in late September, but continued to suffer appalling losses; the "Luftwaffe" had complete control of the skies.
The burden of the initial defence of the city fell on the 1077th Anti-Aircraft Regiment, a unit made up mainly of young female volunteers who had no training for engaging ground targets. Despite this, and with no support available from other units, the AA gunners stayed at their posts and took on the advancing panzers. The German 16th Panzer Division reportedly had to fight the 1077th's gunners "shot for shot" until all 37 anti-aircraft guns were destroyed or overrun. The 16th Panzer was shocked to find that, due to Soviet manpower shortages, it had been fighting female soldiers. In the early stages of the battle, the NKVD organised poorly armed "Workers' militias" similar to those that had defended the city twenty-four years earlier, composed of civilians not directly involved in war production for immediate use in the battle. The civilians were often sent into battle without rifles. Staff and students from the local technical university formed a "tank destroyer" unit. They assembled tanks from leftover parts at the tractor factory. These tanks, unpainted and lacking gun-sights, were driven directly from the factory floor to the front line. They could only be aimed at point-blank range through the bore of their gun barrels.
By the end of August, Army Group South (B) had finally reached the Volga, north of Stalingrad. Another advance to the river south of the city followed, while the Soviets abandoned their Rossoshka position for the inner defensive ring west of Stalingrad. The wings of the 6th Army and the 4th Panzer Army met near Jablotchni along the Zaritza on 2 Sept. By 1 September, the Soviets could only reinforce and supply their forces in Stalingrad by perilous crossings of the Volga under constant bombardment by artillery and aircraft.
On 5 September, the Soviet 24th and 66th Armies organized a massive attack against XIV Panzer Corps. The "Luftwaffe" helped repel the offensive by heavily attacking Soviet artillery positions and defensive lines. The Soviets were forced to withdraw at midday after only a few hours. Of the 120 tanks the Soviets had committed, 30 were lost to air attack.
Soviet operations were constantly hampered by the "Luftwaffe". On 18 September, the Soviet 1st Guards and 24th Army launched an offensive against VIII Army Corps at Kotluban. "VIII. Fliegerkorps" dispatched wave after wave of Stuka dive-bombers to prevent a breakthrough. The offensive was repelled. The Stukas claimed 41 of the 106 Soviet tanks knocked out that morning, while escorting Bf 109s destroyed 77 Soviet aircraft.
Amid the debris of the wrecked city, the Soviet 62nd and 64th Armies, which included the Soviet 13th Guards Rifle Division, anchored their defence lines with strong-points in houses and factories.
Fighting within the ruined city was fierce and desperate. Lieutenant General Alexander Rodimtsev was in charge of the 13th Guards Rifle Division, and received one of two Heroes of the Soviet Union awarded during the battle for his actions. Stalin's Order No. 227 of 27 July 1942 decreed that all commanders who ordered unauthorised retreats would be subject to a military tribunal. Deserters and presumed malingerers were captured or executed after fighting. During the battle the 62nd Army had the most arrests and executions: 203 in all, of which 49 were executed, while 139 were sent to penal companies and battalions. The Germans pushing forward into Stalingrad suffered heavy casualties.
By 12 September, at the time of their retreat into the city, the Soviet 62nd Army had been reduced to 90 tanks, 700 mortars and just 20,000 personnel. The remaining tanks were used as immobile strong-points within the city. The initial German attack on 14 September attempted to take the city in a rush. The 51st Army Corps' 295th Infantry Division went after the Mamayev Kurgan hill, the 71st attacked the central rail station and toward the central landing stage on the Volga, while 48th Panzer Corps attacked south of the Tsaritsa River. Rodimtsev's 13th Guards Rifle Division had been hurried up to cross the river and join the defenders inside the city. Assigned to counterattack at the Mamayev Kurgan and at Railway Station No. 1, it suffered particularly heavy losses.
Though initially successful, the German attacks stalled in the face of Soviet reinforcements brought in from across the Volga. The Soviet 13th Guards Rifle Division, assigned to counterattack at the Mamayev Kurgan and at Railway Station No. 1, suffered particularly heavy losses. Over 30 percent of its soldiers were killed in the first 24 hours, and just 320 out of the original 10,000 survived the entire battle. Both objectives were retaken, but only temporarily. The railway station changed hands 14 times in six hours. By the following evening, the 13th Guards Rifle Division had ceased to exist.
Combat raged for three days at the giant grain elevator in the south of the city. About fifty Red Army defenders, cut off from resupply, held the position for five days and fought off ten different assaults before running out of ammunition and water. Only forty dead Soviet fighters were found, though the Germans had thought there were many more due to the intensity of resistance. The Soviets burned large amounts of grain during their retreat in order to deny the enemy food. Paulus chose the grain elevator and silos as the symbol of Stalingrad for a patch he was having designed to commemorate the battle after a German victory.
German military doctrine was based on the principle of combined-arms teams and close cooperation between tanks, infantry, engineers, artillery and ground-attack aircraft. Some Soviet commanders adopted the tactic of always keeping their front-line positions as close to the Germans as physically possible; Chuikov called this "hugging" the Germans. This slowed the German advance and reduced the effectiveness of the German advantage in supporting fire.
The Red Army gradually adopted a strategy to hold for as long as possible all the ground in the city. Thus, they converted multi-floored apartment blocks, factories, warehouses, street corner residences and office buildings into a series of well defended strong-points with small 5–10-man units. Manpower in the city was constantly refreshed by bringing additional troops over the Volga. When a position was lost, an immediate attempt was usually made to re-take it with fresh forces.
Bitter fighting raged for every ruin, street, factory, house, basement, and staircase. Even the sewers were the sites of firefights. The Germans called this unseen urban warfare "Rattenkrieg" ("Rat War"), and bitterly joked about capturing the kitchen but still fighting for the living room and the bedroom. Buildings had to be cleared room by room through the bombed-out debris of residential areas, office blocks, basements and apartment high-rises. Some of the taller buildings, blasted into roofless shells by earlier German aerial bombardment, saw floor-by-floor, close-quarters combat, with the Germans and Soviets on alternate levels, firing at each other through holes in the floors. Fighting on and around Mamayev Kurgan, a prominent hill above the city, was particularly merciless; indeed, the position changed hands many times.
In another part of the city, a Soviet platoon under the command of Sergeant Yakov Pavlov fortified a four-story building that oversaw a square 300 meters from the river bank, later called "Pavlov's House". The soldiers surrounded it with minefields, set up machine-gun positions at the windows and breached the walls in the basement for better communications. The soldiers found about ten Soviet civilians hiding in the basement. They were not relieved, and not significantly reinforced, for two months. The building was labelled "Festung" ("Fortress") on German maps. Sgt. Pavlov was awarded the Hero of the Soviet Union for his actions.
The Germans made slow but steady progress through the city. Positions were taken individually, but the Germans were never able to capture the key crossing points along the river bank. By 27 Sept. the Germans occupied the southern portion of the city, but the Soviets held the centre and northern part. Most importantly, the Soviets controlled the ferries to their supplies on the east bank of the Volga.
The Germans used aircraft, tanks and heavy artillery to clear the city with varying degrees of success. Toward the end of the battle, the gigantic railroad gun nicknamed "Dora" was brought into the area. The Soviets built up a large number of artillery batteries on the east bank of the Volga. This artillery was able to bombard the German positions or at least provide counter-battery fire.
Snipers on both sides used the ruins to inflict casualties. The most famous Soviet sniper in Stalingrad was Vasily Zaytsev, with 225 confirmed kills during the battle. Targets were often soldiers bringing up food or water to forward positions. Artillery spotters were an especially prized target for snipers.
A significant historical debate concerns the degree of terror in the Red Army. The British historian Antony Beevor noted the "sinister" message from the Stalingrad Front's Political Department on 8 October 1942 that: "The defeatist mood is almost eliminated and the number of treasonous incidents is getting lower" as an example of the sort of coercion Red Army soldiers experienced under the Special Detachments (later to be renamed SMERSH). On the other hand, Beevor noted the often extraordinary bravery of the Soviet soldiers in a battle that was only comparable to Verdun, and argued that terror alone cannot explain such self-sacrifice. Richard Overy addresses the question of just how important the Red Army's coercive methods were to the Soviet war effort compared with other motivational factors such as hatred for the enemy. He argues that, though it is "easy to argue that from the summer of 1942 the Soviet army fought because it was forced to fight," to concentrate solely on coercion is nonetheless to "distort our view of the Soviet war effort." After conducting hundreds of interviews with Soviet veterans on the subject of terror on the Eastern Front – and specifically about Order No. 227 ("Not a step back!") at Stalingrad – Catherine Merridale notes that, seemingly paradoxically, "their response was frequently relief." Infantryman Lev Lvovich's explanation, for example, is typical for these interviews; as he recalls, "[i]t was a necessary and important step. We all knew where we stood after we had heard it. And we all – it's true – felt better. Yes, we felt better."
Many women fought on the Soviet side, or were under fire. As General Chuikov acknowledged, "Remembering the defence of Stalingrad, I can't overlook the very important question ... about the role of women in war, in the rear, but also at the front. Equally with men they bore all the burdens of combat life and together with us men, they went all the way to Berlin." At the beginning of the battle there were 75,000 women and girls from the Stalingrad area who had finished military or medical training, and all of whom were to serve in the battle. Women staffed a great many of the anti-aircraft batteries that fought not only the Luftwaffe but German tanks. Soviet nurses not only treated wounded personnel under fire but were involved in the highly dangerous work of bringing wounded soldiers back to the hospitals under enemy fire. Many of the Soviet wireless and telephone operators were women who often suffered heavy casualties when their command posts came under fire. Though women were not usually trained as infantry, many Soviet women fought as machine gunners, mortar operators, and scouts. Women were also snipers at Stalingrad. Three air regiments at Stalingrad were entirely female. At least three women won the title Hero of the Soviet Union while driving tanks at Stalingrad.
For both Stalin and Hitler, Stalingrad became a matter of prestige far beyond its strategic significance. The Soviet command moved units from the Red Army strategic reserve in the Moscow area to the lower Volga, and transferred aircraft from the entire country to the Stalingrad region.
The strain on both military commanders was immense: Paulus developed an uncontrollable tic in his eye, which eventually afflicted the left side of his face, while Chuikov experienced an outbreak of eczema that required him to have his hands completely bandaged. Troops on both sides faced the constant strain of close-range combat.
After 27 September, much of the fighting in the city shifted north to the industrial district. Having slowly advanced over 10 days against strong Soviet resistance, the 51st Army Corps was finally in front of the three giant factories of Stalingrad: the Red October Steel Factory, the Barrikady Arms Factory and Stalingrad Tractor Factory. It took a few more days for them to prepare for the most savage offensive of all, which was unleashed on 14 October with a concentration of gunfire never seen before.
Exceptionally intense shelling and bombing paved the way for the first German assault groups. The main attack (led by the 14th Panzer and 305th Infantry Divisions) attacked towards the tractor factory, while another assault led by the 24th Panzer Division hit to the south of the giant plant.
The German onslaught crushed the 37th Guards Rifle Division of Major General Viktor Zholudev and in the afternoon the forward assault group reached the tractor factory before arriving at the Volga River, splitting the 62nd Army into two. In response to the German breakthrough to the Volga, the front headquarters committed three battalions from the 300th Rifle Division and the 45th Rifle Division of Colonel Vasily Sokolov, a substantial force of over 2,000 men, to the fighting at the Red October Factory.
Fighting raged inside the Barrikady Factory until the end of October. The Soviet-controlled area shrank down to a few strips of land along the western bank of the Volga, and in November the fighting concentrated around what Soviet newspapers referred to as "Lyudnikov's Island", a small patch of ground behind the Barrikady Factory where the remnants of Colonel Ivan Lyudnikov's 138th Rifle Division resisted all ferocious assaults thrown by the Germans and became a symbol of the stout Soviet defence of Stalingrad.
From 5 to 12 September, "Luftflotte 4" conducted 7,507 sorties (938 per day). From 16 to 25 September, it carried out 9,746 missions (975 per day). Determined to crush Soviet resistance, "Luftflotte" 4's "Stukawaffe" flew 900 individual sorties against Soviet positions at the Stalingrad Tractor Factory on 5 October. Several Soviet regiments were wiped out; the entire staff of the Soviet 339th Infantry Regiment was killed the following morning during an air raid.
The "Luftwaffe" retained air superiority into November, and Soviet daytime aerial resistance was nonexistent. However, the combination of constant air support operations on the German side and the Soviet surrender of the daytime skies began to affect the strategic balance in the air. From 28 June to 20 September, "Luftflotte" 4 original strength of 1,600 aircraft, of which 1,155 were operational, fell to 950, of which only 550 were operational. The fleet's total strength decreased by 40 percent. Daily sorties decreased from 1,343 per day to 975 per day. Soviet offensives in the central and northern portions of the Eastern Front tied down Luftwaffe reserves and newly built aircraft, reducing "Luftflotte" 4 percentage of Eastern Front aircraft from 60 percent on 28 June to 38 percent by 20 September. The "Kampfwaffe" (bomber force) was the hardest hit, having only 232 out of an original force of 480 left. The "VVS" remained qualitatively inferior, but by the time of the Soviet counter-offensive, the "VVS" had reached numerical superiority.
In mid-October, after receiving reinforcements from the Caucasus theatre, the "Luftwaffe" intensified its efforts against remaining Red Army positions holding the west bank. "Luftflotte" 4 flew 1,250 sorties on 14 October and its Stukas dropped 550 tonnes of bombs, while German infantry surrounded the three factories. "Stukageschwader" 1, 2, and 77 had largely silenced Soviet artillery on the eastern bank of the Volga before turning their attention to the shipping that was once again trying to reinforce the narrowing Soviet pockets of resistance. The 62nd Army had been cut in two and, due to intensive air attack on its supply ferries, was receiving much less material support. With the Soviets forced into a strip of land on the western bank of the Volga, over 1,208 "Stuka" missions were flown in an effort to eliminate them.
The Soviet bomber force, the "Aviatsiya Dal'nego Deystviya" (Long Range Aviation; ADD), having taken crippling losses over the past 18 months, was restricted to flying at night. The Soviets flew 11,317 night sorties over Stalingrad and the Don-bend sector between 17 July and 19 November. These raids caused little damage and were of nuisance value only.
On 8 November, substantial units from "Luftflotte" 4 were withdrawn to combat the Allied landings in North Africa. The German air arm found itself spread thinly across Europe, struggling to maintain its strength in the other southern sectors of the Soviet-German front.
As historian Chris Bellamy notes, the Germans paid a high strategic price for the aircraft sent into Stalingrad: the "Luftwaffe" was forced to divert much of its air strength away from the oil-rich Caucasus, which had been Hitler's original grand-strategic objective.
The Royal Romanian Air Force was also involved in the Axis air operations at Stalingrad. Starting 23 October 1942, Romanian pilots flew a total of 4,000 sorties, during which they destroyed 61 Soviet aircraft. The Romanian Air Force lost 79 aircraft, most of them captured on the ground along with their airfields.
After three months of slow advance, the Germans finally reached the river banks, capturing 90% of the ruined city and splitting the remaining Soviet forces into two narrow pockets. Ice floes on the Volga now prevented boats and tugs from supplying the Soviet defenders. Nevertheless, the fighting continued, especially on the slopes of Mamayev Kurgan and inside the factory area in the northern part of the city. From 21 August to 20 November, the German 6th Army lost 60,548 men, including 12,782 killed, 45,545 wounded and 2,221 missing.
Recognising that German troops were ill-prepared for offensive operations during the winter of 1942, and that most of them were redeployed elsewhere on the southern sector of the Eastern Front, the Stavka decided to conduct a number of offensive operations between 19 November 1942 and 2 February 1943. These operations opened the Winter Campaign of 1942–1943 (19 November 1942 – 3 March 1943), which involved some fifteen Armies operating on several fronts. According to Zhukov, ""German operational blunders were aggravated by poor intelligence: they failed to spot preparations for the major counter-offensive near Stalingrad where there were 10 field, 1 tank and 4 air armies.""
During the siege, the German and allied Italian, Hungarian, and Romanian armies protecting Army Group B's north and south flanks had pressed their headquarters for support. The Hungarian 2nd Army was given the task of defending a section of the front north of Stalingrad between the Italian Army and Voronezh. This resulted in a very thin line, with some sectors where stretches were being defended by a single platoon. These forces were also lacking in effective anti-tank weapons. Zhukov states, "Compared with the Germans, the troops of the satellites were not so well armed, less experienced and less efficient, even in defence."
Because of the total focus on the city, the Axis forces had neglected for months to consolidate their positions along the natural defensive line of the Don River. The Soviet forces were allowed to retain bridgeheads on the right bank from which offensive operations could be quickly launched. These bridgeheads in retrospect presented a serious threat to Army Group B.
Similarly, on the southern flank of the Stalingrad sector the front southwest of Kotelnikovo was held only by the Romanian 4th Army. Beyond that army, a single German division, the 16th Motorised Infantry, covered 400 km. Paulus had requested permission to "withdraw the 6th Army behind the Don," but was rejected. According to Paulus' comments to Adam, "There is still the order whereby no commander of an army group or an army has the right to relinquish a village, even a trench, without Hitler's consent."
In autumn, the Soviet generals Georgy Zhukov and Aleksandr Vasilevsky, responsible for strategic planning in the Stalingrad area, concentrated forces in the steppes to the north and south of the city. The northern flank was defended by Hungarian and Romanian units, often in open positions on the steppes. The natural line of defence, the Don River, had never been properly established by the German side. The armies in the area were also poorly equipped in terms of anti-tank weapons. The plan was to punch through the overstretched and weakly defended German flanks and surround the German forces in the Stalingrad region.
During the preparations for the attack, Marshal Zhukov personally visited the front and noticing the poor organisation, insisted on a one-week delay in the start date of the planned attack. The operation was code-named "Uranus" and launched in conjunction with Operation Mars, which was directed at Army Group Center. The plan was similar to the one Zhukov had used to achieve victory at Khalkhin Gol three years before, where he had sprung a double envelopment and destroyed the 23rd Division of the Japanese army.
On 19 November 1942, the Red Army launched Operation Uranus. The attacking Soviet units under the command of Gen. Nikolay Vatutin consisted of three complete armies, the 1st Guards Army, 5th Tank Army and 21st Army, including a total of 18 infantry divisions, eight tank brigades, two motorised brigades, six cavalry divisions and one anti-tank brigade. The preparations for the attack could be heard by the Romanians, who continued to push for reinforcements, only to be refused again. Thinly spread, deployed in exposed positions, outnumbered and poorly equipped, the Romanian 3rd Army, which held the northern flank of the German 6th Army, was overrun.
Behind the front lines, no preparations had been made to defend key points in the rear such as Kalach. The response by the "Wehrmacht" was both chaotic and indecisive. Poor weather prevented effective air action against the Soviet offensive. Army Group B was in disarray and faced strong Soviet pressure across all its fronts. Hence it was ineffective in relieving the 6th army.
On 20 November, a second Soviet offensive (two armies) was launched to the south of Stalingrad against points held by the Romanian 4th Army Corps. The Romanian forces, made up primarily of infantry, were overrun by large numbers of tanks. The Soviet forces raced west and met on 23 November at the town of Kalach, sealing the ring around Stalingrad. The link-up of the Soviet forces, not filmed at the time, was later re-enacted for a propaganda film which was shown worldwide..
The surrounded Axis personnel comprised 265,000 Germans, Romanians, Italians, and the Croatians. In addition, the German 6th Army included between 40,000 and 65,000 "Hilfswillige" ("Hiwi"), or "volunteer auxiliaries", a term used for personnel recruited amongst Soviet POWs and civilians from areas under occupation. "Hiwi" often proved to be reliable Axis personnel in rear areas and were used for supporting roles, but also served in some front-line units as their numbers had increased. German personnel in the pocket numbered about 210,000, according to strength breakdowns of the 20 field divisions (average size 9,000) and 100 battalion sized units of the Sixth Army on 19 November 1942. Inside the pocket (, literally "cauldron"), there were also around 10,000 Soviet civilians and several thousand Soviet soldiers the Germans had taken captive during the battle. Not all of the 6th Army was trapped: 50,000 soldiers were brushed aside outside the pocket. These belonged mostly to the other two divisions of the 6th Army between the Italian and Romanian Armies: the 62nd and 298th Infantry Divisions. Of the 210,000 Germans, 10,000 remained to fight on, 105,000 surrendered, 35,000 left by air and the remaining 60,000 died.
Even with the desperate situation of the Sixth Army, Army Group A continued their invasion of the Caucasus further south from 19 November until 19 December. By 19 December the German army was in full retreat out of the Caucasus, while using the Sixth Army to tie down the Soviet forces. Hence Army Group A was never used to help relieve the Sixth Army.
Army Group Don was formed under Field Marshal von Manstein. Under his command were the twenty German and two Romanian divisions encircled at Stalingrad, Adam's battle groups formed along the Chir River and on the Don bridgehead, plus the remains of the Romanian 3rd Army.
The Red Army units immediately formed two defensive fronts: a circumvallation facing inward and a contravallation facing outward. Field Marshal Erich von Manstein advised Hitler not to order the 6th Army to break out, stating that he could break through the Soviet lines and relieve the besieged 6th Army. The American historians Williamson Murray and Alan Millet wrote that it was Manstein's message to Hitler on 24 November advising him that the 6th Army should not break out, along with Göring's statements that the Luftwaffe could supply Stalingrad that "... sealed the fate of the Sixth Army." After 1945, Manstein claimed that he told Hitler that the 6th Army must break out. The American historian Gerhard Weinberg wrote that Manstein distorted his record on the matter. Manstein was tasked to conduct a relief operation, named Operation Winter Storm ("Unternehmen Wintergewitter") against Stalingrad, which he thought was feasible if the 6th Army was temporarily supplied through the air.
Adolf Hitler had declared in a public speech (in the Berlin Sportpalast) on 30 September 1942 that the German army would never leave the city. At a meeting shortly after the Soviet encirclement, German army chiefs pushed for an immediate breakout to a new line on the west of the Don, but Hitler was at his Bavarian retreat of Obersalzberg in Berchtesgaden with the head of the "Luftwaffe", Hermann Göring. When asked by Hitler, Göring replied, after being convinced by Hans Jeschonnek, that the "Luftwaffe" could supply the 6th Army with an "air bridge." This would allow the Germans in the city to fight on temporarily while a relief force was assembled. A similar plan had been used a year earlier at the Demyansk Pocket, albeit on a much smaller scale: a corps at Demyansk rather than an entire army.
The director of "Luftflotte" 4, Wolfram von Richthofen, tried to get this decision overturned. The forces under the 6th Army were almost twice as large as a regular German army unit, plus there was also a corps of the 4th Panzer Army trapped in the pocket. The maximum 107 tonnes they could deliver a day – based on the number of available aircraft and with only the airfield at Pitomnik to land at – was far less than the minimum 750 tonnes needed. To supplement the limited number of Junkers Ju 52 transports, the Germans pressed other aircraft into the role, such as the Heinkel He 177 bomber (some bombers performed adequately – the Heinkel He 111 proved to be quite capable and was much faster than the Ju 52). General Richthofen informed Manstein on 27 November of the small transport capacity of the Luftwaffe and the impossibility of supplying 300 tons a day by air. Manstein now saw the enormous technical difficulties of a supply by air of these dimensions. The next day he made a six-page situation report to the general staff. Based on the information of the expert Richthofen, he declared that contrary to the example of the pocket of Demyansk the permanent supply by air would be impossible. If only a narrow link could be established to Sixth Army, he proposed that this should be used to pull it out from the encirclement, and said that the Luftwaffe should instead of supplies deliver only enough ammunition and fuel for a breakout attempt. He acknowledged the heavy moral sacrifice that giving up Stalingrad would mean, but this would be made easier to bear by conserving the combat power of the Sixth Army and regaining the initiative. He ignored the limited mobility of the army and the difficulties of disengaging the Soviets. Hitler reiterated that the Sixth Army would stay at Stalingrad and that the air bridge would supply it until the encirclement was broken by a new German offensive.
Supplying the 270,000 men trapped in the "cauldron" required 700 tons of supplies a day. That would mean 350 Ju 52 flights a day into Pitomnik. At a minimum, 500 tons were required. However, according to Adam, "On not one single day have the minimal essential number of tons of supplies been flown in." The "Luftwaffe" was able to deliver an average of 85 tonnes of supplies per day out of an air transport capacity of 106 tonnes per day. The most successful day, 19 December, the "Luftwaffe" delivered 262 tonnes of supplies in 154 flights. The outcome of the airlift was the Luftwaffe's failure to provide its transport units with the tools they needed to maintain an adequate count of operational aircraft – tools that included airfield facilities, supplies, manpower, and even aircraft suited to the prevailing conditions. These factors, taken together, prevented the Luftwaffe from effectively employing the full potential of its transport forces, ensuring that they were unable to deliver the quantity of supplies needed to sustain the 6th Army.
In the early parts of the operation, fuel was shipped at a higher priority than food and ammunition because of a belief that there would be a breakout from the city. Transport aircraft also evacuated technical specialists and sick or wounded personnel from the besieged enclave. Sources differ on the number flown out: at least 25,000 to at most 35,000.
Initially, supply flights came in from the field at Tatsinskaya, called 'Tazi' by the German pilots. On 23 December, the Soviet 24th Tank Corps, commanded by Major-General Vasily Mikhaylovich Badanov, reached nearby Skassirskaya and in the early morning of 24 December, the tanks reached Tatsinskaya. Without any soldiers to defend the airfield, it was abandoned under heavy fire; in a little under an hour, 108 Ju 52s and 16 Ju 86s took off for Novocherkassk – leaving 72 Ju 52s and many other aircraft burning on the ground. A new base was established some from Stalingrad at Salsk, the additional distance would become another obstacle to the resupply efforts. Salsk was abandoned in turn by mid-January for a rough facility at Zverevo, near Shakhty. The field at Zverevo was attacked repeatedly on 18 January and a further 50 Ju 52s were destroyed. Winter weather conditions, technical failures, heavy Soviet anti-aircraft fire and fighter interceptions eventually led to the loss of 488 German aircraft.
In spite of the failure of the German offensive to reach the 6th Army, the air supply operation continued under ever more difficult circumstances. The 6th Army slowly starved. General Zeitzler, moved by their plight, began to limit himself to their slim rations at meal times. After a few weeks on such a diet, he had "visibly lost weight", according to Albert Speer, and Hitler "commanded Zeitzler to resume at once taking sufficient nourishment."
The toll on the "Transportgruppen" was heavy. 160 aircraft were destroyed and 328 were heavily damaged (beyond repair). Some 266 Junkers Ju 52s were destroyed; one-third of the fleet's strength on the Eastern Front. The He 111 "gruppen" lost 165 aircraft in transport operations. Other losses included 42 Ju 86s, 9 Fw 200 Condors, 5 He 177 bombers and 1 Ju 290. The "Luftwaffe" also lost close to 1,000 highly experienced bomber crew personnel. So heavy were the "Luftwaffe"s losses that four of "Luftflotte" 4's transport units (KGrzbV 700, KGrzbV 900, I./KGrzbV 1 and II./KGzbV 1) were "formally dissolved."
Soviet forces consolidated their positions around Stalingrad, and fierce fighting to shrink the pocket began. Operation Winter Storm ("Operation Wintergewitter"), the German attempt led by Erich von Manstein to relieve the trapped army from the south, was initially successful. By 18 December, the German Army had pushed to within 48 km (30 mi) of Sixth Army's positions. However the predictable nature of the relief operation brought significant risk for all German forces in the area. The starving encircled forces at Stalingrad made no attempt to break out or link up with Manstein's advance. Some German officers requested that Paulus defy Hitler's orders to stand fast and instead attempt to break out of the Stalingrad pocket. Paulus refused, concerned about the Red Army attacks on the flank of Army Group Don and Army Group B in their advance on Rostov-on-Don, "an early abandonment" of Stalingrad "would result in the destruction of Army Group A in the Caucasus," and the fact that his 6th Army tanks only had fuel for a 30 km advance towards Hoth's spearhead, a futile effort if they did not receive assurance of resupply by air. Of his questions to Army Group Don, Paulus was told, "Wait, implement Operation 'Thunderclap' only on explicit orders!", Operation Thunderclap being the code word initiating the breakout.
On 16 December, the Soviets launched Operation Little Saturn, which attempted to punch through the Axis army (mainly Italians) on the Don and take Rostov-on-Don. The Germans set up a "mobile defence" of small units that were to hold towns until supporting armour arrived. From the Soviet bridgehead at Mamon, 15 divisions – supported by at least 100 tanks – attacked the Italian Cosseria and Ravenna Divisions, and although outnumbered 9 to 1, the Italians initially fought well, with the Germans praising the quality of the Italian defenders, but on 19 December, with the Italian lines disintegrating, ARMIR headquarters ordered the battered divisions to withdraw to new lines.
The fighting forced a total revaluation of the German situation. Sensing that this was the last chance for a breakout, Manstein pleaded with Hitler on 18 December, but Hitler refused. Paulus himself also doubted the feasibility of such a breakout. The attempt to break through to Stalingrad was abandoned and Army Group A was ordered to pull back from the Caucasus. The 6th Army now was beyond all hope of German relief. While a motorised breakout might have been possible in the first few weeks, the 6th Army now had insufficient fuel and the German soldiers would have faced great difficulty breaking through the Soviet lines on foot in harsh winter conditions. But in its defensive position on the Volga, the 6th Army continued to tie down a significant number of Soviet Armies.
On 23 December, the attempt to relieve Stalingrad was abandoned and Manstein's forces switched over to the defensive to deal with new Soviet offensives. As Zhukov states, "The military and political leadership of Nazi Germany sought not to relieve them, but to get them to fight on for as long possible so as to tie up the Soviet forces. The aim was to win as much time as possible to withdraw forces from the Caucasus (Army Group A) and to rush troops from other Fronts to form a new front that would be able in some measure to check our counter-offensive."
The Red Army High Command sent three envoys while simultaneously aircraft and loudspeakers announced terms of capitulation on 7 January 1943. The letter was signed by Colonel-General of Artillery Voronov and the commander-in-chief of the Don Front, Lieutenant-General Rokossovsky. A low-level Soviet envoy party (comprising Major Aleksandr Smyslov, Captain Nikolay Dyatlenko and a trumpeter) carried an offer to Paulus: if he surrendered within 24 hours, he would receive a guarantee of safety for all prisoners, medical care for the sick and wounded, prisoners being allowed to keep their personal belongings, "normal" food rations, and repatriation to any country they wished after the war; but Paulus – ordered not to surrender by Hitler – did not respond. The German High Command informed Paulus, "Every day that the army holds out longer helps the whole front and draws away the Russian divisions from it."
The Germans inside the pocket retreated from the suburbs of Stalingrad to the city itself. The loss of the two airfields, at Pitomnik on 16 January 1943 and Gumrak on the night of 21/22 January, meant an end to air supplies and to the evacuation of the wounded. The third and last serviceable runway was at the Stalingradskaya flight school, which reportedly had the last landings and takeoffs on 23 January. After 23 January, there were no more reported landings, just intermittent air drops of ammunition and food until the end.
The Germans were now not only starving, but running out of ammunition. Nevertheless, they continued to resist, in part because they believed the Soviets would execute any who surrendered. In particular, the so-called "HiWis", Soviet citizens fighting for the Germans, had no illusions about their fate if captured. The Soviets were initially surprised by the number of Germans they had trapped, and had to reinforce their encircling troops. Bloody urban warfare began again in Stalingrad, but this time it was the Germans who were pushed back to the banks of the Volga. The Germans adopted a simple defence of fixing wire nets over all windows to protect themselves from grenades. The Soviets responded by fixing fish hooks to the grenades so they stuck to the nets when thrown. The Germans had no usable tanks in the city, and those that still functioned could, at best, be used as makeshift pillboxes. The Soviets did not bother employing tanks in areas where the urban destruction restricted their mobility.
On 22 January, Paulus requested that he be granted permission to surrender. Hitler rejected it on a point of honour. He telegraphed the 6th Army later that day, claiming that it had made a historic contribution to the greatest struggle in German history and that it should stand fast "to the last soldier and the last bullet." Hitler told Goebbels that the plight of the 6th Army was a "heroic drama of German history." On 24 January, in his radio report to Hitler, Paulus reported "18,000 wounded without the slightest aid of bandages and medicines."
On 26 January 1943, the German forces inside Stalingrad were split into two pockets north and south of Mamayev-Kurgan. The northern pocket consisting of the VIIIth Corps, under General Walter Heitz, and the XIth Corps, was now cut off from telephone communication with Paulus in the southern pocket. Now "each part of the cauldron came personally under Hitler." On 28 January, the cauldron was split into three parts. The northern cauldron consisted of the XIth Corps, the central with the VIIIth and LIst Corps, and the southern with the XIVth Panzer Corps and IVth Corps "without units". The sick and wounded reached 40,000 to 50,000.
On 30 January 1943, the 10th anniversary of Hitler's coming to power, Goebbels read out a proclamation that included the sentence: "The heroic struggle of our soldiers on the Volga should be a warning for everybody to do the utmost for the struggle for Germany's freedom and the future of our people, and thus in a wider sense for the maintenance of our entire continent." Hitler promoted Paulus to the rank of "Generalfeldmarschall". No German field marshal had ever surrendered, and the implication was clear: if Paulus surrendered, he would shame himself and would become the highest ranking German officer ever to be captured. Hitler believed that Paulus would either fight to the last man or commit suicide.
On the next day, the southern pocket in Stalingrad collapsed. Soviet forces reached the entrance to the German headquarters in the ruined GUM department store. General Schmidt negotiated a surrender of the headquarters while Paulus was unaware in another room. When interrogated by the Soviets, Paulus claimed that he had not surrendered. He said that he had been taken by surprise. He denied that he was the commander of the remaining northern pocket in Stalingrad and refused to issue an order in his name for them to surrender.
There was no cameraman to film the capture of Paulus, but one of them (Roman Karmen) was able to record his first interrogation this same day, at Shumilov's 64th Army's HQ, and a few hours later at Rokossovsky's Don Front HQ.
The central pocket, under the command of Heitz, surrendered the same day, while the northern pocket, under the command of Karl Strecker, held out for two more days. When Strecker finally surrendered he and his Chief of Staff, Helmuth Groscurth, drafted the final signal sent from Stalingrad, purposely omitting the customary exclamation to Hitler, replacing it with "Long live Germany!"
Four Soviet armies were deployed against the remaining northern pocket. At four in the morning on 2 February, General Strecker was informed that one of his own officers had gone to the Soviets to negotiate surrender terms. Seeing no point in continuing, he sent a radio message saying that his command had done its duty and fought to the last man. He then surrendered. Around 91,000 exhausted, ill, wounded, and starving prisoners were taken, including 3,000 Romanians (the survivors of the 20th Infantry Division, 1st Cavalry Division and "Col. Voicu" Detachment). The prisoners included 22 generals. Hitler was furious and confided that Paulus ""could have freed himself from all sorrow and ascended into eternity and national immortality, but he prefers to go to Moscow.""
The calculation of casualties depends on what scope is given to the Battle of Stalingrad. The scope can vary from the fighting in the city and suburbs to the inclusion of almost all fighting on the southern wing of the Soviet–German front from the spring of 1942 to the end of the fighting in the city in the winter of 1943. Scholars have produced different estimates depending on their definition of the scope of the battle. The difference is comparing the city against the region. The Axis suffered 647,300 – 968,374 total casualties (killed, wounded or captured) among all branches of the German armed forces and its allies:
Totally, 235,000 German and allied troops from all units, including Manstein's ill-fated relief force, were captured during the battle.
The Germans lost 900 aircraft (including 274 transports and 165 bombers used as transports), 500 tanks and 6,000 artillery pieces. According to a contemporary Soviet report, 5,762 guns, 1,312 mortars, 12,701 heavy machine guns, 156,987 rifles, 80,438 sub-machine guns, 10,722 trucks, 744 aircraft; 1,666 tanks, 261 other armoured vehicles, 571 half-tracks and 10,679 motorcycles were captured by the Soviets. In addition, an unknown amount of Hungarian, Italian, and Romanian materiel was lost.
The situation of the Romanian tanks is known, however. Before "Operation Uranus", the 1st Romanian Armoured Division consisted of 121 R-2 light tanks and 19 German-produced tanks (Panzer III and IV). All of the 19 German tanks were lost, as well as 81 of the R-2 light tanks. Only 27 of the latter were lost in combat, however, the remaining 54 being abandoned after breaking down or running out of fuel. Ultimately, however, Romanian armoured warfare proved to be a tactical success, as the Romanians destroyed 127 Soviet tanks for the cost of their 100 lost units. Romanian forces destroyed 62 Soviet tanks on 20 November for the cost of 25 tanks of their own, followed by 65 more Soviet tanks on 22 November, for the cost of 10 tanks of their own. More Soviet tanks were destroyed as they overran the Romanian airfields. This was accomplished by Romanian Vickers/Reșița 75 mm anti-aircraft guns, which proved effective against Soviet armour. The battle for the German-Romanian airfield at Karpova lasted two days, with Romanian gunners destroying numerous Soviet tanks. Later, when the Tatsinskaya Airfield was also captured, the Romanian 75 mm guns destroyed five more Soviet tanks.
The USSR, according to archival figures, suffered 1,129,619 total casualties; 478,741 personnel killed or missing, and 650,878 wounded or sick. The USSR lost 4,341 tanks destroyed or damaged, 15,728 artillery pieces and 2,769 combat aircraft. 955 Soviet civilians died in Stalingrad and its suburbs from aerial bombing by "Luftflotte" 4 as the German 4th Panzer and 6th Armies approached the city.
The losses of transport planes were especially serious, as they destroyed the capacity for supply of the trapped 6th Army. The destruction of 72 aircraft when the airfield at Tatsinskaya was overrun meant the loss of about 10 percent of the Luftwaffe transport fleet.
These losses amounted to about 50 percent of the aircraft committed and the Luftwaffe training program was stopped and sorties in other theatres of war were significantly reduced to save fuel for use at Stalingrad.
The German public was not officially told of the impending disaster until the end of January 1943, though positive media reports had stopped in the weeks before the announcement. Stalingrad marked the first time that the Nazi government publicly acknowledged a failure in its war effort. On 31 January, regular programmes on German state radio were replaced by a broadcast of the sombre Adagio movement from Anton Bruckner's Seventh Symphony, followed by the announcement of the defeat at Stalingrad. On 18 February, Minister of Propaganda Joseph Goebbels gave the famous "Sportpalast" speech in Berlin, encouraging the Germans to accept a total war that would claim all resources and efforts from the entire population.
Based on Soviet records, over 10,000 German soldiers continued to resist in isolated groups within the city for the next month. Some have presumed that they were motivated by a belief that fighting on was better than a slow death in Soviet captivity. Brown University historian Omer Bartov claims they were motivated by National Socialism. He studied 11,237 letters sent by soldiers inside of Stalingrad between 20 December 1942 and 16 January 1943 to their families in Germany. Almost every letter expressed belief in Germany's ultimate victory and their willingness to fight and die at Stalingrad to achieve that victory. Bartov reported that a great many of the soldiers were well aware that they would not be able to escape from Stalingrad but in their letters to their families boasted that they were proud to "sacrifice themselves for the Führer".
The remaining forces continued to resist, hiding in cellars and sewers but by early March 1943, the last small and isolated pockets of resistance had surrendered. According to Soviet intelligence documents shown in the documentary, a remarkable NKVD report from March 1943 is available showing the tenacity of some of these German groups:
The operative report of the Don Front's staff issued on 5 February 1943, 22:00 said,
The condition of the troops that surrendered was pitiful. British war correspondent Alexander Werth described the following scene in his "Russia at War" book, based on a first-hand account of his visit to Stalingrad on 3–5 February 1943,
Out of the nearly 91,000 German prisoners captured in Stalingrad, only about 5,000 returned. Weakened by disease, starvation and lack of medical care during the encirclement, they were sent on foot marches to prisoner camps and later to labour camps all over the Soviet Union. Some 35,000 were eventually sent on transports, of which 17,000 did not survive. Most died of wounds, disease (particularly typhus), cold, overwork, mistreatment and malnutrition. Some were kept in the city to help rebuild it.
A handful of senior officers were taken to Moscow and used for propaganda purposes, and some of them joined the National Committee for a Free Germany. Some, including Paulus, signed anti-Hitler statements that were broadcast to German troops. Paulus testified for the prosecution during the Nuremberg Trials and assured families in Germany that those soldiers taken prisoner at Stalingrad were safe. He remained in the Soviet Union until 1952, then moved to Dresden in East Germany, where he spent the remainder of his days defending his actions at Stalingrad and was quoted as saying that Communism was the best hope for postwar Europe. General Walther von Seydlitz-Kurzbach offered to raise an anti-Hitler army from the Stalingrad survivors, but the Soviets did not accept. It was not until 1955 that the last of the 5,000–6,000 survivors were repatriated (to West Germany) after a plea to the Politburo by Konrad Adenauer.
Stalingrad has been described as the biggest defeat in the history of the German Army. It is often identified as the turning point on the Eastern Front, in the war against Germany overall, and in the entire Second World War. The Red Army had the initiative, and the Wehrmacht was in retreat. A year of German gains during Case Blue had been wiped out. Germany's Sixth Army had ceased to exist, and the forces of Germany's European allies, except Finland, had been shattered. In a speech on 9 November 1944, Hitler himself blamed Stalingrad for Germany's impending doom.
The destruction of an entire army (the largest killed, captured, wounded figures for Axis soldiers, nearly 1 million, during the war) and the frustration of Germany's grand strategy made the battle a watershed moment. At the time, the global significance of the battle was not in doubt. Writing in his diary on 1 January 1943, British General Alan Brooke, Chief of the Imperial General Staff, reflected on the change in the position from a year before:
At this point, the British had won the Battle of El Alamein in November 1942. However, there were only about 50,000 German soldiers at El Alamein in Egypt, while at Stalingrad 300,000 to 400,000 Germans had been lost.
Regardless of the strategic implications, there is little doubt about Stalingrad's symbolism. Germany's defeat shattered its reputation for invincibility and dealt a devastating blow to German morale. On 30 January 1943, the tenth anniversary of his coming to power, Hitler chose not to speak. Joseph Goebbels read the text of his speech for him on the radio. The speech contained an oblique reference to the battle, which suggested that Germany was now in a defensive war. The public mood was sullen, depressed, fearful, and war-weary. Germany was looking in the face of defeat.
The reverse was the case on the Soviet side. There was an overwhelming surge in confidence and belief in victory. A common saying was: "You cannot stop an army which has done Stalingrad." Stalin was feted as the hero of the hour and made a Marshal of the Soviet Union.
The news of the battle echoed round the world, with many people now believing that Hitler's defeat was inevitable. The Turkish Consul in Moscow predicted that "the lands which the Germans have destined for their living space will become their dying space". Britain's conservative "The Daily Telegraph" proclaimed that the victory had saved European civilisation. The country celebrated "Red Army Day" on 23 February 1943. A ceremonial Sword of Stalingrad was forged by King George VI. After being put on public display in Britain, this was presented to Stalin by Winston Churchill at the Tehran Conference later in 1943. Soviet propaganda spared no effort and wasted no time in capitalising on the triumph, impressing a global audience. The prestige of Stalin, the Soviet Union, and the worldwide Communist movement was immense, and their political position greatly enhanced.
In recognition of the determination of its defenders, Stalingrad was awarded the title Hero City in 1945. A colossal monument called The Motherland Calls was erected in 1967 on Mamayev Kurgan, the hill overlooking the city where bones and rusty metal splinters can still be found. The statue forms part of a war memorial complex which includes the ruins of the Grain Silo and Pavlov's House. On 2 February 2013 Volgograd hosted a military parade and other events to commemorate the 70th anniversary of the final victory. Since then, military parades have always commemorated the victory in the city.
The events of the Battle for Stalingrad have been covered in numerous media works of British, American, German, and Russian origin, for its significance as a turning point in the Second World War and for the loss of life associated with the battle. The term Stalingrad has become almost synonymous with large-scale urban battles with high casualties on both sides. | https://en.wikipedia.org/wiki?curid=4284 |
Biconditional introduction
In propositional logic, biconditional introduction is a valid rule of inference. It allows for one to infer a biconditional from two conditional statements. The rule makes it possible to introduce a biconditional statement into a logical proof. If formula_1 is true, and if formula_2 is true, then one may infer that formula_3 is true. For example, from the statements "if I'm breathing, then I'm alive" and "if I'm alive, then I'm breathing", it can be inferred that "I'm breathing if and only if I'm alive". Biconditional introduction is the converse of biconditional elimination. The rule can be stated formally as:
where the rule is that wherever instances of "formula_1" and "formula_2" appear on lines of a proof, "formula_3" can validly be placed on a subsequent line.
The "biconditional introduction" rule may be written in sequent notation:
where formula_9 is a metalogical symbol meaning that formula_3 is a syntactic consequence when formula_1 and formula_2 are both in a proof;
or as the statement of a truth-functional tautology or theorem of propositional logic:
where formula_14, and formula_15 are propositions expressed in some formal system. | https://en.wikipedia.org/wiki?curid=4286 |
Biconditional elimination
Biconditional elimination is the name of two valid rules of inference of propositional logic. It allows for one to infer a conditional from a biconditional. If formula_1 is true, then one may infer that formula_2 is true, and also that formula_3 is true. For example, if it's true that I'm breathing if and only if I'm alive, then it's true that if I'm breathing, I'm alive; likewise, it's true that if I'm alive, I'm breathing. The rules can be stated formally as:
and
where the rule is that wherever an instance of "formula_1" appears on a line of a proof, either "formula_2" or "formula_3" can be placed on a subsequent line;
The "biconditional elimination" rule may be written in sequent notation:
and
where formula_11 is a metalogical symbol meaning that formula_2, in the first case, and formula_3 in the other are syntactic consequences of formula_1 in some logical system;
or as the statement of a truth-functional tautology or theorem of propositional logic:
where formula_17, and formula_18 are propositions expressed in some formal system. | https://en.wikipedia.org/wiki?curid=4287 |
Base pair
A base pair (bp) is a unit consisting of two nucleobases bound to each other by hydrogen bonds. They form the building blocks of the DNA double helix and contribute to the folded structure of both DNA and RNA. Dictated by specific hydrogen bonding patterns, Watson–Crick base pairs (guanine–cytosine and adenine–thymine) allow the DNA helix to maintain a regular helical structure that is subtly dependent on its nucleotide sequence. The complementary nature of this based-paired structure provides a redundant copy of the genetic information encoded within each strand of DNA. The regular structure and data redundancy provided by the DNA double helix make DNA well suited to the storage of genetic information, while base-pairing between DNA and incoming nucleotides provides the mechanism through which DNA polymerase replicates DNA and RNA polymerase transcribes DNA into RNA. Many DNA-binding proteins can recognize specific base-pairing patterns that identify particular regulatory regions of genes.
Intramolecular base pairs can occur within single-stranded nucleic acids. This is particularly important in RNA molecules (e.g., transfer RNA), where Watson–Crick base pairs (guanine–cytosine and adenine–uracil) permit the formation of short double-stranded helices, and a wide variety of non-Watson–Crick interactions (e.g., G–U or A–A) allow RNAs to fold into a vast range of specific three-dimensional structures. In addition, base-pairing between transfer RNA (tRNA) and messenger RNA (mRNA) forms the basis for the molecular recognition events that result in the nucleotide sequence of mRNA becoming translated into the amino acid sequence of proteins via the genetic code.
The size of an individual gene or an organism's entire genome is often measured in base pairs because DNA is usually double-stranded. Hence, the number of total base pairs is equal to the number of nucleotides in one of the strands (with the exception of non-coding single-stranded regions of telomeres). The haploid human genome (23 chromosomes) is estimated to be about 3.2 billion bases long and to contain 20,000–25,000 distinct protein-coding genes. A kilobase (kb) is a unit of measurement in molecular biology equal to 1000 base pairs of DNA or RNA. The total amount of related DNA base pairs on Earth is estimated at 5.0 and weighs 50 billion tonnes. In comparison, the total mass of the biosphere has been estimated to be as much as 4 TtC (trillion tons of carbon).
Top, a G.C base pair with three hydrogen bonds. Bottom, an A.T base pair with two hydrogen bonds. Non-covalent hydrogen bonds between the bases are shown as dashed lines. The wiggly lines stand for the connection to the pentose sugar and point in the direction of the minor groove.
Hydrogen bonding is the chemical interaction that underlies the base-pairing rules described above. Appropriate geometrical correspondence of hydrogen bond donors and acceptors allows only the "right" pairs to form stably. DNA with high GC-content is more stable than DNA with low GC-content. But, contrary to popular belief, the hydrogen bonds do not stabilize the DNA significantly; stabilization is mainly due to stacking interactions.
The bigger nucleobases, adenine and guanine, are members of a class of double-ringed chemical structures called purines; the smaller nucleobases, cytosine and thymine (and uracil), are members of a class of single-ringed chemical structures called pyrimidines. Purines are complementary only with pyrimidines: pyrimidine-pyrimidine pairings are energetically unfavorable because the molecules are too far apart for hydrogen bonding to be established; purine-purine pairings are energetically unfavorable because the molecules are too close, leading to overlap repulsion. Purine-pyrimidine base-pairing of AT or GC or UA (in RNA) results in proper duplex structure. The only other purine-pyrimidine pairings would be AC and GT and UG (in RNA); these pairings are mismatches because the patterns of hydrogen donors and acceptors do not correspond. The GU pairing, with two hydrogen bonds, does occur fairly often in RNA (see wobble base pair).
Paired DNA and RNA molecules are comparatively stable at room temperature, but the two nucleotide strands will separate above a melting point that is determined by the length of the molecules, the extent of mispairing (if any), and the GC content. Higher GC content results in higher melting temperatures; it is, therefore, unsurprising that the genomes of extremophile organisms such as "Thermus thermophilus" are particularly GC-rich. On the converse, regions of a genome that need to separate frequently — for example, the promoter regions for often-transcribed genes — are comparatively GC-poor (for example, see TATA box). GC content and melting temperature must also be taken into account when designing primers for PCR reactions.
The following DNA sequences illustrate pair double-stranded patterns. By convention, the top strand is written from the 5' end to the 3' end; thus, the bottom strand is written 3' to 5'.
Chemical analogs of nucleotides can take the place of proper nucleotides and establish non-canonical base-pairing, leading to errors (mostly point mutations) in DNA replication and DNA transcription. This is due to their isosteric chemistry. One common mutagenic base analog is 5-bromouracil, which resembles thymine but can base-pair to guanine in its enol form.
Other chemicals, known as DNA intercalators, fit into the gap between adjacent bases on a single strand and induce frameshift mutations by "masquerading" as a base, causing the DNA replication machinery to skip or insert additional nucleotides at the intercalated site. Most intercalators are large polyaromatic compounds and are known or suspected carcinogens. Examples include ethidium bromide and acridine.
An unnatural base pair (UBP) is a designed subunit (or nucleobase) of DNA which is created in a laboratory and does not occur in nature. DNA sequences have been described which use newly created nucleobases to form a third base pair, in addition to the two base pairs found in nature, A-T (adenine – thymine) and G-C (guanine – cytosine). A few research groups have been searching for a third base pair for DNA, including teams led by Steven A. Benner, Philippe Marliere, Floyd Romesberg and Ichiro Hirao. Some new base pairs have been reported.
In 1989 Steven Benner (then working at the Swiss Federal Institute of Technology in Zurich) and his team led with modified forms of cytosine and guanine into DNA molecules "in vitro". The nucleotides, which encoded RNA and proteins, were successfully replicated "in vitro". Since then, Benner's team has been trying to engineer cells that can make foreign bases from scratch, obviating the need for a feedstock.
In 2002, Ichiro Hirao's group in Japan developed an unnatural base pair between 2-amino-8-(2-thienyl)purine (s) and pyridine-2-one (y) that functions in transcription and translation, for the site-specific incorporation of non-standard amino acids into proteins. In 2006, they created 7-(2-thienyl)imidazo[4,5-b]pyridine (Ds) and pyrrole-2-carbaldehyde (Pa) as a third base pair for replication and transcription. Afterward, Ds and 4-[3-(6-aminohexanamido)-1-propynyl]-2-nitropyrrole (Px) was discovered as a high fidelity pair in PCR amplification. In 2013, they applied the Ds-Px pair to DNA aptamer generation by "in vitro" selection (SELEX) and demonstrated the genetic alphabet expansion significantly augment DNA aptamer affinities to target proteins.
In 2012, a group of American scientists led by Floyd Romesberg, a chemical biologist at the Scripps Research Institute in San Diego, California, published that his team designed an unnatural base pair (UBP). The two new artificial nucleotides or "Unnatural Base Pair" (UBP) were named d5SICS and dNaM. More technically, these artificial nucleotides bearing hydrophobic nucleobases, feature two fused aromatic rings that form a (d5SICS–dNaM) complex or base pair in DNA. His team designed a variety of "in vitro" or "test tube" templates containing the unnatural base pair and they confirmed that it was efficiently replicated with high fidelity in virtually all sequence contexts using the modern standard "in vitro" techniques, namely PCR amplification of DNA and PCR-based applications. Their results show that for PCR and PCR-based applications, the d5SICS–dNaM unnatural base pair is functionally equivalent to a natural base pair, and when combined with the other two natural base pairs used by all organisms, A–T and G–C, they provide a fully functional and expanded six-letter "genetic alphabet".
In 2014 the same team from the Scripps Research Institute reported that they synthesized a stretch of circular DNA known as a plasmid containing natural T-A and C-G base pairs along with the best-performing UBP Romesberg's laboratory had designed and inserted it into cells of the common bacterium "E. coli" that successfully replicated the unnatural base pairs through multiple generations. The transfection did not hamper the growth of the "E. coli" cells and showed no sign of losing its unnatural base pairs to its natural DNA repair mechanisms. This is the first known example of a living organism passing along an expanded genetic code to subsequent generations. Romesberg said he and his colleagues created 300 variants to refine the design of nucleotides that would be stable enough and would be replicated as easily as the natural ones when the cells divide. This was in part achieved by the addition of a supportive algal gene that expresses a nucleotide triphosphate transporter which efficiently imports the triphosphates of both d5SICSTP and dNaMTP into "E. coli" bacteria. Then, the natural bacterial replication pathways use them to accurately replicate a plasmid containing d5SICS–dNaM. Other researchers were surprised that the bacteria replicated these human-made DNA subunits.
The successful incorporation of a third base pair is a significant breakthrough toward the goal of greatly expanding the number of amino acids which can be encoded by DNA, from the existing 20 amino acids to a theoretically possible 172, thereby expanding the potential for living organisms to produce novel proteins. The artificial strings of DNA do not encode for anything yet, but scientists speculate they could be designed to manufacture new proteins which could have industrial or pharmaceutical uses. Experts said the synthetic DNA incorporating the unnatural base pair raises the possibility of life forms based on a different DNA code.
The following abbreviations are commonly used to describe the length of a D/RNA molecule:
For single-stranded DNA/RNA, units of nucleotides are used—abbreviated nt (or knt, Mnt, Gnt)—as they are not paired.
To distinguish between units of computer storage and bases, kbp, Mbp, Gbp, etc. may be used for base pairs.
The centimorgan is also often used to imply distance along a chromosome, but the number of base pairs it corresponds to varies widely. In the Human genome, the centimorgan is about 1 million base pairs. | https://en.wikipedia.org/wiki?curid=4292 |
Baltimore Ravens
The Baltimore Ravens are a professional American football team based in Baltimore. The Ravens compete in the National Football League (NFL) as a member club of the American Football Conference (AFC) North division. The team plays its home games at M&T Bank Stadium and is headquartered in Owings Mills.
The club was established in 1996, after Art Modell, who was then the owner of the Cleveland Browns, announced plans to relocate the franchise from Cleveland to Baltimore in 1995. As part of a settlement between the league and the city of Cleveland, Modell was required to leave the Browns' history and records in Cleveland for a replacement team and replacement personnel that would resume play in 1999. In return, he was allowed to take his own personnel and team to Baltimore, where such personnel would then form an expansion team.
The Ravens have been one of the more successful franchises since their inception, having qualified for the NFL playoffs 12 times since 2000, with two Super Bowl championship titles (Super Bowl XXXV and Super Bowl XLVII), two AFC Championship titles (2000 and 2012), 15 playoff victories, four AFC Championship game appearances (2000, 2008, 2011 and 2012), six AFC North division titles (2003, 2006, 2011, 2012, 2018, and 2019), and are currently the only team in the NFL to hold a perfect record in multiple Super Bowl appearances. The Ravens organization was led by general manager Ozzie Newsome from 1996 until his retirement following the 2018 season, and has had three head coaches: Ted Marchibroda, Brian Billick, and since 2008, John Harbaugh. Starting with a record-breaking defensive performance in their 2000 season, the Ravens have established a reputation for strong defensive play throughout team history, featuring Hall of Fame players like middle linebacker Ray Lewis and safety Ed Reed as well as Pro Bowl players like defensive end Terrell Suggs and defensive tackle Haloti Ngata. The team is owned by Steve Bisciotti and valued at $2.59 billion, making the Ravens the 29th-most valuable sports franchise in the world.
The name "Ravens" was inspired by Edgar Allan Poe's poem "The Raven". Chosen in a fan contest that drew 33,288 voters, the allusion honors Poe, who spent the early part of his career in Baltimore and is buried there. As the "Baltimore Sun" reported at the time, fans also "liked the tie-in with the other birds in town, the Orioles, and found it easy to visualize a tough, menacing black bird.".
Before the football team, there was the Baltimore Ravens wheelchair basketball team — the original Baltimore Ravens. In 1972, the wheelchair basketball team was founded by Ralph Smith, long-term resident of Baltimore and second Vice President of the National Wheelchair Basketball Association. The name "Ravens" was inspired by Bob Ardinger, a member of the Ravens wheelchair basketball team. In the 1990s, the naming rights were later sold to the football team when they came to the city and the wheelchair basketball team became known as the Maryland Ravens, Inc.
After the controversial relocation of the Colts to Indianapolis, several attempts were made to bring an NFL team back to Baltimore. In 1993, ahead of the 1995 league expansion, the city was considered a favorite, behind only St. Louis, to be granted one of two new franchises. League officials and team owners feared litigation due to conflicts between rival bidding groups if St. Louis was awarded a franchise, and in October Charlotte, North Carolina was the first city chosen. Several weeks later, Baltimore's bid for a franchise—dubbed the Baltimore Bombers, in honor of the locally produced Martin B-26 Marauder bomber—had three ownership groups in place and a state financial package which included a proposed $200 million, rent-free stadium and permission to charge up to $80 million in personal seat license fees. Baltimore, however, was unexpectedly passed over in favor of Jacksonville, Florida, despite Jacksonville's minor TV market status and that the city had withdrawn from contention in the summer, only to return with then-Commissioner Paul Tagliabue's urging. Although league officials denied that any city had been favored, it was reported that Tagliabue and his longtime friend Washington Redskins owner Jack Kent Cooke had lobbied against Baltimore due to its proximity to Washington, D.C., and that Tagliabue had used the initial committee voting system to prevent the entire league ownership from voting on Baltimore's bid. This led to public outrage and the "Baltimore Sun" describing Tagliabue as having an "Anybody But Baltimore" policy. Maryland governor William Donald Schaefer said afterward that Tagliabue had led him on, praising Baltimore and the proposed owners while working behind-the-scenes to oppose Baltimore's bid.
By May 1994, Baltimore Orioles owner Peter Angelos had gathered a new group of investors, including author Tom Clancy, to bid on teams whose owners had expressed interest in relocating. Angelos found a potential partner in Georgia Frontiere, who was open to moving the Los Angeles Rams to Baltimore. Jack Kent Cooke opposed the move, intending to build the Redskins' new stadium in Laurel, Maryland, close enough to Baltimore to cool outside interest in bringing in a new franchise. This led to heated arguments between Cooke and Angelos, who accused Cooke of being a "carpetbagger." The league eventually persuaded Rams team president John Shaw to relocate to St. Louis instead, leading to a league-wide rumor that Tagliabue was again steering interest away from Baltimore, a claim which Tagliabue denied. In response to anger in Baltimore, including Governor Schaefer's threat to announce over the loudspeakers Tagliabue's exact location in Camden Yards any time he attended a Baltimore Orioles game, Tagliabue remarked of Baltimore's financial package: "Maybe (Baltimore) can open another museum with that money." Following this, Angelos made an unsuccessful $200 million bid to bring the Tampa Bay Buccaneers to Baltimore.
Having failed to obtain a franchise via the expansion, the city, despite having "misgivings," turned to the possibility of obtaining the Cleveland Browns, whose owner Art Modell was financially struggling and at odds with the city of Cleveland over needed improvements to the team's stadium.
Enticed by Baltimore's available funds for a first-class stadium and a promised yearly operating subsidy of $25 million, Modell announced on November 6, 1995 his intention to relocate the team from Cleveland to Baltimore the following year. The resulting controversy ended when representatives of Cleveland and the NFL reached a settlement on February 8, 1996. Tagliabue promised the city of Cleveland that an NFL team would be located in Cleveland, either through relocation or expansion, "no later than 1999". Additionally, the agreement stipulated that the Browns' name, colors, uniform design and franchise records would remain in Cleveland. The franchise history includes Browns club records and connections with Pro Football Hall of Fame players. Modell's Baltimore team, while retaining all current player contracts, would, for purposes of team history, appear as an expansion team, a new franchise. Not all players, staff or front office would make the move to Baltimore, however.
After relocation, Modell hired Ted Marchibroda as the head coach for his new team in Baltimore. Marchibroda was already well known because of his work as head coach of the Baltimore Colts during the 1970s and the Indianapolis Colts during the early 1990s. Ozzie Newsome, the Browns' tight end for many seasons, joined Modell in Baltimore as director of football operations. He was later promoted to vice-president/general manager.
The home stadium for the Ravens first two seasons was Baltimore's Memorial Stadium, previously home to the Baltimore Colts, the Baltimore Orioles, and the Canadian Football League’s Baltimore Stallions. The Ravens moved to their own new stadium, now known as M&T Bank Stadium, next to Camden Yards in 1998.
In the 1996 NFL Draft, the Ravens, with two picks in the first round, drafted offensive tackle Jonathan Ogden at No. 4 overall and linebacker Ray Lewis at No. 26 overall. Both Ogden and Lewis went on to play for the Ravens for their entire professional careers and were both inducted into the Pro Football Hall of Fame.
The 1996 Ravens won their opening game against the Oakland Raiders, but finished the season 4–12 despite receiver Michael Jackson leading the league with 14 touchdown catches. The 1997 Ravens started 3–1. Peter Boulware, a rookie defender from Florida State, recorded 11.5 sacks and was named AFC Defensive Rookie of the Year. The team finished 6–9–1. On October 26, the team made its first trip to Landover, Maryland to play their new regional rivals, the Washington Redskins. The Ravens won the game 20–17. On December 14, 1997, the Ravens played the final professional sporting event at Baltimore’s historic Memorial Stadium, winning 21–19 over the Tennessee Oilers.
1998 marked the opening of a new stadium for the Ravens, currently known as M&T Bank Stadium, but originally named “PSINet Stadium” after the now defunct internet service provider which purchased the original naming rights. Quarterback Vinny Testaverde left for the New York Jets before the season, and was replaced by former Indianapolis Colt Jim Harbaugh, and later Eric Zeier. Cornerback Rod Woodson joined the team after a successful stint with the Pittsburgh Steelers, and Priest Holmes started getting the first playing time of his career and ran for 1,000 yards. The Ravens finished 1998 with a 6–10 record. On November 29, the Ravens welcomed the Colts back to Baltimore for the first time in 15 years. Amidst a shower of negative cheers towards the Colts, the Ravens won 38–31.
Three consecutive losing seasons under Marchibroda led to a change in the head coach. Brian Billick took over as head coach in 1999. Billick had been offensive coordinator for the record-setting Minnesota Vikings the season before. Quarterback Tony Banks came to Baltimore from the St. Louis Rams and had the best season of his career with 17 touchdown passes and an 81.2 pass rating. He was joined by receiver Qadry Ismail, who posted a 1,000-yard season. The Ravens initially struggled with a record of 4–7 but managed to finish with an 8–8 record.
Due to continual financial hardships for the organization, the NFL took an unusual move and directed Modell to initiate the sale of his franchise. On March 27, 2000, NFL owners approved the sale of 49% of the Ravens to Steve Bisciotti. In the deal, Bisciotti had an option to purchase the remaining 51% for $325 million in 2004 from Art Modell. On April 9, 2004 the NFL approved Steve Bisciotti's purchase of the majority stake in the club.
Banks shared playing time in the 2000 regular season with Trent Dilfer. Both players put up decent numbers (and a 1,364-yard rushing season by rookie Jamal Lewis helped too) but the defense became the team's hallmark and bailed a struggling offense out in many instances through the season. Ray Lewis was named Defensive Player of the Year. Two of his defensive teammates, Sam Adams and Rod Woodson, made the Pro Bowl. Baltimore's season started strong with a 5–1 record. But the team struggled through mid-season, at one point going five games without scoring an offensive touchdown. The team regrouped and won each of their last seven games, finishing 12–4 and making the playoffs for the first time.
During the 2000 season, the Ravens dominating defense broke two notable NFL records. They held opposing teams to 165 total points, surpassing the 1985 Chicago Bears mark of 198 points for a 16-game season as well as surpassing the 1986 Chicago Bears mark of 187 points for a 16-game season, which at that time was the current NFL record these things along with outstanding play by the defense places the 2000 Ravens in the discussion as one of the Greatest if not the Greatest NFL defense of All Time along with the 1985 Chicago Bears, 2002 Tampa Bay Buccaneers and the 2015 Denver Broncos defenses.
Since the divisional rival Tennessee Titans had a record of 13–3, the Ravens had to play in the wild card round. They dominated the Denver Broncos 21–3 in their first game. In the divisional playoff, they went on the road to Tennessee. With the score tied 10–10 in the fourth quarter, an Al Del Greco field goal attempt was blocked and returned for a touchdown by Anthony Mitchell, and a Ray Lewis interception return for a score put the game squarely in Baltimore's favor. The 24–10 win put the Ravens in the AFC Championship against the Oakland Raiders. The game was rarely in doubt. Shannon Sharpe's 96-yard touchdown catch early in the second quarter followed by an injury to Raiders quarterback Rich Gannon were crucial as the Ravens won easily, 16–3.
Baltimore then went to Tampa for Super Bowl XXXV against the New York Giants. The Ravens’ defense carried them to a win. They recorded four sacks and forced five turnovers, one of which was a Kerry Collins interception returned for a touchdown by Duane Starks. The Giants' only score was a Ron Dixon kickoff return for a touchdown; however, the Ravens immediately countered with a touchdown return on the ensuing kickoff by Jermaine Lewis. The Ravens became champions with a 34–7 win.
In 2001, the Ravens attempted to defend their title with Elvis Grbac as their new starting quarterback, but a season-ending injury to Jamal Lewis on the first day of training camp and poor offensive performances stymied the team. After a 3–3 start, the Ravens defeated the Minnesota Vikings in the final week to clinch a wild card berth at 10–6. In the first round the Ravens showed flashes of their previous year with a 20–3 win over the Miami Dolphins, in which the team forced three turnovers and out-gained the Dolphins 347 yards to 151. In the divisional playoff the Ravens played the Pittsburgh Steelers. Three interceptions by Grbac ended the Ravens' season, as they lost 27–10.
Baltimore ran into salary cap problems entering the 2002 season and was forced to part with a number of impact players. In the NFL Draft, the team selected Ed Reed with the 24th overall pick. Reed would go on to become one of the best safeties in NFL history, making nine Pro Bowls until leaving the Ravens for the Houston Texans in 2013. Despite low expectations, the Ravens stayed somewhat competitive in 2002 until a losing streak in December eliminated any chances of a post-season berth and a 7–9 finish.
In 2003, the Ravens drafted their new quarterback, Kyle Boller, but he was injured midway through the season and was replaced by Anthony Wright. Jamal Lewis ran for 2,066 yards (including a then-NFL record 295 yards in one game against the Cleveland Browns on September 14). With a 10–6 record, Baltimore won their first AFC North division title. Their first playoff game, at home against the Tennessee Titans, went back and forth, with the Ravens being held to only 54 yards total rushing. The Titans won 20–17 on a late field goal, and Baltimore's season ended early.
Ray Lewis was also named Defensive Player of the year for the second time in his career.
In April 2003, Art Modell sold 49% of the team to Steve Bisciotti, a local businessman who had made his fortune in the temporary staffing field. After the season, Art Modell sold his remaining 51% ownership to Bisciotti, ending over 40 years of tenure as an NFL franchise owner.
The Ravens did not make the playoffs in 2004 and finished the season with a record of 9–7 with Boller spending the season at QB. They did get good play from veteran corner Deion Sanders and third year safety Ed Reed, who won the NFL Defensive Player of the Year award. They were also the only team to defeat the 15–1 Pittsburgh Steelers in the regular season.
The next offseason, the Ravens looked to augment their receiving corps (which was second-worst in the NFL in 2004) by signing Derrick Mason from the Titans and drafting Oklahoma wide receiver Mark Clayton in the first round of the 2005 NFL Draft. However, the Ravens ended their season 6–10.
The 2006 Baltimore Ravens season began with the team trying to improve on their 6–10 record of 2005. The Ravens, for the first time in franchise history, started 4–0, under the leadership of former Titans quarterback Steve McNair.
In 2006, The Ravens lost two straight games mid-season on offensive troubles, prompting coach Billick to drop their offensive coordinator Jim Fassel in their week seven bye. After the bye, and with Billick calling the offense, Baltimore would record a five-game win streak before losing to the Cincinnati Bengals in week 13. Still ranked second overall to first-place San Diego Chargers, the Ravens continued on. They defeated the Kansas City Chiefs, and held the defending Super Bowl champion Pittsburgh Steelers to only one touchdown at Heinz Field, allowing the Ravens to clinch the AFC North.
The Ravens ended the regular season with a franchise-best 13–3 record. Baltimore had secured the AFC North title, the No. 2 AFC playoff seed, and clinched a 1st-round bye by season's end. The Ravens were slated to face the Indianapolis Colts in the second round of the playoffs, in the first meeting of the two teams in the playoffs. Many Baltimore and Indianapolis fans saw this historic meeting as a sort of "Judgment Day" with the new team of Baltimore facing the old team of Baltimore (the former Baltimore Colts having left Baltimore under questionable circumstances in 1984). Both Indianapolis and Baltimore were held to scoring only field goals as the two defenses slugged it out all over M&T Bank Stadium. McNair threw two costly interceptions, including one at the 1-yard line. The eventual Super Bowl champion Colts won 15–6, ending Baltimore's season.
The Ravens hoped to improve upon their 13–3 record but injuries and poor play plagued the team. The Ravens finished the 2007 season in the AFC North cellar with a disappointing 5–11 record. A humiliating 22–16 overtime loss to the previously winless Miami Dolphins on December 16 ultimately led to Billick's dismissal after the end of the regular season. He was replaced by John Harbaugh, the special teams coach of the Philadelphia Eagles and the older brother of former Ravens quarterback Jim Harbaugh (1998).
With rookies at head coach (John Harbaugh) and quarterback (Joe Flacco), the Ravens entered the 2008 campaign with lots of uncertainty. Baltimore smartly recovered in 2008, winning eleven games and achieving a wild card spot in the postseason. On the strength of four interceptions, one resulting in an Ed Reed touchdown, the Ravens began its postseason run by winning a rematch over Miami 27–9 at Dolphin Stadium on January 4, 2009 in a wild-card game. Six days later, they advanced to the AFC Championship Game by avenging a Week 5 loss to the Titans 13–10 at LP Field on a Matt Stover field goal with 53 seconds left in regulation time. The Ravens fell one victory short of Super Bowl XLIII by losing to the Steelers 23–14 at Heinz Field on January 18, 2009.
In 2009, the Ravens won their first three matches, then lost the next three, including a close match in Minnesota. The rest of the season was an uneven string of wins and losses, which included a home victory over Pittsburgh in overtime followed by a Monday Night loss in Green Bay. That game was notable for the number of penalties committed, costing a total of 310 yards, and almost tying with the record set by Tampa Bay and Seattle in 1976. Afterwards, the Ravens easily crushed the Lions and Bears, giving up less than ten points in both games. The next match was against the Steelers, where Baltimore lost a close one before beating the Raiders to end the season. With a record of 9–7, the team finished second in the division and gained another wild card. Moving into the playoffs, they overwhelmed the Patriots; nevertheless they did not reach the AFC Championship because they were routed 20–3 by the Colts in the Divisional Round a week later.
Baltimore managed to beat the Jets 10–9 on the 2010 opener, but then lost a poorly-played game against Cincinnati the following week. The Ravens rebounded against the other two division teams, beating Cleveland 24–17 in Week 3 and then . The Ravens scored a fine win (31–17) at home against Denver in Week 5. The Ravens finished the season 12–4, second in the division due to a tiebreaker with Pittsburgh, and earning a wild card spot. Baltimore headed to Kansas City and defeated the Chiefs 30–7, but once again were knocked from the playoffs by Pittsburgh in a hard-fought game.
The Ravens hosted their arch-enemy in Week 1 of the 2011 season. On a hot, humid day in M&T Bank Stadium, crowd noise and multiple Steelers mistakes allowed Baltimore to crush them with three touchdowns 35–7. The frustrated Pittsburgh players also committed several costly penalties. Thus, the Ravens had gained their first ever victory over the Steelers with Ben Roethlisberger playing and avenged themselves of repeated regular and postseason losses in the series.
But in Week 2, the Ravens collapsed in Tennessee and lost 26–13. They rebounded by routing the Rams in Week 3 and then overpowering the Jets 34–17 in Week 4.
Week 5, the Ravens had a bye week, following a game against the Texans. But in Week 7, Baltimore had a stunning MNF upset loss in Jacksonville as they were held to one touchdown in a 12–7 loss. Their final scoring drive failed as Joe Flacco threw an interception in the closing seconds of the game.
After beating the Cincinnati Bengals in Week 17 of the regular season, the Ravens advanced to the playoffs as the Number 2 seed in the AFC with a record of 12–4. They gained the distinction of AFC North Champions over Pittsburgh (12-4) due to a tie breaker.
Ravens' Lee Evans was stripped of a 14-yard touchdown pass by the Patriots Sterling Moore with 22 seconds left and Ravens kicker Billy Cundiff pushed a 32-yard field goal attempt wide left on fourth down as the Patriots held on to beat the Ravens 23-20 during the AFC championship game and advance to Super Bowl XLVI.
The Ravens' attempt to convert Joe Flacco into a pocket passer remained a work in progress as the 2012 season began. Terrell Suggs suffered a tendon injury during an off-season basketball game and was unable to play for at least several weeks. In the opener on September 10, Baltimore routed Cincinnati 44–13. After this easy win, the team headed to Philadelphia, but lost 24–23.
Returning home for a primetime rematch of the AFC Championship, another bizarre game ensued. New England picked apart the Baltimore defense (which was considerably weakened without Terrell Suggs and some other players lost over the off-season) for the first half. Trouble began early in the game when a streaker ran out onto the field and had to be tackled by security, and accelerated when, at 2:18 in the 4th quarter, the referees made a holding call on RG Marshal Yanda. Enraged fans repeatedly chanted an obscenity at this penalty. The Ravens finally drove downfield and on the last play of the game, Justin Tucker kicked a 27-yard field goal to win the game 31–30, capping off a second intense and controversially-officiated game in a row for the Ravens.
The Ravens would win the AFC North with a 10–6 record, but finished 4th in the AFC playoff seeding, and thus had to play a wild-card game. After defeating the Indianapolis Colts 24–9 at home (the final home game of Ray Lewis), the Ravens traveled to Denver to play against the top seeded Broncos. In a very back-and-forth contest, the Ravens pulled out a 38-35 victory in two overtimes. They then won their 2nd AFC championship by coming back from a 13-7 halftime deficit to defeat the Patriots once again, 28–13.
The Ravens played the Super Bowl XLVII against the San Francisco 49ers. Baltimore built a 28–6 lead early in the third quarter before a partial power outage in the Superdome suspended play for 34 minutes (earning the game the added nickname of the Blackout Bowl). After play resumed, San Francisco scored 17 unanswered third-quarter points to cut the Ravens' lead, 28–23, and continued to chip away in the fourth quarter. With the Ravens leading late in the game, 34–29, the 49ers advanced to the Baltimore 7-yard line just before the two-minute warning but turned the ball over on downs. The Ravens then took an intentional safety in the waning moments of the game to preserve the victory. Baltimore quarterback Joe Flacco, who completed 22 of 33 passes for 287 yards and three touchdowns, was named Super Bowl MVP.
Coming off as the defending Super Bowl champions, this was the first year in franchise history for the team without Ray Lewis. The Ravens started out 3–2, and started the 2-0 Houston Texans 14-loss streak by shutting them 30–9 in Week 3. However, the Ravens lost their next 3 games, losing to the Green Bay Packers and Pittsburgh Steelers in last-minute field goals and were shut out in an attempt to tie the game against the Cleveland Browns 24–18.
After winning and losing their next game, the Ravens came out 4–6, but managed winning their next four games in dominating the Jets 19–3, a Steelers win 22-20 during Thanksgiving, a booming ending in Baltimore against the Vikings 29–26, and an 18–16 win at Detroit, including Justin Tucker's 61-yard game-winning field goal. The Ravens were 8–6, with the 6th seed, but after losing their next two games, and the San Diego Chargers winning their next two to clinch the 6th seed, the Ravens finished 8-8 and missed the playoffs for the first time since 2007.
On January 27, 2014, the Ravens hired former Houston Texans head coach Gary Kubiak to be their new offensive coordinator after Jim Caldwell accepted the new available head coaching job with the Detroit Lions. On February 15, 2014, star running back Ray Rice and his fiancée Janay Palmer were arrested and charged with assault after a physical altercation at Revel Casino in Atlantic City, New Jersey. Celebrity news website TMZ posted a video of Rice dragging Palmer's body out of an elevator after apparently knocking her out. For the incident, Rice was initially suspended for the first two games of the 2014 NFL season on July 25, 2014, which led to widespread criticism of the NFL.
In Week 1, on September 7, the Baltimore Ravens lost to the Cincinnati Bengals, 23–16. The next day, on September 8, 2014, TMZ released additional footage from an elevator camera showing Rice punching Palmer. The Baltimore Ravens terminated Rice's contract as a result, and was later indefinitely suspended by the NFL, although a judge later vacated this indefinite suspension. In Week 12, the Ravens traveled down for an interconference battle with the New Orleans Saints, which the Ravens won. In Week 16, the Ravens traveled to Houston to take on the Texans. In one of Flacco's worst performances, the offense sputtered against the Houston defense and Flacco threw three interceptions, falling to the Texans 25–13. With their playoff chances and season hanging in the balance, the Ravens took on the Browns in Week 17 at home. After three quarters had gone by and down 10–3, Joe Flacco led the Ravens on a comeback scoring 17 unanswered points, winning 20–10. With the win, and the Kansas City Chiefs defeating the San Diego Chargers, the Ravens clinched their sixth playoff berth in seven seasons.
In the wild card round, the Ravens won 30–17 against their divisional rivals, the Pittsburgh Steelers, at Heinz Field. In the next game in the Divisional round, the Ravens faced the New England Patriots. Despite a strong offensive effort and having a 14-point lead twice in the game, the Ravens were defeated by the Patriots 35–31, ending their season.
The 2015 season marked 20 seasons of the franchise's existence competing in the NFL, which the franchise recognized with a special badge being worn on their uniforms during the 2015 NFL season.
The Ravens lost key players such as Joe Flacco, Justin Forsett, Terrell Suggs, Steve Smith Sr., and Eugene Monroe to season-ending injuries. Injuries and their inability to win close games early in the season led to the first losing season in the Harbaugh-Flacco era. The 2016 Ravens finished 8–8, but failed to qualify the playoffs for the second straight year. They were eliminated from playoff contention after their Week 16 loss to their division rivals, the Steelers. This was the first time the Ravens missed the playoffs in consecutive seasons since 2004–2005, as well as the first in the Harbaugh/Flacco era.
During the 2017 season, the Ravens improved upon their 8–8 record from 2016 by one win, finishing the season 9-7 and missing the playoffs for the third year in a row. This marked the first time the Ravens failed to make the playoffs in three straight seasons since the team's first three years of existence (1996-1998). The Ravens suffered a heartbreaking loss at home to the Cincinnati Bengals in the final game of the season that prevented them from earning a playoff berth.
The Ravens drafted QB Lamar Jackson with the 32nd pick in the 2018 draft. After the team started the season with a 4–5 record, Jackson took over as the starting QB in Week 11 when Joe Flacco was sidelined with a hip injury. The team won six of its next seven games, finishing the 2018 season with a 10–6 record and winning the AFC North, giving them their first playoff appearance since 2014 and their first division title since 2012. The Ravens lost to the Los Angeles Chargers in the Wild Card round with Jackson at quarterback, making him the youngest QB in NFL history to start a playoff game. At the conclusion of the season, Ozzie Newsome stepped down as the team's general manager. He was replaced by longtime assistant Eric DeCosta.
On March 13, 2019, the Ravens traded Joe Flacco to the Denver Broncos in exchange for a fourth-round pick in the 2019 NFL Draft. That season, Lamar Jackson led the Ravens to a franchise-best 14–2 record, including a 12-game winning streak to finish the regular season. On December 22, they clinched home field advantage for the first time in franchise history following a win over the Cleveland Browns. On December 8, Jackson became only the second player in NFL history to rush for over 1,000 yards from the quarterback position. Four days later, Jackson broke Michael Vick's single-season quarterback rushing record of 1,037 yards. Thirteen Ravens were selected to the 2019 Pro Bowl, matching the all-time NFL record.
The Ravens finished the 2019 regular season with 3,296 rushing yards, the most rushing yards by any team in NFL history during a season and they became the first team in NFL history to average at least 200 passing yards and 200 rushing yards per game in the same season.
Despite earning the number-one seed in the playoffs, the Ravens were eliminated by the sixth-seeded Tennessee Titans in the Divisional Round of the playoffs, 28–12. Lamar Jackson was unanimously voted AP NFL MVP, becoming only the second player in NFL history to do so, after Tom Brady in 2010.
By far the team's biggest rival is the Pittsburgh Steelers. Pittsburgh and Baltimore are separated by a less-than-5-hour drive along Interstate 70. Both teams are known for their hard-hitting physical style of play. They play twice a year in the AFC North, and have met four times in the playoffs. Games between these two teams usually come down to the wire as most within the last 5 years have come down to under 4 points.
The rivalry is considered one of the most significant and intense in the NFL today.
The Ravens also have divisional rivalries with the Cleveland Browns and Cincinnati Bengals. The rivalry with the Browns has been very one-sided; Baltimore holds an advantage of 31–11 against Cleveland. The rivalry with Cincinnati has been closer, with the Ravens slightly holding the edge in the all-time series 25–23 as of the end of the 2019 NFL season.
The Ravens first met the New England Patriots in 1996, but the rivalry truly started in 2007 when the Ravens suffered a bitter 27–24 loss in the Patriots' quest for perfection. The rivalry began to escalate in 2009 when the Patriots beat the Ravens 27–21 in a game that involved a confrontation between Patriots quarterback Tom Brady and Ravens linebacker Terrell Suggs. Both players would go on to take verbal shots at each other through the media after the game.
While the Patriots lead the overall series, 10–4, the teams have split four postseason meetings, 2–2. The Ravens won the , 33–14, and the , 28–13. The Patriots won the 23–20 and the , 35–31.
The team's first helmet logo, used from 1996 through the 1999 Pro Bowl, featured raven wings outspread from a shield displaying a letter "B" framed by the word "Ravens" overhead and a cross bottony underneath. The US Fourth Circuit Court of Appeals affirmed a jury verdict that the logo infringed on a copyright retained by Frederick E. Bouchat, an amateur artist and security guard in Maryland, but that he was entitled to only three dollars in damages from the NFL.
Bouchat had submitted his design to the Maryland Stadium Authority by fax after learning that Baltimore was to acquire an NFL team. He was not credited for the design when the logo was announced. Bouchat sued the team, claiming to be the designer of the emblem; representatives of the team asserted that the image had been designed independently. The court ruled in favor of Bouchat, noting that team owner Modell had access to Bouchat's work. Bouchat's fax had gone to John Moag, the Maryland Stadium Authority chairman, whose office was located in the same building as Modell's. Bouchat ultimately was not awarded monetary compensation in the damages phase of the case.
The "Baltimore Sun" ran a poll showing three designs for new helmet logos. Fans participating in the poll expressed a preference for a raven's head in profile over other designs. Art Modell announced that he would honor this preference but still wanted a letter "B" to appear somewhere in the design. The new Ravens logo, introduced in 1999, featured a raven's head in profile with the letter superimposed. The secondary logo is a shield that honors Baltimore's history of heraldry. Alternating Calvert and Crossland emblems (seen also in the flag of Maryland and the flag of Baltimore) are interlocked with stylized letters "B" and "R".
The design of the Ravens uniform has remained essentially unchanged since the team's inaugural season in 1996. Art Modell admitted to ESPN's Roy Firestone that the Ravens' colors, introduced in early 1996, were inspired by the Northwestern Wildcats 1995 dream season. Helmets are black with purple "talon" stripes rising from the facemask to the crown. Players normally wear purple jerseys at home and white jerseys on the road. In 1996 the team wore black pants with a single large white stripe for all games.
In 1997 the Ravens opted for a more classic NFL look with white pants sporting stripes in purple and black, along with the jerseys sporting a different font for the uniform numbers. The white pants were worn with both home and road jerseys. The road uniform (white pants with white jerseys) was worn by the Ravens in Super Bowl XXXV, at the end of the 2000 NFL season.
In the 2002 season the Ravens began the practice of wearing white jerseys for the home opener that has a 1:00 kickoff. In recent seasons, the practice has come when the home game is played in week one. Since John Harbaugh became the head coach in 2008, the Ravens have also worn their white jerseys at home for preseason games.
In November 2004 the team introduced an alternate uniform design featuring black jerseys and solid black pants with black socks. The all-black uniform was first worn for a home game against the Cleveland Browns, entitled "Pitch Black" night, that resulted in a Ravens win. The uniform has since been worn for select prime-time national game broadcasts and other games of significance.
The Ravens began wearing black pants again with the white jersey in 2008. On December 7, 2008, during a Sunday Night Football game against the Washington Redskins, the Ravens introduced a new combination of black jersey with white pants. It was believed to be due to the fact that John Harbaugh doesn't like the "blackout" look. However, on December 19, 2010, the Ravens wore their black jerseys and black pants in a 30–24 victory over the New Orleans Saints.
Since 2010, the Ravens have worn their black jerseys at least twice each season. From 2011 to 2013 and again in 2015, they wore the all blacks once and the black on white once. In 2014 and 2016, they wore all black both times they wore alternate uniforms. In 2017, they wore all black twice and black on white once (although the league is supposed to limit teams to wearing alternate jerseys a maximum of two times a season).
On December 5, 2010, the Ravens reverted to the black pants with the purple jerseys versus the Pittsburgh Steelers during NBC's "Sunday Night Football" telecast. The Ravens lost to the Steelers 13–10. They wore the same look again for their game against the Cleveland Browns on December 24, 2011, and they won, 20–14. They wore this combination a third time against the Houston Texans on January 15, 2012 in the AFC Divisional playoff. They won 20–13. They would again wear this combination on January 6, 2013, during the AFC Wild Card playoff and what turned out to be Ray Lewis' final home game, where they defeated the Indianapolis Colts 24–9.
From their inaugural season until 2006, the Ravens wore white cleats with their uniforms; they switched to black cleats in 2007.
On December 20, 2015, the team unexpectedly debuted gold pants for the first time, wearing them with their regular purple jerseys against the Kansas City Chiefs. Although gold is an official accent color of the Ravens, the pants got an overwhelmingly negative response on social media by both Ravens fans and fans of other NFL teams, with some comparisons being made to the rival Pittsburgh Steelers' pants, and mustard.
During the 2015 season, the NFL announced a jersey promotion called Color Rush in which teams would wear uniforms typically of one color head-to-toe during select prime-time games. The promotion was used three times that season; all the games that featured them were on Thursday Night and had both teams wear them in each. The following season, the league released uniforms for all 32 teams and announced they would be worn during all Thursday Night games that year, as well as on Christmas. The Ravens had one Thursday Night game in 2016; they wore their all-purple Color Rush uniforms and won 28–7 over the division rival Cleveland Browns. They had one other Thursday Night game the following season, in which they again wore the jerseys and won 40–0 over the Miami Dolphins. In their Christmas 2016 game against the Steelers, the Ravens wore their regular all-white uniforms while their rivals wore their Color Rush uniforms.
On September 13, 2018, the Ravens debuted a new combination in a road game against the Cincinnati Bengals, wearing white jerseys with purple pants. The purple pants are similar to the ones used for Color Rush except that it has side stripes of black and white; the Color Rush purple pants have gold and white stripes. Then on October 21 against the New Orleans Saints, the Ravens paired their new purple pants with their regular purple uniforms. For the regular season finale against the Browns on December 30, the Ravens wore their black uniforms with purple pants.
The team marching band is called Baltimore's Marching Ravens. They began as the Colts' marching band and have operated continuously from September 7, 1947 to the present. They helped campaign for football to return to Baltimore after the Colts moved. Because they stayed in Baltimore after the Colts left, the band is nicknamed "the band that would not die" and were the subject of an episode of ESPN's "30 for 30". The Washington Redskins are the only other NFL team that currently has a marching band.
Note: The following lists players who officially played for the Ravens. For other Hall of Famers, players whose numbers were retired, and players who played for the Baltimore Colts, see Indianapolis Colts. Bold number notes player inducted as a member of the Ravens. For Cleveland Browns players, including those in the Hall of Fame and those whose numbers were retired, see Cleveland Browns.
The Ravens do not officially have retired numbers. However, the number 19 has not been issued out of respect for Baltimore Colts quarterback Johnny Unitas, except for quarterback Scott Mitchell in his lone season in Baltimore in 1999. In addition, numbers 75, 52, and 20, in honor of Jonathan Ogden, Ray Lewis, and Ed Reed respectively, have not been issued since those players' retirements from football.
The Ravens have a "Ring of Honor" which is on permanent display encircling the field of M&T Bank Stadium. The ring currently honors 19 members, including eight former members of the Baltimore Colts.
Key/Legend
The Baltimore Ravens had their first draft in 1996, where they selected offensive lineman from UCLA and current NFL Hall of Famer, and 11-time Pro-Bowler Jonathan Ogden. Along with their pick in the next year's draft, this was the highest first-round draft pick that the Ravens have had. They also selected Ray Lewis with the 26th pick. In both 1996 and 2000, the Ravens had two first-round draft picks. However, in 2004 they had none. In their history, the Ravens have drafted 4 offensive linemen, 3 linebackers, 2 wide receivers, 2 cornerbacks, 2 quarterbacks, a running back, tight end, safety, and defensive tackle. The Ravens have 56 combined Pro-Bowl appearances from their first-round draft picks.
+ = min. 500 attempts, # = min. 100 attempts, ∗ = minimum 15 attempts,
∗ = minimum 15 attempts, # = min. 100 attempts, + = min. 500 attempts
∗ = minimum 4 receptions, # = min. 20 receptions, + = min. 200 receptions
"All records as of December 18, 2019 per Pro-Football Reference.com"
The Ravens' flagship radio stations are Hearst-owned WIYY (98 Rock) and WBAL 1090 AM, with Gerry Sandusky (WBAL-TV Sports Anchor since 1988) as the play-by-play announcer and analysts Dennis Pitta (Baltimore Ravens TE 2010–2016) and Jarret Johnson (Baltimore Ravens LB 2003–2011).
The team's flagship station is WIYY/WBAL sister station WBAL-TV, which broadcasts pre-season games and team programming throughout the season. The programming is syndicated to WJLA-TV in Washington, WGAL in the Harrisburg–Lebanon–York–Lancaster, Pennsylvania market, and until 2017, was carried through the remainder of the team's region by CSN Mid-Atlantic. In January 2017, the Ravens announced that it had cut ties with CSN Mid-Atlantic, as the network was cutting back on its day-to-day coverage of other teams in the region in order to focus more extensively on the Washington Capitals and Wizards—whose games are broadcast by CSN Mid-Atlantic, and whose owner holds a stake in the network. The team announced that it would seek a new partner; until 2010, these rights were held by MASN.
Ravens regular season games are typically broadcast by WJZ-TV as part of CBS's rights to the AFC, but games may occasionally be broadcast on WBAL ("Sunday Night Football" and simulcasts of games on cable) or WBFF-TV (also "Thursday Night Football"). | https://en.wikipedia.org/wiki?curid=4293 |
British National Party
The British National Party (BNP) is a far-right, fascist political party in the United Kingdom. It is headquartered in Wigton, Cumbria, and its current leader is Adam Walker. A minor party, it has no elected representatives at any level of UK government. Founded in 1982, the party reached its greatest level of success in the 2000s, when it had over fifty seats in local government, one seat on the London Assembly, and two Members of the European Parliament.
Taking its name from that of a defunct 1960s far-right party, the BNP was created by John Tyndall and other former members of the fascist National Front (NF). During the 1980s and 1990s, the BNP placed little emphasis on contesting elections, in which it did poorly. Instead, it focused on street marches and rallies, creating the Combat 18 paramilitary—its name a coded reference to Nazi German leader Adolf Hitler—to protect its events from anti-fascist protesters. A growing 'moderniser' faction was frustrated by Tyndall's leadership, and ousted him in 1999. The new leader Nick Griffin sought to broaden the BNP's electoral base by presenting a more moderate image, targeting concerns about rising immigration rates, and emphasising localised community campaigns. This resulted in increased electoral growth throughout the 2000s, to the extent that it became the most electorally successful far-right party in British history. Concerns regarding financial mismanagement resulted in Griffin being removed from office in 2014. By this point the BNP's membership and vote share had declined dramatically, groups like Britain First and National Action had splintered off, and the English Defence League had supplanted it as the UK's foremost far-right group.
Ideologically positioned on the extreme-right or far-right of British politics, the BNP has been characterised as fascist or neo-fascist by political scientists. Under Tyndall's leadership, it was more specifically regarded as neo-Nazi. The party is ethnic nationalist, and it espouses the view that only white people should be citizens of the United Kingdom. It calls for an end to non-white migration into the UK and for non-white Britons to be stripped of citizenship and removed from the country. Initially, it called for the compulsory expulsion of non-whites, although since 1999 has advocated voluntary removals with financial incentives. It promotes biological racism and the white genocide conspiracy theory, calling for global racial separatism and condemning interracial relationships. Under Tyndall, the BNP emphasised anti-semitism and Holocaust denial, promoting the conspiracy theory that Jews seek to dominate the world through both communism and international capitalism. Under Griffin, the party's focus switched from anti-semitism towards Islamophobia. It promotes economic protectionism, Euroscepticism, and a transformation away from liberal democracy, while its social policies oppose feminism, LGBT rights, and societal permissiveness.
Operating around a highly centralised structure that gave its chair near total control, the BNP built links with far-right parties across Europe and created various sub-groups, including a record label and trade union. The BNP attracted most support from within White British working-class communities in northern and eastern England, particularly among middle-aged and elderly men. Polls suggested that most Britons favoured a ban on the party. It faced much opposition from anti-fascists, religious organisations, the mainstream media, and most politicians, and BNP members were banned from various professions.
The British National Party (BNP) was founded by the extreme-right political activist John Tyndall. Tyndall had been involved in neo-Nazi groups since the late 1950s before leading the far-right National Front (NF) throughout most of the 1970s. Following an argument with senior party member Martin Webster, he resigned from the NF in 1980. In June 1980 Tyndall established a rival, the New National Front (NNF). At the recommendation of Ray Hill—who was secretly an anti-fascist spy seeking to sow disharmony among Britain's far-right—Tyndall decided to unite an array of extreme-right groups as a single party. To this end, Tyndall established a Committee for Nationalist Unity (CNU) in January 1982. In March 1982, the CNU held a conference at the Charing Cross Hotel in London, at which 50 far-right activists agreed to the formation of the BNP.
The BNP was formally launched on 7 April 1982 at a press conference in Victoria. Led by Tyndall, most of its early members came from the NNF, although others were defectors from the NF, British Movement, British Democratic Party, and Nationalist Party. Tyndall remarked that there was "scarcely any difference [between the BNP and NF] in ideology or policy save in the minutest detail", and most of the BNP's leading activists had formerly been senior NF figures. Under Tyndall's leadership the party was neo-Nazi in orientation and engaged in nostalgia for Nazi Germany. It adopted the NF's tactic of holding street marches and rallies, believing that these boosted morale and attracted new recruits. Their first march took place in London on St. George's Day 1982. These marches often involved clashes with anti-fascist protesters and resulted in multiple arrests, helping to cement the BNP's association with political violence and older fascist groups in the public eye. As a result, BNP organisers began to favour indoor rallies, although street marches continued to be held throughout the mid-to-late 1980s.
In its early years, the BNP's involvement in elections was "irregular and intermittent", and for its first two decades it faced consistent electoral failure. It suffered from low finances and few personnel, and its leadership was aware that its electoral viability was weakened by the anti-immigration rhetoric of Conservative Party Prime Minister Margaret Thatcher. In the 1983 general election the BNP stood 54 candidates, although it only campaigned in five seats. Although it was able to air its first party political broadcast, it averaged a vote share of 0.06% in the seats it contested.
After the Representation of the People Act 1985 raised the electoral deposit to £500, the BNP adopted a policy of "very limited involvement" in elections. It abstained in the 1987 general election, and stood only 13 candidates in the 1992 general election. In a 1993 local by-election the BNP gained one council seat—won by Derek Beackon in the East London district of Millwall—after a campaign that played to local whites who were angry at the perceived preferential treatment received by Bangladeshi migrants in social housing. Following an anti-BNP campaign launched by local religious groups and the Anti-Nazi League, it lost this seat during the 1994 local elections. In the 1997 general election, it contested 55 seats and gained an average 1.4% of the vote.
In the early 1990s, the paramilitary group Combat 18 (C18) was formed to protect BNP events from anti-fascists. In 1992, C18 carried out attacks on left-wing targets like an anarchist bookshop and the headquarters of the "Morning Star". Tyndall was angered by C18's growing influence on the BNP's street activities, and by August 1993, C18 activists were physically clashing with other BNP members. In December 1993, Tyndall issued a bulletin to BNP branches declaring C18 to be a proscribed organisation, furthermore suggesting that it may have been established by agents of the state to discredit the party. To counter the group's influence among militant British nationalists, he secured the American white nationalist militant William Pierce as a guest speaker at the BNP's annual rally in November 1995.
In the early 1990s, a "moderniser" faction emerged within the party, favouring a more electorally palatable strategy and an emphasis on building grassroots support to win local elections. They were impressed by the electoral gains made by a number of extreme-right parties in continental Europe—such as Jörg Haider's Austrian Freedom Party and Jean-Marie Le Pen's National Front—which had been achieved by both switching focus from biological racism to the perceived cultural incompatibility of different racial groups and by replacing anti-democratic platforms with populist ones.
The modernisers called for community campaigns among the white working-class populations of London's East End, and Northern England. While the modernisers gained some concessions from the party's hard-liners, Tyndall opposed many of their ideas and sought to stem their growing influence. In his view, "we should not be looking for ways of applying ideological cosmetic surgery to ourselves in order to make our features more appealing to the public".
After the BNP's poor performance at the 1997 general election, opposition to Tyndall's leadership grew. The modernisers called the party's first leadership election, and in October 1999 Tyndall was ousted when two-thirds of those voting backed Nick Griffin, who offered an improved administration, financial transparency, and greater support for local branches. Often characterised as a political chameleon, Griffin had once been considered a party hardliner before switching allegiance to the modernisers in the late 1990s. In his youth, he had been involved in the NF as well as Third Positionist groups like Political Soldier and the International Third Position. Criticising his predecessors for fuelling the image of the BNP as "thugs, losers and troublemakers", Griffin inaugurated a period of change in the party.
Influenced by Le Pen's National Front in France, Griffin sought to widen the BNP's appeal to individuals who were concerned about immigration but had not previously voted for the extreme-right. The BNP replaced Tyndall's policy of compulsory deportation of non-whites to a voluntary system whereby non-whites would be given financial incentives to emigrate. It downplayed biological racism and stressed the cultural incompatibility of different racial groups. This emphasis on culture allowed it to foreground Islamophobia, and following the September 11 attacks in 2001 it launched a "Campaign Against Islam". It stressed the claim that the BNP was ""not" a racist party" but an "organised response to anti-white racism". At the same time Griffin sought to reassure the party's base that these reforms were based on pragmatism and not a change in principle.
Griffin also sought to shed the BNP's image as a single-issue party, by embracing a diverse array of social and economic issues. Griffin renamed the party's monthly newspaper from "British Nationalist" to "The Voice of Freedom", and established a new journal, "Identity". The party developed community-based campaigns, through which it targeted local issues, particularly in those areas with large numbers of skilled white working-class people who were disaffected with the Labour Party government. For instance, in Burnley it campaigned for lower speed limits on housing estates and against the closure of a local swimming bath, while in South Birmingham it targeted pensioners' concerns about youth gangs. In 2006 the party urged its activists to carry out local activities like cleaning up children's play areas and removing graffiti while wearing high-vis jackets emblazoned with the party logo.
Griffin believed that Peak Oil and a growth in Third World migrants arriving in Britain would result in a BNP government coming to power by 2040.
The close of the twentieth century produced more favourable conditions for the extreme-right in Britain as a result of increased public concerns about immigration and established Muslim communities coupled with growing dissatisfaction with the established mainstream parties. In turn, the BNP gained rapidly growing levels of support over the coming years. In July 2000, it came second in the council elections for the North End of the London Borough of Bexley, its best result since 1993. At the 2001 general election it gained 16% of the vote in one constituency and over 10% in two others. In the 2002 local elections the BNP gained four councillors, three of whom were in Burnley, where it had capitalised on white anger surrounding the disproportionately high levels of funding being directed to the Asian-dominated Daneshouse ward. This breakthrough generated public anxieties about the party, with a poll finding that six in ten supported a ban on it. In the 2003 local elections the BNP gained 13 additional councillors, including seven more in Burnley, having attained over 100,000 votes. Concerned that much of their potential vote was going to the UK Independence Party (UKIP), in 2003 the BNP offered UKIP an electoral pact but was rebuffed. Griffin then accused UKIP of being a Labour Party scheme to steal the BNP's votes. They invested much in the campaign for the 2004 European Parliament election, at which they gained 800,000 votes but failed to secure a parliamentary seat. In the 2004 local elections, they secured four more seats, including three in Epping.
For the 2005 general election, the BNP expanded its number of candidates to 119 and targeted specific regions. Its average vote in the areas it contested rose to 4.3%. It gained significantly more support in three seats, achieving 10% in Burnley, 13% in Dewsbury, and 17% in Barking. In the 2006 local elections the party gained 220,000 votes, with 33 additional councillors, having averaged a vote share of 18% in the areas it contested. In Barking and Dagenham, it saw 12 of its 13 candidates elected to the council. At the 2008 London Assembly election, the BNP gained 130,000 votes, reaching the 5% mark and thus gaining an Assembly seat. At the 2009 European Parliament election, the party gained almost 1 million votes, with two of its candidates, Nick Griffin and Andrew Brons, being elected as Members of the European Parliament for North West England and Yorkshire and the Humber respectively. That election also saw extreme-right parties winning seats for various other EU member-states. This victory marked a major watershed for the party. Amid significant public controversy, Griffin was invited to appear on the BBC show "Question Time" in October 2009, the first time that the BNP had been invited to share a national television platform with mainstream panellists. Griffin's performance was however widely regarded as poor.
Despite its success, there was dissent in the party. In 2007 a group of senior members known as the "December rebels" challenged Griffin, calling for internal party democracy and financial transparency, but were expelled. In 2008, a group of BNP activists in Bradford split to form the Democratic Nationalists. In November 2008, the BNP membership list was posted to WikiLeaks, after appearing briefly on a weblog. A year later, in October 2009, another list of BNP members was leaked.
Eddy Butler then led a challenge to Griffin's leadership, alleging financial corruption, but he had insufficient support. The rebels who supported him split into two groups: one section remained as the internal Reform Group, the other left the BNP to form the British Freedom Party.
By 2010, there was discontent among the party's grassroots, a result of the change to its white-only membership policy and rumours of financial corruption among its leadership. Some defected to the National Front or left to form parties like the Britannica Party. Anti-fascist groups like Hope not Hate had campaigned extensively in Barking to stop the area's locals voting for the BNP. At the 2010 general election, the BNP had hoped to make a breakthrough by gaining a seat in the House of Commons, although it failed to achieve this. It nevertheless gained the fifth largest national vote share, with 1.9% of the vote, representing the most successful electoral performance for an extreme-right party in UK history. In the 2010 local elections, it lost all of its councillors in Barking and Dagenham. Nationally, the party's number of councillors dropped from over fifty to 28. Griffin described the results as "disastrous".
In a 2011 leadership election, Griffin secured a narrow victory, beating Brons by nine votes of a total of 2,316 votes cast. In October 2012, Brons left the party, leaving Griffin as its sole MEP. In the 2012 local elections, the party lost all of its seats and saw its vote share fall dramatically; whereas it gained over 240,000 votes in 2008, this had fallen to under 26,000 by 2012. Commenting on the result, the political scientist Matthew Goodwin noted: "Put simply, the BNP's electoral challenge is over." In the 2012 London mayoral election, the BNP candidate came seventh, with 1.3% of first-preference votes, its poorest showing in the London mayoral contest. The 2012 election results established that the BNP's steady growth had ended. In the 2013 local elections, the BNP fielded 99 candidates but failed to win any council seats, leaving it with only two.
In June 2013, Griffin visited Syria along with members of Hungarian far-right party Jobbik to meet with government officials, including the Speaker of the Syrian People's Assembly, Mohammad Jihad al-Laham, and the Prime Minister Wael Nader al-Halqi. Griffin claims he was influential in the speaker of Syria's Parliament writing an open letter to British MPs urging them to "turn Great Britain from the warpath" by not intervening in the Syrian conflict. Griffin lost his European Parliament seat in the May 2014 European election. The party blamed the UK Independence Party for its decline, accusing the latter of stealing BNP policies and slogans. In July 2014, Griffin resigned and was succeeded by Adam Walker as acting chairman. In October, Griffin was expelled from the party for "trying to cause disunity [in the party] by deliberately fabricating a state of crisis".
In January 2015, membership of the party numbered 500, down from 4,220 in December 2013. At the general election in 2015, the BNP fielded eight candidates, down from 338 in 2010. The party's vote share declined 99.7% from its 2010 result. In January 2016, the Electoral Commission de-registered the BNP for failing to pay its annual registration fee of £25 GBP. At this time, it was estimated that BNP assets totalled less than £50,000 GBP. According to the Commission, "BNP candidates cannot, at present, use the party's name, descriptions or emblems on the ballot paper at elections." A month later, the party was re-registered. There were ten BNP candidates at the general election in 2017. At the 2018 local elections, the party's last remaining councillor—Brian Parker of Pendle—decided not to stand for re-election, leaving the party without representation at any level of UK government. The BNP fielded only one candidate at the 2019 general election in Hornchurch and Upminster, where it came last.
Many academic historians and political scientists have described the BNP as a far-right party, or as an extreme-right party. As the political scientist Matthew Goodwin used it, the term referred to "a particular form of political ideology that is defined by two anti-constitutional and anti-democratic elements: first, right-wing extremists are "extremist" because they reject or undermine the values, procedures and institutions of the democratic constitutional state; and second they are "right-wing" because they reject the principle of fundamental human equality".
Various political scientists and historians have described the BNP as being fascist in ideology. Others have instead described it as neo-fascist, a term which the historian Nigel Copsey argued was more exact. Academic observers—including the historians Copsey, Graham Macklin, and Roger Griffin, and the political theologian Andrew P. Davey—have argued that Nick Griffin's reforms were little more than a cosmetic process to obfuscate the party's fascist roots. According to Copsey, under Griffin the BNP was "fascism recalibrated – a form of neo-fascism – to suit contemporary sensibilities". Macklin noted that despite Griffin's 'modernisation' project, the BNP retained its ideological continuity with earlier fascist groups and thus had not transformed itself into a genuinely "post-fascist" party. In this it was distinct from parties like the Italian National Alliance of Gianfranco Fini, which has been credited with successfully shedding its fascist past and becoming post-fascist.
The anti-fascist activist Gerry Gable referred to the BNP as a "Nazi organisation", while the Anti-Nazi League published leaflets describing the BNP as the "British Nazi Party". Copsey suggested that while the BNP under Tyndall could be described as neo-Nazi, it was not "crudely mimetic" of the original German Nazism. Davey characterised the BNP as a "populist ethno-nationalist" party.
In his writings, Griffin acknowledged that much of his 'modernisation' was an attempt to hide the BNP's core ideology behind more electorally palatable policies. Like the National Front, the BNP's private discourse differed from its public one, with Griffin stating that "Of course we must teach the truth to the hardcore... [but] when it comes to influencing the public, forget about racial differences, genetics, Zionism, historical revisionism and so on... we must at all times present them with an image of moderate reasonableness". The BNP has eschewed the labels "fascist" and "Nazi", claiming that it is neither. In its 1992 electoral manifesto, it insisted that "Fascism was Italian. Nazism was German. We are British. We will do things our own way; we will not copy foreigners". In 2009, Griffin that the term "fascism" was simply "a smear that comes from the far left"; he added that the term should be reserved for groups that engaged in "political violence" and desired a state that "should impose its will on people", claiming that it was the anti-fascist group Unite Against Fascism—and not the BNP—who were the real fascists. More broadly, many on Britain's extreme-right sought to avoid the term "British fascism" because of its electorally unpalatable connotations, utilising "British nationalism" in its place.
After Griffin took control of the party, it made increasing use of nativist themes in order to emphasise its "British" credentials. In its published material, the party made appeals to the idea of Britain and Britishness in a manner not dissimilar to mainstream political parties. In this material it has also made prominent use of the Union flag and the colours red, white, and blue. Roger Griffin noted that the terms "Britain" and "England" appear "confusingly interchangeable" in BNP literature, while Copsey has pointed out that the BNP's form of British nationalism is "Anglo-centric". The party employed militaristic rhetoric under both Tyndall and Griffin's leadership; under the latter for example its published material spoke of a "war without uniforms" and a "war for our survival as a people". Tyndall described the BNP as a revolutionary party, calling it a "guerrilla army operating in occupied territory".
The BNP adheres to biological racist ideas, displaying an obsession with the perceived differences of racial groups. Both Tyndall and Griffin believed that there was a biologically distinct white-skinned "British race" which was one branch of a wider Nordic race, a view akin to those of earlier fascists such as Hitler and Arnold Leese.
The BNP adheres to an ideology of ethnic nationalism. It promotes the idea that not all citizens of the United Kingdom belong to the British nation. Instead, it claims that the nation only belongs to "the English, Scots, Irish and Welsh along with the limited numbers of peoples of European descent, who have arrived centuries or decades ago and who have fully integrated into our society". This is a group that Griffin referred to as the "home people" or "the folk". According to Tyndall, "The BNP is a racial nationalist party which believes in Britain for the British, that is to say racial separatism." Richard Edmonds in 1993 told "The Guardian"'s Duncan Campbell that "we [the BNP] are 100% racist". The BNP does not regard UK citizens who are not ethnic white Europeans as "British", and party literature calls on supporters to avoid referring to such individuals as "Black Britons" or "Asian Britons", instead describing them as "racial foreigners".
Tyndall believed the white British and the broader Nordic race to be superior to other races, and under his leadership, the BNP promoted pseudoscientific claims in support of white supremacy. Following Griffin's ascendency to power in the party, it officially repudiated racial supremacism and insisted that no racial group was superior or inferior to another. Instead it foregrounded an "ethno-pluralist" racial separatism, claiming that different racial groups had to be kept separate and distinct for their own preservation, maintaining that global ethno-cultural diversity was something to be protected. This switch in focus owed much to the discourse of the French Nouvelle Droite movement which had emerged within France's extreme-right during the 1960s.
At the same time the BNP switched focus from openly promoting biological racism to stressing what it perceived as the cultural incompatibility of racial groups. It placed great focus on opposing what it referred to as "multiculturalism", characterising this as a form of "cultural genocide", and claiming that it promoted the interests of non-whites at the expense of the white British population. However, internal documents produced and circulated under Griffin's leadership demonstrated that—despite the shift in its public statements—it remained privately committed to biological racist ideas.
The party emphasises what it sees as the need to protect the racial purity of the white British. It condemns miscegenation and "race mixing", claiming that this is a threat to the British race. Tyndall stated that he "felt deeply sorry for the child of a mixed marriage" but had "no sympathy whatsoever for the parents". Griffin similarly stated that mixed-race children were "the most tragic victims of enforced multi-racism", and that the party would not "accept miscegenation as moral or normal ... we never will". In its 1983 election manifesto, the BNP stated that "family size is a private matter" but still called for white Britons who are "of intelligent, healthy and industrious stock" to have large families and thus raise the white British birth-rate. The encouragement of high birth rates among white British families continued under Griffin's leadership.
Under Tyndall's leadership, the BNP promoted eugenics, calling for the forced sterilisation of those with genetically transmittable disabilities. In party literature, it talked of improving the British 'racial stock' by removing "inferior strains within the indigenous races of the British Isles". Tyndall argued that medical professionals should be responsible for determining whom to sterilise, while a lowering of welfare benefits would discourage breeding among those he deemed to be genetic inferiors. In his magazine "Spearhead", Tyndall also stated that "the gas chamber system" should be used to eliminate "sub-human elements", "perverts", and "asocials" from British society.
Opposition to immigration has been central to the BNP's political platform. It has engaged in xenophobic campaigns which emphasise the idea that immigrants and ethnic minorities are both different from, and a threat to, the white British and white Irish populations. In its campaign material it presented non-whites both as a source of crime in the UK, and as a socio-economic threat to the white British population by taking jobs, housing, and welfare away from them. It engaged in welfare chauvinism, calling for white Britons to be prioritised by the UK's welfare state. Party literature included such as claims as that the BNP was the only party which could "do anything effective about the swamping of Britain by the Third World" or "lead the native peoples of Britain in our version of the New Crusade that must be organised if Europe is not to sink under the Islamic yoke".
Much of its published material made claims about a forthcoming race war and promoted the conspiracy theory about white genocide. In a 2009 radio interview, Griffin referred to this as a "bloodless genocide". It presents the idea that white Britons are engaged in a battle against their own extinction as a racial group. It reiterated a sense of urgency about the situation, claiming that both high immigration rates and high birth rates among ethnic minorities were a threat to the white British. In 2010, it for instance was promoting the idea that at current levels, "indigenous Britons" would be a minority within the UK by 2060.
The BNP calls for the non-white population of Britain to either be reduced in size or removed from the country altogether. Under Tyndall's leadership it promoted the compulsory removal of non-whites from the UK, stating that under a BNP government they would be "repatriated" to their countries of origin. In the early 1990s it produced stickers with the slogan "Our Final Solution: Repatriation". Tyndall understood this to be a two-stage process that would take ten to twenty years, with some non-whites initially leaving willingly and the others then being forcibly deported. During the 1990s, party modernisers suggested that the BNP move away from a policy of compulsory repatriation and toward a voluntary system, whereby non-white persons would be offered financial incentives to leave the UK. This idea, adopted from Powellism, was deemed more electorally palatable.
When Griffin took control of the party, the policy of voluntary repatriation was officially adopted, with the party suggesting that this could be financed through the use of the UK's pre-existing foreign aid budget. It stated that any non-whites who refused to leave would be stripped of their British citizenship and categorised as "permanent guests", while continuing to be offered incentives to emigrate. Griffin's BNP also stressed its support for an immediate halt to non-white immigration into Britain and for the deportation of any migrants illegally in the country. Speaking on the BBC's "Andrew Marr Show" in 2009, Griffin declared that, unlike Tyndall, he "does not want all-white UK" because "nobody out there wants it or would pay for it".
Under Tyndall's leadership, the BNP was openly anti-Semitic. From A. K. Chesterton, Tyndall had inherited a belief that there was a global conspiracy of Jews bent on world domination, viewing the "Protocols of the Elders of Zion" as genuine evidence for this. He believed that Jews were responsible for both communism and international finance capitalism and that they were responsible for undermining the British Empire and the British race. He believed that both democratic government and immigration into Europe were parts of the Jewish conspiracy to weaken other races. In an early edition of "Spearhead" published in the 1960s, Tyndall wrote that "if Britain were to become Jew-clean she would have no nigger neighbours to worry about... It is the Jews who are our misfortune: T-h-e J-e-w-s. Do you hear me? THE JEWS?" Tyndall added Holocaust denial to the anti-Semitic beliefs inherited from Chesterton, believing that the Holocaust was a hoax created by the Jews to gain sympathy for themselves and thus aid their plot for world domination. Among those to endorse such anti-Semitic conspiracy theories was Griffin, who promoted them in his 1997 pamphlet, "Who are the Mind Benders?" Griffin also engaged in Holocaust denial, publishing articles promoting such ideas in "The Rune", a magazine produced by the Croydon BNP. In 1998, these articles resulted in Griffin being convicted of inciting racial hatred.
When Griffin took power, he sought to banish overt anti-Semitic discourse from the party. He informed party members that "we can get away with criticising Zionists, but any criticism of Jews is likely to be legal and political suicide". In 2006, he complained that the "obsession" that many BNP members had with "the Jews" was "insane and politically disastrous". In 2004, the party selected a Jewish candidate, Pat Richardson, to stand for it during local council elections, something Tyndall lambasted as a "gimmick". References to Jews in BNP literature were often coded to hide the party's electorally unpalatable anti-Semitic ideas. For instance, the term "Zionists" was often used in party literature as a euphemism for "Jews". As noted by Macklin, Griffin still framed many of his arguments "within the parameters of recognizably anti-Semitic discourse". The BNP's literature is replete with references to a conspiratorial group who have sought to suppress nationalist sentiment among the British population, who have encouraged immigration and mixed-race relationships, and who are promoting the Islamification of the country. This group is likely a reference to the Jews, being an old fascist canard.
Sectors of the extreme-right were highly critical of Griffin's softening on the subject of the Jews, claiming that he had "sold out" to the 'Zionist Occupied Government'. In 2006, John Bean, editor of "Identity", included an article in which he reassured BNP members that the party had not "sold out to the Jews" or "embraced Zionism" but that it remained "committed to fighting... subversive Jews". Under Griffin, the BNP's website linked to other web pages that explicitly portrayed immigration as part of a Jewish conspiracy, while it also sold books that promoted Holocaust denial. In 2004, secretly filmed footage was captured in which Griffin was seen claiming that "the Jews simply bought the West, in terms of press and so on, for their own political ends".
Copsey noted that a "culture of anti-Semitism" still pervaded the BNP. In 2004, a London activist told reporters that "most of us hate Jews", while a Scottish BNP group was observed making Nazi salutes while shouting "Auschwitz". The party's Newcastle upon Tyne Central candidate compared the Auschwitz concentration camp to Disneyland, while their Luton North candidate stated her refusal to buy from "the kikes that run Tesco". In 2009, a BNP councillor from Stoke-on-Trent resigned from the party, complaining that it still contained Holocaust deniers and Nazi sympathisers.
Griffin informed BNP members that rather than "bang on" about the Jews—which would be deemed extremist and prove electorally unpopular—their party should focus on criticising Islam, an issue that would be more resonant among the British public. After Griffin took over, the party increasingly embraced an Islamophobic stance, launching a "Campaign Against Islam" in September 2001. In "Islam: A Threat to Us All", a leaflet distributed to London households in 2007, the BNP claimed that it would stand up to both Islamic extremism and "the threat that 'mainstream' Islam poses to our British culture". In contrast to the mainstream British view that the actions of militant Islamists—such as those who perpetrated the 7 July 2005 London bombings—are not representative of mainstream Islam, the BNP insists that they are. In some of its literature it presents the view that every Muslim in Britain is a threat to the country. Griffin referred to Islam as an "evil, wicked faith", and elsewhere publicly described it as a "cancer" that needed to be removed from Europe through "chemotherapy".
The BNP has called for the prohibition of immigration from Muslim countries and for the banning of the burka, halal meat, and the building of new mosques in the UK. It also called for the immediate deportation of radical Islamist preachers from the country. In 2005 the party claimed that its primary issue of concern was the "growth of fundamentalist-militant Islam in the UK and its ever-increasing threat to Western civilization and our implicit values". To broaden its anti-Islamic agenda, Griffin's BNP made overtures to the UK's Hindu, Sikh, and Jewish communities; Griffin's claim that Jews can make "good allies" in the fight against Islam caused controversy within the international far-right.
Tyndall believed that liberal democracy was damaging to British society, claiming that liberalism was a "doctrine of decay and degeneration". Under Tyndall, the party sought to dismantle the UK's liberal democratic system of parliamentary governance, although was vague about what it sought to replace this system with. In his 1988 work "The Eleventh Hour", Tyndall wrote of the need for "an utter rejection of liberalism and a dedication to the resurgence of authority". Tyndall's BNP perceived itself as a revolutionary force that would bring about a national rebirth in Britain, entailing a radical transformation of society. It proposed a state in which the Prime Minister would have full executive powers, and would be elected directly by the population for an indefinite period of time. This Prime Minister could be dismissed from office in a further election that could be called if Parliament produced a vote of no confidence in them. It stated that rather than having political parties, candidates standing for election to the parliament would be independent. During the period of Griffin's leadership, the party downplayed its anti-democratic themes and instead foregrounded populist ones. Its campaign material called for the devolution of greater powers to local communities, the reestablishment of county councils, and the introduction of citizens' initiative referendums based on those used in Switzerland.
The BNP has adopted a hard Eurosceptic platform from its foundation. Under Tyndall's leadership, the BNP had overt anti-Europeanist tendencies. Throughout the 1980s and 1990s he maintained the party's opposition to the European Economic Community. Antagonism toward what became the European Union was retained under Griffin's leadership, which called for the UK to leave the Union. One of Vote Leave's biggest donors during the Brexit referendum was former BNP member Gladys Bramall, and the party has claimed that its anti-Establishment rhetoric "created the road" to Britain's vote to leave the European Union.
Tyndall suggested replacing the EEC with a trading association among the "White Commonwealth", namely countries like Canada, Australia, and New Zealand. Tyndall held imperialist views and was sympathetic to the re-establishment of the British Empire through the recolonization of parts of Africa. However, officially the BNP had no plans to re-establish the British Empire or secure dominion over non-white nations. In the 2000s, it called for an immediate military withdrawal from both the Iraq War and the Afghan War. It has advocated ending overseas aid to provide economic support within the UK and to finance the voluntary repatriation of legal immigrants.
Under Tyndall, the BNP rejected both Welsh nationalism and Scottish nationalism, claiming that they were bogus because they caused division among the wider 'British race'. Tyndall also led the BNP in support of Ulster loyalism, for instance by holding public demonstrations against the Irish republican party Sinn Féin, and endorsing Ulster loyalist paramilitaries. Under Griffin, the BNP continued to support Ulster's membership of the United Kingdom, calling for the crushing of the Irish Republican Army and the scrapping of the Anglo-Irish Agreement. Griffin later expressed the view that "the only solution that could possibly be acceptable to loyalists and republicans alike" would be the reintegration of the Irish Republic into the United Kingdom, which would be reorganised along federal lines. However, while retaining the party's commitment to Ulster loyalism, under Griffin the importance of the issue was downplayed, something that was criticised by Tyndall loyalists.
Tyndall described his approach to the economy as "National Economics", expressing the view that "politics must lead, and not be led by, economic forces". His approach rejected economic liberalism because it did not serve "the national interest", although still saw advantages in a capitalist system, looking favourably on individual enterprise. He called on capitalist elements to be combined with socialist ones, with the government playing a role in planning the economy. He promoted the idea of the UK becoming an autarky which was economically self-sufficient, with domestic production protected from foreign competition. This attitude was heavily informed by the corporatist system that had been introduced in Benito Mussolini's Fascist Italy.
A number of senior members, including Griffin and John Bean, had anti-capitalist leanings, having been influenced by Strasserism and National Bolshevism. Under Griffin's leadership, the BNP promoted economic protectionism and opposed globalisation. Its economic policies reflect a vague commitment to distributist economics, ethno-socialism, and national autarky. The BNP maintains a policy of protectionism and economic nationalism, although in comparison with other far-right nationalist parties, the BNP focuses less on corporatism. It has called for British ownership of its own industries and resources and the "subordination of the power of the City to the power of the government". It has promoted the regeneration of farming in the United Kingdom, with the object of achieving maximum self-sufficiency in food production. In 2002, the party criticised corporatism as a "mixture of big capitalism and state control", saying it favoured a "distributionist tradition established by home-grown thinkers" favouring small business.
When it comes to environmentalism, the BNP refers to itself as the "real green party", claiming that the Green Party of England and Wales engages in "watermelon" politics by being green (environmentalist) on the outside but red (leftist) on the inside. Influenced by the Nouvelle Droite, it framed its arguments regarding environmentalism in an anti-immigration manner, talking about the need for 'sustainability'. It engages in climate change denial, with Griffin claiming that global warming is a hoax orchestrated by those trying to establish the New World Order.
The BNP is opposed to feminism and has pledged that—if in government—it would introduce financial incentives to encourage women to leave employment and become housewives. It would also seek to discourage children being born out of wedlock. It has stated that it would criminalise abortion, except in cases where the child has been conceived as a result of rape, the mother's life is threatened, or the child will be disabled. There are nevertheless circumstances where it has altered this anti-abortion stance; an article in "British Nationalist" stated that a white woman bearing the child of a black man should "abort the pregnancy... for the good of society". More widely, the party censures inter-racial sex and accuses the British media of encouraging inter-racial relationships.
Under Tyndall, the BNP called for the re-criminalisation of homosexual activity. Following Griffin's takeover, it moderated its policy on homosexuality. However, it opposed the 2004 introduction of civil partnerships for same-sex couples. During his 2009 "Question Time" appearance, Griffin described the sight of two men kissing as "really creepy". The party has also condemned the availability of pornography; its 1992 manifesto stated that the BNP would give the "pedlars of this filth... the criminal status that they deserve". The BNP promoted the reintroduction of capital punishment, and the sterilisation of some criminals. It also called for the reintroduction of national service in the UK, adding that on completion of this service adults would be permitted to keep their standard issue assault rifle.
According to the academic Steven Woodbridge, the BNP had a "rather ambivalent attitude toward Christian belief and religious themes in general" during most of its history, but under Griffin's modernisation the party increasingly utilised Christian terminology and themes in its discourse. Various members of the party presented themselves as "true Christians", and defenders of the faith, with key ideologues claiming that the religion has been "betrayed" and "sold out" by mainstream clergy and the British establishment. British Christianity, the BNP claimed, was under threat from Islam, Marxism, multiculturalism, and "political correctness". On analysing the BNP's use of Christianity, Davey argued that the party's emphasis was not on Christian faith itself, but on the inheritance of European Christian culture.
The BNP long considered the mainstream media to be one of its major impediments to electoral success. Tyndall claimed that the media represents a "state above the state" which was committed to the "left-liberal" goals of internationalism, liberal democracy, and racial integration. The party has claimed that the mainstream media has given disproportionate coverage to the achievements of ethnic minority sportsmen and to the victims of anti-black racism while ignoring white victims of racial prejudice and the BNP's activities. Both Tyndall and Griffin have claimed that the mainstream media is controlled by Jews, who use it for their own devices; the latter promoted this idea in his "Who are the Mind Benders?" Griffin has described the BBC as "a thoroughly unpleasant, ultra-leftist establishment". The BNP has stated that if it took power, it would end "the dictatorship of the media over free debate". It claims that it would introduce a law prohibiting the media from disseminating falsehoods about an individual or organisation for financial or political gain, and that it would ban the media from promoting racial integration.
BNP policy pledges to protect freedom of speech, as part of which it would repeal all laws banning racial or religious hate speech. It would repeal the 1998 Human Rights Act and withdraw from the European Convention on Human Rights.
In contrast to the UK's mainstream parties, the BNP received few donations from businesses, wealthy donors, or trade unions. Instead it relied on finances produced by its membership. Under Tyndall, the party operated on a shoestring budget with a lack of transparency; in 1992 it collected £5000 and in 1997 it collected £10,000. It also tried raising money by selling extreme-right literature, and opened a bookshop in Welling in 1989, although this was closed in 1996 after being attacked by anti-fascists and proving too costly to run. In 1992 the party formed a dining club of its wealthier supporters, which was renamed the Trafalgar Club in 2000. By the 1997 general election it admitted that its expenses had "far out-stripped" its income, and it was appealing for donations to pay off loans it had taken out.
Griffin placed greater emphasis on fundraising, and from 2001 through to 2008 the BNP's annual turnover increased almost fivefold. Membership subscriptions grew from £35,000 to £166,000, while its donations raised from £38,000 to £660,000. However, expenses also rose as the BNP spent more on its electoral campaigns, and the party reported a financial deficit in 2004 and again in 2005. Between 2007 and 2009 the BNP accumulated debts of £500,000.
For most of its history, the BNP had a whites-only membership policy. In 2009, the state's Equality and Human Rights Commission stated that this was a violation of the Race Relations Act 1976 and called on the party to amend its constitution accordingly. Responding to this, in early 2010 members voted to remove the racial restriction to membership, although it is unlikely that many non-whites joined.
At its creation, the BNP had approximately 1,200 members. By the 1983 general election this had grown to approximately 2,500, although by 1987 had slumped to 1000, with no significant further growth until the 21st century. After taking control Griffin began publishing the party's membership figures: 2,174 in 2001, 3,487 in 2002, 5,737 in 2003, and 7,916 in 2004. Membership dropped slightly to 6,281 in 2005, but had grown to 9,297 in 2007 and to 10,276 in spring 2010. In 2011, it was noted that this meant that the BNP had experienced the most rapid growth since 2001 of any minor party in the UK.
A party membership list dating from late 2007 was leaked onto the internet by a disgruntled activist, containing the names and addresses of 12,000 members. This included names, addresses and other personal details. People on the list included prison officers (barred from BNP membership), teachers, soldiers, civil servants and members of the clergy. The leaked list indicated that membership was concentrated in particular areas, namely the East Midlands, Essex, and Pennine Lancashire, but with particular clusters in Charnwood, Pendle, and Amber Valley. Many of these areas had long been targeted by extreme-right campaigns, dating back to the NF activity of the 1970s, suggesting that such longstanding activism may have had an effect on levels of BNP membership. This information also revealed that membership was most likely in urban areas with low rates of educational attainment and large numbers of economically insecure people employed in manufacturing, with further correlations to nearby Muslim communities. Following an investigation by Welsh police and the Information Commissioner's Office, two people were arrested in December 2008 for breach of the Data Protection Act concerning the leak. Matthew Single was subsequently found guilty and fined £200. The 'low' fine was criticised as an "absolute disgrace" by a BNP spokesman and a detective sergeant involved said he was "disappointed" with the outcome.
The leaked membership list showed that the party was 17.22% female. While women have occupied key positions within the BNP, men dominated at every level of the party. In 2009, over 80% of the party's Advisory Council was male and from 2002 to 2009, three-quarters of its councillors were male. The average percentage of female candidates presented at local elections in 2001 was 6%, although this had risen to 16% by 2010. Since 2006, the party had made a point of selecting female candidates, with Griffin stating that this was necessary to "soften" the party's image. Goodwin suggested that membership fell into three camps: the "activist old guard" who had previously been involved in the NF during the 1970s, the "political wanderers" who had defected from other parties to the BNP, and the "new recruits" who had joined post-2001 and who had little or no political interest or experience beforehand.
Having performed qualitative research among the BNP by interviewing various members, Goodwin noted that few of those he interviewed "conformed to the popular stereotypes of them being irrational and uninformed crude racists". He noted that most strongly identified with the working class and claimed to have either been former Labour voters or from a Labour-voting family. None of those interviewed claimed a family background in the ethnic nationalist movement. Instead, he noted that members claimed that they joined the party as a result of a "profound sense of anxiety over immigration and rising ethno-cultural diversity" in Britain, along with its concomitant impact on "British culture and society". He noted that among these members, the perceived cultural threat of immigrants and ethnic minorities was given greater prominence than the perceived economic threat that they posed to white Britons. He noted that in his interviews with them, members often framed Islam in particular as a threat to British values and society, expressing the fear that British Muslims wanted to Islamicise the country and eventually impose "sharia" law on its population.
Goodwin described the BNP's voters as being "socially distinct and concerned about a specific set of issues". Under Griffin's leadership, the party targeted areas with high proportions of skilled white working-class voters, particularly those who were disenchanted with the Labour government. It has attempted to appeal to disaffected Labour voters with slogans such as "We are the Labour Party your Grandfather Voted For". The BNP had little success in gaining support from women, the middle classes, and the more educated.
Goodwin noted a "strong male bias" in the party's support base, with statistical polling revealing that between 2002 and 2006, seven out of ten BNP voters were male. That same research also indicated that BNP voters were disproportionately middle-aged and elderly, with three quarters being aged over 35, and only 11% aged between 18 and 24. This contrasted to the NF's support base during the 1970s, when 40% of its voters were aged between 18 and 24. Goodwin suggested two possibilities for the BNP's failure to appeal to younger voters: one was the 'life cycle effect', that older people have obtained more during their life and thus have more to lose, feeling both more threatened by change and more socially conservative in their views. The other explanation was the 'generational effect', with younger Britons who have grown up since the onset of mass immigration having had greater social exposure to ethnic minorities and thus being more tolerant toward them. Conversely, many older voters came of age during the 1970s, under the impact of the anti-immigrant rhetoric promoted by Powellism, Thatcherism, and the NF, and thus have less tolerant attitudes.
Most BNP voters had no formal qualifications and the party's support was centred largely in areas with low educational attainment. According to the 2002–06 data, two-thirds of BNP voters had either no formal qualifications or had left education after their O-levels/GCSEs. Only one in ten BNP voters possessed an A-level, and an even smaller percentage had a university degree. Most of the BNP's voting base were from the financially insecure lower classes. Research conducted from 2002 to 2006 indicated that seven out of ten BNP voters were either skilled or unskilled workers or unemployed. A 2009 poll found that six out of ten BNP voters fitted this profile. Goodwin suggested that it was the skilled working classes rather than their unskilled or unemployed neighbours who were the main support base behind the BNP, because they owned some assets and thus felt that they had more to lose as a result of the economic threat posed by immigrants and ethnic minorities.
Research indicated that BNP voters also held opinions that were distinct from the average British citizen. They were far more pessimistic about their economic prospects than average, with seven out of ten BNP voters expecting their economic prospects to decline in future, contrasted with four out of ten who held this view in the wider population. In the 2002–06 period, 59% of BNP voters considered immigration to be the most important issue facing the UK, compared with only 16% of the wider population who agreed. By 2009, 87% of BNP voters identified immigration and asylum as the most important issue, to 49% of the wider population. BNP voters were also more likely to identify law and order, the EU, and Islamic extremism as the most important issues facing the UK than other voters, and less likely than average to rate the economy, NHS, pensions, and housing market as the most important.
BNP voters were also more likely than average to believe both that white Britons face unfair discrimination, and that Muslims, non-whites, and homosexuals had unfair advantages in British society. 78% of BNP voters endorsed the belief that the Labour Party prioritised immigrants and ethnic minorities over white British people, to 44% of the wider population. When asked questions about immigration and Muslims, BNP voters were found to be far more hostile to them than the average Briton, and also more willing than average to support outright racially discriminatory policies toward them. Copsey believed that "popular racism"—namely against asylum seekers and Muslims—generated the BNP's "largest reservoir of support", and that in many Northern English towns the main factors behind BNP support were white resentment toward Asian communities, anger at Asian-on-white crime, and the perception that Asians received disproportionately high levels of public funding.
Research also indicated that BNP voters were more mistrustful of the establishment than average citizens. In 2002–06, 92% of BNP voters described themselves as being dissatisfied with the government, to 62% of the wider population. Over 80% of BNP voters were found to distrust their local Member of Parliament, council officials, and civil servants, and were also more likely than average to think that politicians were personally corrupt. There was also a tendency for BNP voters to read tabloids like the "Daily Mail", "Daily Express", and "The Sun", all of which promote anti-immigration sentiment. Whether these voters gained such sentiment as a result of reading these tabloids or they read these tabloids because it endorsed their pre-existing views is unclear.
The early stronghold of the BNP was in London, where it established enclaves of support in the boroughs of Enfield, Hackney, Lewisham, Southwark, and Tower Hamlets, with smaller units in Bexley, Camden, Greenwich, Hillingdon, Lambeth, and Redbridge. By the late 1990s, the party was increasingly retreating from its original East End heartland, finding that its electoral support had declined in the area. Griffin expressed the view that it was too dangerous for BNP activists to campaign in the East End, suggesting that they would likely be attacked by opponents. Instead the party shifted its focus to parts of Outer London, in particular the boroughs of Barking, Bexley, Dagenham, Greenwich, and Havering. After Griffin took power, the party focused on building support in the North of England, taking advantage of the anxieties generated by the ethnic riots that took place in Bradford, Oldham, and Burnley in 2001. In the period between 2002 and 2006, over 40% of the BNP's voters were in Northern England.
The decline of the BNP as an electoral force around 2014 helped to open the way for the growth of another right-wing party, UKIP. In a study Goodwin produced with Robert Ford, the two political scientists noted that UKIP's support base mirrored the BNP's in that it had the same "very clear social profile": the "old, male, working class, white and less educated". One area where the two differed, they noted, was in the fact that BNP support had been highest among the middle-aged before tailing off among the over 55s, whereas UKIP retained strong support with those over 55. Ford and Goodwin suggested that this might be because more over 55s had "direct or indirect experiences" of the Second World War, in which Britain defeated the fascist powers, resulting in them being less inclined to support fascist parties than their younger counterparts. Despite these commonalities, UKIP proved far more successful at mobilising these social groups than did the BNP. This was likely in part because UKIP had a "reputational shield"; it emerged from within the Eurosceptic tradition of British politics rather than from the far-right and thus, while often ridiculed by the mainstream, was regarded as a legitimate democratic actor in a way that the BNP was not.
On its formation, the BNP avoided the National Front's committee-rule system of collective leadership in the hope of evading the infighting and factionalism that had damaged the NF. Instead it was founded around what it called the "leadership principle", with a central chairman having complete control over the party, which was then arranged in a highly hierarchical structure. The BNP lacked any internal democracy, with the grassroots membership having no formal powers. On taking power, Griffin retained the leadership principle inherited from Tyndall. He nevertheless established an Advisory Council which would meet several times a year; the members were to be selected by Griffin himself and would serve as his advisors.
The party's branches and local groups were referred to as "units" within the party. These were designed to recruit followers, raise funds, and campaign during elections. Under Tyndall, the party operated with a skeleton organisation. It had no full-time staff and for most of the 1980s lacked a telephone number. Instead it relied on a handful of geographically scattered, unpaid regional organisers. Its early activists were recruited from within the extreme-right movement, and thus lacked the experience and skills in electoral campaigning. When Griffin took control, he introduced a variety of internal departments to help manage the party's activities: the administration and enquiries department, department for group development, legal affairs department, security department, and communications department. Griffin tried to build a more professional party machine by educating and training BNP members, providing them with incentives, establishing a steady income stream, and overcoming factionalism and dissent. He launched an "annual college" for activists in 2001 and formed an education and training department in 2007. In 2008 and 2010 he oversaw the establishment of "summer schools" for high-ranking officials. The party also began employing full-time members of staff, having three in 2001 and 13 in 2007.
To incentivise members to remain committed to the party, Griffin followed the example of the Swedish National Democrats by implementing a new "voting membership" scheme in 2007. This meant that those who had been BNP members for two years could become a "voting member", at which they would go on a year's probation. During this year they were required to attend educational and training seminars, to engage in a certain amount of activism, and to donate a specified amount of money to the party. Once completed, they were allowed to vote on certain matters at general members' meetings and annual conferences, to participate in policy debates, and to be eligible for intermediate and senior positions. This policy ensured that those who reached the higher echelons of the BNP were fully trained in the party's ideology and electoral strategy.
Griffin hoped to build a wider social movement around the BNP by establishing affiliated networks and organisations. In many cases, these were presented to the public in a way that concealed any direct connection to the BNP. Most of these affiliated groups were poorly funded and had few members. The party established its own record label, Great White Records, a radio station, and a trade union known as Solidarity – The Union for British Workers. It formed a group for young people known as the Young BNP, although in 2010 renamed this group as the BNP Crusaders, "to pay homage to our ancestors from the Middle Ages who saved Christian Europe from the onslaught of Islam". It established a Land and People group to recruit support in rural areas, a Family Circle to recruit women and families, and both a Veterans Group and an Association of British ex-Servicemen for former military servicemen. A group called Families Against Immigrant Racism was established to counter perceived racism against white Britons, while an Ethnic Liaison Committee was created to build links with anti-Muslim Hindu and Sikh groups active in Britain. Another group was the American Friends of the British National Party (AFBNP), set up by Mark Cotterill in 1999 to gain support from sympathisers in the United States. In 2001 it had 100 members, and by 2008 had 107.
A group called Islands of the North Atlantic (IONA) was established to promote the BNP's view of British culture and identity. The British Students Association was founded to promote the party's views among university students in 2000.
Albion Life Insurance was set up in September 2006 as an insurance brokerage company established on behalf of the BNP to raise funds for its activities. The firm ceased to operate in November 2006. In 2006, the BNP launched the Christian Council of Britain (CCB), a group designed to rival the Muslim Council of Britain and oppose the growing "Islamification" of inner city areas. The CCB was established and run by BNP member Robert West, who claimed to have been ordained by the Apostolic Church, a claim that the church denies. West is a Calvinist and espouses a theology of nations which is influenced by Calvinist theologians like Abraham Kuyper, holding that God wishes every race and nation to remain separate until end time.
Griffin's BNP also established an annual Red, White and Blue festival, which was based on the 'Blue Blanc Rouge' organised by France's National Front. The festival brought party activists together and aimed to promote a more family friendly image for the group, although it also provided a venue for white power skinhead bands like Stigger, Nemesis and Warlord. Around 1,000 BNP members attended the party's 2001 festival.
Under Griffin's leadership, the BNP zealously embraced the use of alternative media to promote itself in a way different from the negative portrayal that featured in the mainstream media. On its website—which had been established in 1995—it created an internet television channel, 'BNPtv'. It has created blogs that cover different themes without being explicitly political in order to promote the party's message. The BNP established an online marketing platform, Excalibur, through which to sell its merchandise. In 2003, the BNP claimed that it had the most viewed website of a political party in Britain, and by 2011 was claiming to have the most viewed such website in Europe. In September 2007, "The Daily Telegraph" newspaper reported that Hitwise, the online competitive intelligence service, said that the BNP website had more hits than any other website of a British political party.
Under Griffin, the BNP forged stronger links with various extreme-right parties elsewhere in Europe, among them France's National Front, Germany's National Democratic Party (NPD), Sweden's National Democrats, and Hungary's Jobbik. Griffin unsuccessfully urged the NPD to move away from neo-Nazism and embark on the same 'modernisation' project that he had taken the BNP. Jean-Marie Le Pen of the French Front National was the guest of honour at an "Anglo-French Patriotic Dinner" held by the BNP in April 2004.
Griffin met leaders of the Hungarian far right party Jobbik to discuss co-operation between the two parties and spoke at a Jobbik party rally in August 2008. In April 2009, Simon Darby, deputy chairman of the BNP, was welcomed with fascist salutes by members of the Italian nationalist Forza Nuova during a trip to Milan. Darby stated that the BNP would look to form an alliance with France's Front National in the European Parliament. Following the election of two BNP MEPs in 2009, the following year saw the BNP join with other extreme-right parties to form the Alliance of European National Movements, with Griffin becoming its vice president. The party also had close links with the Historical Review Press, a publisher focused on promoting Holocaust denial.
Britain's extreme-right has long faced internal and public divisions. Disgruntled BNP members left the party to found or join a wide range of rivals, among them the British Freedom Party, White Nationalist Party, Nationalist Alliance, Wolf's Hook White Brotherhood, British People's Party, England First Party, Britain First, Democratic Nationalists, and the New Nationalist Party. Various BNP members were involved in the nascent English Defence League (EDL)—with EDL leader Tommy Robinson having been a former BNP activist—although Griffin proscribed the organisation and condemned it as having been manipulated by "Zionists". The political scientist Chris Allen noted that the EDL shared much of the BNP's ideology, but that its "strategies and actions" were very different, with the EDL favouring street marches over electoral politics. By 2014, both the BNP and EDL were in decline, and Britain First—founded by former BNP members James Dowson and Paul Golding—had risen to prominence. It combined the electoral tactics of the BNP with the street marches of the EDL.
The Steadfast Trust was established as a charity in 2004 with the stated aims of reducing poverty among those of Anglo-Saxon descent and supporting English culture. It has many former and current BNP, NF and British Ku Klux Klan members. It was deregistered as a charity by the Charity Commission in February 2014. In 2014, after Nick Griffin lost the leadership of BNP, he set up British Voice, but before it was launched, he decided to set up a different group, British Unity.
Some members of the BNP were radicalised during their involvement with the party and subsequently sought to carry out acts of violence and terrorism. Tony Lecomber was imprisoned for three years for possessing explosives, after a nail bomb exploded while he was transporting it to the offices of the Workers' Revolutionary Party in 1985. He was imprisoned for three years in 1991 whilst serving as the BNP's Director of Propaganda for assaulting a Jewish teacher. In 1999, the ex-BNP member David Copeland used nail bombs to target homosexuals and ethnic minorities in London. In 2005, the BNP's Burnley candidate Robert Cottage was convicted of stockpiling chemicals for use in what he believed was a coming civil war, while a Yorkshire BNP member, Terry Gavan, was convicted in 2010 for stockpiling firearms and nail bombs.
The BNP has contested seats in England, Wales, Scotland and Northern Ireland. Research from Robert Ford and Matthew Goodwin shows that BNP support is concentrated among older and less educated working-class men living in the declining industrial towns of the North and Midlands regions, in contrast to previous significant far-right parties like the National Front, which drew support from a younger demographic.
The BNP placed comparatively little emphasis on elections to the British House of Commons, aware that the first past the post voting system was a major obstacle.
The British National Party has contested general elections since 1983.
The BNP in the 2001 general election saved five deposits (out of 33 contested seats) and secured its best general election result in Oldham West and Royton (which had recently been the scene of racially motivated rioting between white and Asian youths) where party leader Nick Griffin secured 16% of the vote.
The 2005 general election was considered a major breakthrough by the BNP, as they picked up 192,746 votes in the 119 constituencies it contested, took a 0.7% share of the overall vote, and retained a deposit in 40 of the seats.
The BNP put forward candidates for 338 out of 650 seats for the 2010 general election gaining 563,743 votes (1.9%), finishing in fifth place and failing to win any seats. However, a record of 73 deposits were saved. Party chairman Griffin came third in the Barking constituency, behind Margaret Hodge of Labour and Simon Marcus of the Conservatives, who were first and second respectively. At 14.6%, this was the BNP's best result in any of the seats it contested that year.
The BNP's first electoral success came in 1993, when Derek Beackon was returned as a councillor in Millwall, London. He lost his seat in elections the following year. The next BNP success in local elections was not until the 2002 local elections, when three BNP candidates gained seats on the Burnley council. The BNP's first councillor for six years was John Haycock, elected as a parish councillor for Bromyard and Winslow in Herefordshire in 2000. Haycock failed to attend any council meetings for six months and was later disqualified from office.
The party had 55 councillors for a time in 2009. After the 2013 local county council elections, the BNP was left with a total of two borough councillors in England:
As of 2011, the BNP had yet to make "a major breakthrough" on local councils.
The BNP's councillors usually had "an extremely limited impact on local politics" because they were isolated as individuals or small groups on the council. Councillors from the main parties often disliked their BNP colleagues and deemed having to work alongside them as an affront to dignity and decency.
Questions were often raised as to whether BNP councillors could adequately represent the interests of all of their local constituents. On being elected, Beackon for instance stated that he refused to serve his Asian constituents in Millwall. There were also allegations made that BNP councillors had particularly low attendance at council meetings, although research indicated that this was not the case, with the BNP's attendance record being largely average.
There is evidence to suggest that racially and religiously motivated crime increased in those areas where BNP councillors had been elected. For instance, after the 1993 election of Beackon, there was a spike in racist attacks in the borough of Tower Hamlets. BNP members were directly responsible for some of this; the party's national organiser Richard Edmonds was sentenced to three months imprisonment for his part in an attack on a black man and his white girlfriend.
BNP lead candidate Richard Barnbrook won a seat in the London Assembly in May 2008, after the party gained 5.3% of the London-wide vote. However, in August 2010, he resigned the party whip and became an independent.
In the 2007 Welsh Assembly elections, the BNP fielded 20 candidates, four in each of the five regional lists, with Nick Griffin standing in the South Wales West region. It did not win any seats, but was the only minor party to have saved deposits in the electoral regions, one in the North Wales region and the other in the South Wales West region. In total the BNP polled 42,197 votes (4.3%).
In the 2011 Welsh Assembly elections, the BNP fielded 20 candidates, four in each of the five regional lists and for the first time 7 candidates were fielded in FPTP constituencies. On the regional lists, the BNP polled 22,610 votes (2.4%), down 1.9% from 2007. In 2 out of the 7 FPTP constituencies contested the BNP saved deposits: (Swansea East and Islwyn).
In the 2007 Scottish Parliament election, the party fielded 32 candidates, entitling it to public funding and an election broadcast, prompting criticism. The BNP received 24,616 votes (1.2%), no seats were won, nor were any deposits saved. In the 2011 Scottish Parliament election, the BNP fielded 32 candidates in the regional lists. 15,580 votes were polled (0.78%).
The BNP fielded 3 candidates for the first time in three constituencies each in the 2011 Northern Ireland Legislative Assembly elections (Belfast East, East Antrim and South Antrim). 1,252 votes were polled (0.2%), winning no seats for the party.
The BNP has taken part in European Parliament elections since 1999, when they received 1.13% of the total vote (102,647 votes).
In the 2004 elections to the European Parliament, the BNP won 4.9% of the vote, making it the sixth biggest party overall, but did not win any seats.
The BNP won two seats in the European Parliament in the 2009 elections. Andrew Brons was elected in the Yorkshire and the Humber regional constituency with 9.8% of the vote. Party chairman Nick Griffin was elected in the North West region, with 8% of the vote. Nationally, the BNP received 6.26%.
The British Government announced in 2009 that the BNP's two MEPs would be denied some of the access and information afforded to other MEPs. The BNP would be subject to the "same general principles governing official impartiality" and they would receive "standard written briefings as appropriate from time to time", but diplomats would not be "proactive" in dealing with the BNP MEPs and that any requests for policy briefings from them would be treated differently and on a discretionary basis.
The BNP did not stand any candidates in the 2019 European Parliament election in the United Kingdom.
The leaders and senior officers of the BNP have criminal convictions for inciting racial hatred.
John Hagan claims that the BNP has conducted right-wing extremist violence to gain "institutionalized power". Critics of the BNP, such as Human Rights Watch in a 1997 report, have asserted that the party recruits from skinhead groups and that it promotes racist violence.
In the past, Nick Griffin has defended the threat of violence to further the party's aims. After the BNP won its first council seat in 1993, he wrote that the BNP should not be a "postmodernist rightist party" but "a strong, disciplined organisation with the ability to back up its slogan 'Defend Rights for Whites' with well-directed boots and fists. When the crunch comes, power is the product of force and will, not of rational debate". In 1997 he said: "It is more important to control the streets of a city than its council chambers."
A BBC "Panorama" programme reported on a number of BNP members who have had criminal convictions, some racially motivated. Some of the more notable convictions include:
In 2011, Goodwin described the BNP as being "the most successful party in the history of the extreme right in Britain". That same year, John E. Richardson noted that it had achieved "a level of electoral success that is unparalleled in the history of British fascism". The historian Alan Sykes stated that "in electoral terms", the BNP achieved "more in the first three years of the twenty-first century" than the British far right "as a whole achieved in the previous seventy". However, Copsey noted that the party's belief that one day the conditions would be right for it to win a general election belonged to the "Never-Never Land of British politics". Copsey also noted that the BNP's electoral successes had been modest in comparison to those achieved by extreme-right groups elsewhere in Western Europe such as France's National Front, Italy's National Alliance, and Belgium's Vlaams Blok.
The BNP's growth met a hostile reaction, and in 2011 the political scientists Copsey and Macklin described it as "Britain's most disliked party". It was widely reviled as racist and even following Griffin's "modernisation" project it was still heavily tainted by its associations with neo-Nazism. For many years it remained closely associated with the National Front in the British public imagination.
The BNP remained unable to gain a broad appeal or widespread credibility. In a 2004 poll, seven out of ten voters said that they would never consider voting for the BNP. A 2009 poll found that two-thirds would "under no circumstances" consider voting BNP, while only 4% of respondents would "definitely consider" voting for them.
The Conservative leader Michael Howard stated that the BNP were a "stain" on British democracy, adding that "this is not a political movement, this is a bunch of thugs dressed up as a political party". His successor David Cameron described it as a "completely unacceptable" organisation which "thrives on hatred". The Labour prime minister, Tony Blair, called it a "nasty, extreme organisation", while the Liberal Democrat leader Nick Clegg termed it a "party of thugs and fascists". In 2004, the General Synod of the Church of England declared that supporting the BNP was incompatible with Christianity, comparing it to "spitting in the face of God". Christian groups throughout Britain have maintained that the BNP's hostility toward cultural and ethnic diversity in the country was at odds with mainstream Christianity's emphasis on inclusiveness, tolerance, and interfaith dialogue. Winston Churchill's family has criticised the BNP's use of his image and quotations, labelling it "offensive and disgusting". The singer Vera Lynn condemned the party for selling a CD featuring her recordings on its website. In 2009, the Royal British Legion asked Griffin—at first privately and then publicly—to not wear their poppy symbol.
The British police, Fire Brigades Union, and Church of England, prohibited its members from joining the BNP. In 2002, Martin Narey, banned BNP membership among prison workers; he subsequently received death threats. In 2010, the Education Secretary Michael Gove announced bans allowing headteachers to ban their staff from being party members.
Individuals whose membership of the party was made public sometimes faced ostracism and the loss of their job: examples include a school headmaster who had to resign, a caretaker who was sacked after attending a BNP rally, and a police officer dismissed from his position. After BNP membership lists were leaked on the Internet, a number of police forces investigated officers whose names appeared on the lists.
In 2005, an invitation to Nick Griffin by the University of St Andrews Union Debating Society to participate in a debate on multiculturalism was withdrawn after protests. The BNP says that National Union of Journalists guidelines on reporting "far right" organisations forbid unionised journalists from reporting uncritically on the party. In April 2007, an election broadcast was cancelled by BBC Radio Wales whose lawyers believed that the broadcast was defamatory of the Chief Constable of North Wales Police, Richard Brunstrom. The BNP claimed that BBC editors were following an agenda.
Attitudes toward the BNP in both mainstream broadcast media and print journalism have been overwhelmingly negative, and no mainstream newspaper has endorsed the party. This hostile coverage has even been found in right-wing tabloids like the "Daily Mail", "Daily Express" and "The Sun" which otherwise share the BNP's hostile attitude toward issues like immigration. In 2003, the "Daily Mail" described the BNP as "poisonous bigots", while in 2004 "The Sun" printed the headline of "BNP: Bloody Nasty People". Senior BNP figures nevertheless believed that these tabloids' hostile coverage of immigration and Islam helped to legitimise and normalise the party and its views among much of the British public, a view echoed by some academic observers. When, in 2004, anti-racist activists picketed outside the "Daily Mail" office in central London to protest against its negative coverage of asylum seekers, BNP members organised a counter-picket at which they displayed the placard "Vote BNP, Read the "Daily Mail"".
The BNP initially faced a 'no platform for fascists' policy from the broadcast media, although this eroded as Griffin was invited on to a number of television programmes amid the party's growing electoral success. When the BBC invited him to appear on "Question Time" in 2009 it was criticised by several trade unions, sections of the media, and several Labour politicians, all of whom believed that the BNP should not be given a public platform. Anti-fascist protesters assembled outside of the television studio to protest Griffin's inclusion.
The first academic attention to be directed at the BNP appeared after it gained a councillor in the 1993 local elections. Nevertheless, throughout the 1990s it remained the subject of little academic research. Academic interest increased following its victories at local elections from 2002 onward. The first detailed monograph study to be devoted to the party was Nigel Copsey's "Contemporary British Fascism", first published in 2004. In September 2008, an academic symposium on the BNP was held at Teesside University.
Opposition to the BNP also came from the organised anti-fascist movement. By the mid-1990s, the BNP's attempts to stage public events in Scotland, the North West and the Midlands were largely thwarted by the militant disruption of the Anti-Fascist Action (AFA) group. The BNP's modernisation and move away from street demonstrations and toward electoral campaigning caused problems for the AFA, who proved unable to successfully change their tactics; on those occasions when AFA activists tried to forcibly disrupt BNP activities, they were prevented and arrested by riot police.
More liberal sections of the anti-fascist movement sought to counter the BNP through community-based initiatives. "Searchlight" encouraged trade unions to establish localised campaigns that would ensure that ethnic minority and other anti-BNP locals voted. It suggested that such campaigns should avoid associating with the mainstream parties from which BNP voters felt disenfranchised and that they should not be afraid of calling out Islamic fundamentalists and extremists active in the area. The Unite Against Fascism group also sought to maximise anti-BNP turnout at elections, calling on the electorate to vote for "anyone but fascists". Evidence suggests that such anti-fascist activities did little to erode the far-right vote; this was in part because anti-fascist groups had encouraged the stereotype that BNP candidates were violent skinheads, something which conflicted with the more normal, friendly image that BNP activists cultivated when canvassing.
The BNP often received a hostile response from other sections of the British extreme-right. Some extreme-right-wingers, such as the British Freedom Party, expressed frustration at the party's inability to moderate itself further on the issue of race, while those such as Colin Jordan and the NF accused the BNP—particularly under Griffin's leadership—of being too moderate. This latter view was articulated by an extreme-right groupuscule, the International Third Position, when it claimed that the BNP "has been openly courting the Jewish vote and pumping out material which confirms what most us knew years ago: the BNP has become a multi-racist, Zionist, queer-tolerant anti-Muslim pressure group".
In "ASLEF v. United Kingdom", the European Court of Human Rights overturned an employment appeal tribunal ruling that awarded BNP member and train driver Jay Lee damages for expulsion from a trade union. In Redfearn v United Kingdom, the court ruled that members of racist organisations could lawfully be dismissed on health and safety grounds if there was a danger of violence occurring in the workplace. In November 2012, the European Court of Human Rights made a majority ruling (4 to 3) that in Redfearn's case against the UK government, his rights under Article 11 (free association) had been infringed, but not those under Article 10 (free expression) or Article 14 (discrimination). | https://en.wikipedia.org/wiki?curid=4294 |
Batavi (Germanic tribe)
The Batavi were an ancient Germanic tribe that lived around the modern Dutch Rhine delta in the area that the Romans called Batavia, from the second half of the first century BC to the third century AD. The name is also applied to several military units employed by the Romans that were originally raised among the Batavi. The tribal name, probably a derivation from "batawjō" ("good island", from Germanic "bat-" "good, excellent," which is also in the English "better," and "awjō" "island, land near water"), refers to the region's fertility, today known as the "fruitbasket of the Netherlands" (the Betuwe).
Finds of wooden tablets show that at least some were literate.
The Batavi themselves are not mentioned by Julius Caesar in his commentary "Commentarii de Bello Gallico", although he is often thought to have founded his dynasty's Germanic bodyguard, which was at least in later generations dominated by Batavi. But he did mention the "Batavian island" in the Rhine river. The island's easternmost point is at a split in the Rhine, one arm being the Waal the other the Lower Rhine/Old Rhine (hence the Latin name "Insula Batavorum", "Island of the Batavi"). Much later Tacitus wrote that they had originally been a tribe of the Chatti, a tribe in Germany also never mentioned by Caesar (unless they were his "Suebi"), who were forced by internal dissension to move to their new home. The time when this happened is unknown, but Caesar does describe forced movements of tribes from the east in his time, such as the Usipetes and Tencteri.
Tacitus also reports that before their arrival the area had been "an uninhabited district on the extremity of the coast of Gaul, and also of a neighbouring island, surrounded by the ocean in front, and by the river Rhine in the rear and on either side". This view, however, is contradicted by the archeological evidence, which shows continuous habitation from at least the third century BC onward.
The strategic position, to wit the high bank of the Waal offering an unimpeded view far into Germania Transrhenana (Germania Beyond the Rhine), was recognized first by Drusus, who built a massive fortress ("castra") and a headquarters ("praetorium") in imperial style. The latter was in use until the Batavian revolt.
Archeological evidence suggests they lived in small villages, composed of six to 12 houses in the very fertile lands between the rivers, and lived by agriculture and cattle-raising. Finds of horse skeletons in graves suggest a strong equestrian preoccupation. On the south bank of the Waal (in what is now Nijmegen) a Roman administrative center was built, called "Oppidum Batavorum". An Oppidum was a fortified warehouse, where a tribe's treasures were stored and guarded. This centre was razed during the Batavian Revolt. The Smetius Collection was instrumental in settling the debate about the exact location of the Batavians.
The first Batavi commander we know of is named Chariovalda, who led a charge across the Vīsurgis (Weser) river against the Cherusci led by Arminius during the campaigns of Germanicus in "Germania Transrhenana".
Tacitus ("De origine et situ Germanorum" XXIX) described the Batavi as the bravest of the tribes of the area, hardened in the Germanic wars, with cohorts under their own commanders transferred to Britannia. They retained the honour of the ancient association with the Romans, not required to pay tribute or taxes and used by the Romans only for war: "They furnished to the Empire nothing but men and arms", Tacitus remarked. Well regarded for their skills in horsemanship and swimming—for men and horses could cross the Rhine without losing formation, according to Tacitus. Dio Cassius describes this surprise tactic employed by Aulus Plautius against the "barbarians"—the British Celts— at the battle of the River Medway, 43:
The barbarians thought that Romans would not be able to cross it without a bridge, and consequently bivouacked in rather careless fashion on the opposite bank; but he sent across a detachment of Germanic tribesmen, who were accustomed to swim easily in full armour across the most turbulent streams. [...] Thence the Britons retired to the river Thames at a point near where it empties into the ocean and at flood-tide forms a lake. This they easily crossed because they knew where the firm ground and the easy passages in this region were to be found; but the Romans in attempting to follow them were not so successful. However, the Germans swam across again and some others got over by a bridge a little way up-stream, after which they assailed the barbarians from several sides at once and cut down many of them. (Cassius Dio, Roman History, Book 60:20)
It is uncertain how they were able to accomplish this feat. The late 4th century writer on Roman military affairs Vegetius mentions soldiers using reed rafts, drawn by leather leads, to transport equipment across rivers. But the sources suggest the Batavi were able to swim across rivers actually wearing full armour and weapons. This would only have been possible by the use of some kind of buoyancy device: Ammianus Marcellinus mentions that the "Cornuti" regiment swam across a river floating on their shields "as on a canoe" (357). Since the shields were wooden, they may have provided sufficient buoyancy
The Batavi were used to form the bulk of the Emperor's personal Germanic bodyguard from Augustus to Galba. They also provided a contingent for their indirect successors, the Emperor's horse guards, the "Equites singulares Augusti".
A Batavian contingent was used in an amphibious assault on Ynys Mon (Anglesey), taking the assembled Druids by surprise, as they were only expecting Roman ships.
Numerous altars and tombstones of the cohorts of Batavi, dating to the 2nd century and 3rd century, have been found along Hadrian's Wall, notably at Castlecary and Carrawburgh. As well as in Germany, Yugoslavia, Hungary, Romania and Austria.
Despite the alliance, one of the high-ranking Batavi, Julius Paullus, to give him his Roman name, was executed by Fonteius Capito on a false charge of rebellion. His kinsman Gaius Julius Civilis was paraded in chains in Rome before Nero; though he was acquitted by Galba, he was retained at Rome, and when he returned to his kin in the year of upheaval in the Roman Empire, 69, he headed a Batavian rebellion. He managed to capture Castra Vetera, the Romans lost two legions while two others (I Germanica and XVI Gallica) were controlled by the rebels. The rebellion became a real threat to the Empire when the conflict escalated to northern Gaul and Germania. The Roman army retaliated and invaded the "insula Batavorum". A bridge was built over the river Nabalia, where the warring parties approached each other on both sides to negotiate peace. The narrative was told in great detail in Tacitus' History, book iv, although, unfortunately, the narrative breaks off abruptly at the climax. Following the uprising, Legio X "Gemina" was housed in a stone "castra" to keep an eye on the Batavians.
The Batavi were still mentioned in 355 during the reign of Constantius II (317 - 361), when their island was already dominated by the Salii, a Frankish tribe that had sought Roman protection there in 297 after having been expelled from their own country by the Saxons.
Constantius Gallus added inhabitants of Batavia to his legions, "of whose discipline we still make use." It has been assumed they merged with the Salii shortly before or after and, after having been expelled by another tribe (it has been proposed this was the Chamavi), shared their subsequent migration to Toxandria. In the Late Roman army there was a unit called "Batavi".
The name of the Bavarian town of Passau descends from the Roman "Batavis", which was named after the Batavi. The town's name is old as it shows the typical effects of the High German consonant shift (b > p, t > ss).
In the 16th-century emergence of a popular foundation story and origin myth for the Dutch people, the Batavians came to be regarded as their ancestors during their national struggle for independence during the Eighty Years' War. The mix of fancy and fact in the "Cronyke van Hollandt, Zeelandt ende Vriesland" (called the "Divisiekroniek") by the Augustinian friar and humanist Cornelius Gerardi Aurelius, first published in 1517, brought the spare remarks in Tacitus' newly rediscovered "Germania" to a popular public; it was being reprinted as late as 1802. Contemporary Dutch virtues of independence, fortitude and industry were fully recognizable among the Batavians in more scholarly history represented in Hugo Grotius' "Liber de Antiquitate Republicae Batavicorum" (1610). The origin was perpetuated by Romeyn de Hooghe's "Spiegel van Staat der Vereenigden Nederlanden" ("Mirror of the State of the United Netherlands," 1706), which also ran to many editions, and it was revived in the atmosphere of Romantic nationalism in the late eighteenth-century reforms that saw a short-lived Batavian Republic and, in the colony of the Dutch East Indies, a capital that was named Batavia. Though since Indonesian independence the city is called Jakarta, its inhabitants up to the present still call themselves "Betawi" or "Orang Betawi", i.e. "People of Batavia" - a name ultimately derived from the ancient Batavians.
The success of this tale of origins was mostly due to resemblance in anthropology, which was based on tribal knowledge. Being politically and geographically inclusive, this historical vision filled the needs of Dutch nation-building and integration in the 1890-1914 era.
However, a disadvantage of this historical nationalism soon became apparent. It suggested there were no strong external borders, while allowing for the fairly clear-cut internal borders that were emerging as the society pillarized into three parts. After 1945 the tribal knowledge lost its grip on anthropology and mostly vanished. Modern variants of the Batavian founding myth are made more accurate by pointing out that the Batavians were one part of the ancestry of the Dutch people - together with the Frisians, Franks and Saxons - by tracing patterns of DNA. Echoes of this cultural continuity can still be found among various areas of Dutch modern culture, such as the very popular replica of the ship "Batavia" that can today be found in Lelystad. | https://en.wikipedia.org/wiki?curid=4295 |
Baptism
Baptism (from the Greek noun βάπτισμα "baptisma") is a Christian rite of admission and adoption, almost invariably with the use of water, into Christianity. It may be performed by sprinkling or pouring water on the head, or by immersing in water either partially or completely. The synoptic gospels recount that John the Baptist baptised Jesus. Baptism is considered a sacrament in most churches, and as an ordinance in others. Baptism is also called christening, although some reserve the word "christening" for the baptism of infants. It has also given its name to the Baptist churches and denominations.
Martyrdom was identified early in Church history as "baptism by blood", enabling the salvation of martyrs who had not been baptized by water. Later, the Catholic Church identified a baptism of desire, by which those preparing for baptism who die before actually receiving the sacrament are considered saved. Some Christians have regarded baptism as necessary for salvation, though others, such as Huldrych Zwingli (1484–1531), denied its necessity.
Quakers and the Salvation Army do not practice water baptism at all. Among denominations that practice water baptism, differences occur in the manner and mode of baptizing and in the understanding of the significance of the rite. Most Christians baptize using the trinitarian formula "in the name of the Father, and of the Son, and of the Holy Spirit" (following the Great Commission), but some baptize using Jesus' name only. Much more than half of all Christians baptize infants; many others regard only adult baptism as true baptism.
The term "baptism" has also been used metaphorically to refer to any ceremony, trial, or experience by which a person is initiated, purified, or given a name.
The English word "baptism" is derived indirectly through Latin from the neuter Greek concept noun "baptisma" (Greek , "washing-ism"), which is a neologism in the New Testament derived from the masculine Greek noun "baptismos" (), a term for ritual washing in Greek language texts of Hellenistic Judaism during the Second Temple period, such as the Septuagint. Both of these nouns are derived from the verb "baptizō" (, "I wash" transitive verb), which is used in Jewish texts for ritual washing, and in the New Testament both for ritual washing and also for the apparently new rite of "baptisma". The Greek verb "baptō" (), "dip", from which the verb "baptizo" is derived, is in turn hypothetically traced to a reconstructed Indo-European root *"gʷabh-", "dip". The Greek words are used in a great variety of meanings. and in Hellenism had the general usage of "immersion," "going under" (as a material in a liquid dye) or "perishing" (as in a ship sinking or a person drowning), with the same double meanings as in English "to sink into" or "to be overwhelmed by," with bathing or washing only occasionally used and usually in sacral contexts.
Baptism has similarities to Tvilah, a Jewish purification ritual of immersing in water, which is required for, among other things, conversion to Judaism, but which differs in being repeatable, while baptism is to be performed only once. (In fact, the Modern Hebrew term for "baptism" is "Christian Tvilah".) John the Baptist, who is considered a forerunner to Christianity, used baptism as the central sacrament of his messianic movement. The apostle Paul distinguished between the baptism of John, ("baptism of repentance") and baptism in the name of Jesus, and it is questionable whether Christian baptism was in some way linked with that of John. Christians consider Jesus to have instituted the sacrament of baptism.
Though some form of immersion was likely the most common method of baptism in the early church, many of the writings from the ancient church appeared to view this mode of baptism as inconsequential. The Didache 7.1–3 (AD 60–150) allowed for affusion practices in situations where immersion was not practical. Likewise, Tertullian (AD 196–212) allowed for varying approaches to baptism even if those practices did not conform to biblical or traditional mandates (cf. De corona militis 3; De baptismo 17). Finally, Cyprian (ca. AD 256) explicitly stated that the amount of water was inconsequential and defended immersion, affusion, and aspersion practices (Epistle 75.12). As a result, there was no uniform or consistent mode of baptism in the ancient church prior to the fourth century.
By the third and fourth centuries, baptism involved catechetical instruction as well as chrismation, exorcisms, laying on of hands, and recitation of a creed.
In the early middle ages infant baptism became common and the rite was significantly simplified. In Western Europe Affusion became the normal mode of baptism between the twelfth and fourteenth centuries, though immersion was still practiced into the sixteenth. In the medieval period, some radical Christians rejected the sacrament of baptism. Sects such as the Tondrakians, Cathars, Arnoldists, Petrobrusians, Henricans, Brethren of the Free Spirit and the Lollards were regarded as heretics by the Catholic Church. In the sixteenth century, Martin Luther retained baptism as a sacrament, but Swiss reformer Huldrych Zwingli considered baptism and the Lord's supper to be symbolic. Anabaptists denied the validity of the practice of infant baptism, and rebaptized converts.
Baptism is practiced in several different ways. Aspersion is the sprinkling of water on the head, and affusion is the pouring of water over the head.
The word "immersion" is derived from late Latin "immersio", a noun derived from the verb "immergere" ("in" – "into" + "mergere" "dip"). In relation to baptism, some use it to refer to any form of dipping, whether the body is put completely under water or is only partly dipped in water; they thus speak of immersion as being either total or partial. Others, of the Anabaptist belief, use "immersion" to mean exclusively plunging someone entirely under the surface of the water. The term "immersion" is also used of a form of baptism in which water is poured over someone standing in water, without submersion of the person. On these three meanings of the word "immersion", see Immersion baptism.
When "immersion" is used in opposition to "submersion", it indicates the form of baptism in which the candidate stands or kneels in water and water is poured over the upper part of the body. Immersion in this sense has been employed in West and East since at least the 2nd century and is the form in which baptism is generally depicted in early Christian art. In the West, this method of baptism began to be replaced by affusion baptism from around the 8th century, but it continues in use in Eastern Christianity.
The word submersion comes from the late Latin ("sub-" "under, below" + "mergere" "plunge, dip") and is also sometimes called "complete immersion". It is the form of baptism in which the water completely covers the candidate's body. Submersion is practiced in the Orthodox and several other Eastern Churches. In the Latin Church of the Catholic Church, baptism by submersion is used in the Ambrosian Rite and is one of the methods provided in the Roman Rite of the baptism of infants. It is seen as obligatory among some groups that have arisen since the Protestant Reformation, such as Baptists.
The Greek-English Lexicon of Liddell and Scott gives the primary meaning of the verb "baptizein", from which the English verb "baptize" is derived, as "dip, plunge", and gives examples of plunging a sword into a throat or an embryo and for drawing wine by dipping a cup in the bowl; for New Testament usage it gives two meanings: "baptize", with which it associates the Septuagint mention of Naaman dipping himself in the Jordan River, and "perform ablutions", as in Luke 11:38.
Although the Greek verb "baptizein" does not exclusively mean dip, plunge or immerse (it is used with literal and figurative meanings such as "sink", "disable", "overwhelm", "go under", "overborne", "draw from a bowl"), lexical sources typically cite this as a meaning of the word in both the Septuagint and the New Testament.
"While it is true that the basic root meaning of the Greek words for baptize and baptism is immerse/immersion, it is not true that the words can simply be reduced to this meaning, as can be seen from Mark 10:38–39, Luke 12:50, Matthew 3:11 Luke 3:16 and Corinthians10:2."
Two passages in the Gospels indicate that the verb "baptizein" did not always indicate submersion. The first is Luke 11:38, which tells how a Pharisee, at whose house Jesus ate, "was astonished to see that he did not first wash ("ἐβαπτίσθη", aorist passive of "βαπτίζω"—literally, "was baptized") before dinner". This is the passage that Liddell and Scott cites as an instance of the use of ' to mean "perform ablutions". Jesus' omission of this action is similar to that of his disciples: "Then came to Jesus scribes and Pharisees, which were of Jerusalem, saying, Why do thy disciples transgress the tradition of the elders? for they wash (') not their hands when they eat bread". The other Gospel passage pointed to is: "The Pharisees...do not eat unless they wash ("", the ordinary word for washing) their hands thoroughly, observing the tradition of the elders; and when they come from the market place, they do not eat unless they wash themselves (literally, "baptize themselves"—"βαπτίσωνται", passive or middle voice of "βαπτίζω")".
Scholars of various denominations claim that these two passages show that invited guests, or people returning from market, would not be expected to immerse themselves ("baptize themselves") totally in water but only to practise the partial immersion of dipping their hands in water or to pour water over them, as is the only form admitted by present Jewish custom. In the second of the two passages, it is actually the hands that are specifically identified as "washed" (Mark 7:3), not the entire person, for whom the verb used is "baptizomai", literally "be baptized", "be immersed" (Mark 7:4), a fact obscured by English versions that use "wash" as a translation of both verbs. Zodhiates concludes that the washing of the hands was done by immersing them. The Liddell–Scott–Jones Greek-English Lexicon (1996) cites the other passage (Luke 11:38) as an instance of the use of the verb "baptizein" to mean "perform ablutions", not "submerge". References to the cleaning of vessels which use βαπτίζω also refer to immersion.
As already mentioned, the lexicographical work of Zodhiates says that, in the second of these two cases, the verb "baptizein" indicates that, after coming from the market, the Pharisees washed their hands by immersing them in collected water. Balz & Schneider understand the meaning of βαπτίζω, used in place of ῥαντίσωνται (sprinkle), to be the same as βάπτω, to dip or immerse, a verb used of the partial dipping of a morsel held in the hand into wine or of a finger into spilled blood.
A possible additional use of the verb "baptizein" to relate to ritual washing is suggested by Peter Leithart (2007) who suggests that Paul's phrase "Else what shall they do who are baptized for the dead?" relates to Jewish ritual washing. In Jewish Greek the verb "baptizein" "baptized" has a wider reference than just "baptism" and in Jewish context primarily applies to the masculine noun "baptismos" "ritual washing" The verb "baptizein" occurs four times in the Septuagint in the context of ritual washing, "baptismos"; Judith cleansing herself from menstrual impurity, Naaman washing seven times to be cleansed from leprosy, etc. Additionally, in the New Testament only, the verb "baptizein" can also relate to the neuter noun "baptisma" "baptism" which is a neologism unknown in the Septuagint and other pre-Christian Jewish texts. This broadness in the meaning of "baptizein" is reflected in English Bibles rendering "wash", where Jewish ritual washing is meant: for example Mark 7:4 states that the Pharisees "except they wash (Greek "baptize"), they do not eat", and "baptize" where "baptisma", the new Christian rite, is intended.
Two nouns derived from the verb "baptizo" (βαπτίζω) appear in the New Testament: the masculine noun "baptismos" (βαπτισμός) and the neuter noun "baptisma" (βάπτισμα):
Until the Middle Ages, most baptisms were performed with the candidates naked—as is evidenced by most of the early portrayals of baptism (some of which are shown in this article), and the early Church Fathers and other Christian writers. Deaconesses helped female candidates for reasons of modesty.
Typical of these is Cyril of Jerusalem who wrote "On the Mysteries of Baptism" in the 4th century (c. 350 AD):
The symbolism is threefold:
1. Baptism is considered to be a form of rebirth—"by water and the Spirit"—the nakedness of baptism (the second birth) paralleled the condition of one's original birth. For example, St. John Chrysostom calls the baptism "λοχείαν", i.e., giving birth, and "new way of creation...from water and Spirit" ("to John" speech 25,2), and later elaborates:
2. The removal of clothing represented the "image of putting off the old man with his deeds" (as per Cyril, above), so the stripping of the body before for baptism represented taking off the trappings of sinful self, so that the "new man", which is given by Jesus, can be put on.
3. As St. Cyril again asserts above, as Adam and Eve in scripture were naked, innocent and unashamed in the Garden of Eden, nakedness during baptism was seen as a renewal of that innocence and state of original sinlessness. Other parallels can also be drawn, such as between the exposed condition of Christ during His crucifixion, and the crucifixion of the "old man" of the repentant sinner in preparation for baptism.
Changing customs and concerns regarding modesty probably contributed to the practice of permitting or requiring the baptismal candidate to either retain their undergarments (as in many Renaissance paintings of baptism such as those by da Vinci, Tintoretto, Van Scorel, Masaccio, de Wit and others) or to wear, as is almost universally the practice today, baptismal robes. These robes are most often white, symbolizing purity. Some groups today allow any suitable clothes to be worn, such as trousers and a T-shirt—practical considerations include how easily the clothes will dry (denim is discouraged), and whether they will become see-through when wet.
There are differences in views about the effect of baptism for a Christian. Some Christian groups assert baptism is a requirement for salvation and a sacrament, and speak of "baptismal regeneration". Its importance is related to their interpretation of the meaning of the "Mystical Body of Christ" as found in the New Testament. This view is shared by the Catholic and Eastern Orthodox denominations, and by Churches formed early during the Protestant Reformation such as Lutheran and Anglican. For example, Martin Luther said:
The Churches of Christ," Jehovah's Witnesses, Christadelphians, and The Church of Jesus Christ of Latter-day Saints also espouse baptism as necessary for salvation.
For Roman Catholics, baptism by water is a sacrament of initiation into the life of the children of God ("Catechism of the Catholic Church", 1212–13). It configures the person to Christ (CCC 1272), and obliges the Christian to share in the Church's apostolic and missionary activity (CCC 1270). The Catholic holds that there are three types of baptism by which one can be saved: sacramental baptism (with water), baptism of desire (explicit or implicit desire to be part of the Church founded by Jesus Christ), and baptism of blood (martyrdom). In his encyclical "Mystici corporis Christi" of June 29, 1943, Pope Pius XII spoke of baptism and profession of the true faith as what makes members of the one true Church, which is the body of Jesus Christ himself, as God the Holy Spirit has taught through the Apostle Paul:
By contrast, Anabaptist and Evangelical Protestants recognize baptism as an outward sign of an inward reality following on an individual believer's experience of forgiving grace. Reformed and Methodist Protestants maintain a link between baptism and regeneration, but insist that it is not automatic or mechanical, and that regeneration may occur at a different time than baptism. Churches of Christ consistently teach that in baptism a believer surrenders his life in faith and obedience to God, and that God "by the merits of Christ's blood, cleanses one from sin and truly changes the state of the person from an alien to a citizen of God's kingdom. Baptism is not a human work; it is the place where God does the work that only God can do." Thus, they see baptism as a passive act of faith rather than a meritorious work; it "is a confession that a person has nothing to offer God".
The liturgy of baptism for Catholics, Eastern Orthodox, Lutheran, Anglican, and Methodist makes clear reference to baptism as not only a symbolic burial and resurrection, but an actual supernatural transformation, one that draws parallels to the experience of Noah and the passage of the Israelites through the Red Sea divided by Moses. Thus, baptism is literally and symbolically not only cleansing, but also dying and rising again with Christ. Catholics believe baptism is necessary to cleanse the taint of original sin, and so commonly baptise infants.
The Eastern Churches (Eastern Orthodox Church and Oriental Orthodoxy) also baptize infants on the basis of texts, such as Matthew 19:14, which are interpreted as supporting full Church membership for children. In these denominations, baptism is immediately followed by Chrismation and Communion at the next Divine Liturgy, regardless of age. Orthodox likewise believe that baptism removes what they call the ancestral sin of Adam. Anglicans believe that Baptism is also the entry into the Church and therefore allows them access to all rights and responsibilities as full members, including the privilege to receive Holy Communion. Most Methodists and Anglicans agree that it also cleanses the taint of what in the West is called original sin, in the East ancestral sin.
Eastern Orthodox Christians usually insist on complete threefold immersion as both a symbol of death and rebirth into Christ, and as a washing away of sin. Latin Church Catholics generally baptize by affusion (pouring); Eastern Catholics usually by submersion, or at least partial immersion. However, submersion is gaining in popularity within the Latin Catholic Church. In newer church sanctuaries, the baptismal font may be designed to expressly allow for baptism by immersion. Anglicans baptize by immersion or affusion.
According to evidence which can be traced back to about the year 200, sponsors or godparents are present at baptism and vow to uphold the Christian education and life of the baptized.
Baptists argue that the Greek word originally meant "to immerse". They interpret some Biblical passages concerning baptism as requiring submersion of the body in water. They also state that only submersion reflects the symbolic significance of being "buried" and "raised" with Christ. Baptist Churches baptize in the name of the Trinity—the Father, the Son, and the Holy Spirit. However, they do not believe that baptism is necessary for salvation; but rather that it is an act of Christian obedience.
Some "Full Gospel" charismatic churches such as Oneness Pentecostals baptize only in the name of Jesus Christ, citing Peter's preaching baptism in the name of Jesus as their authority.
In 1982 the World Council of Churches published the ecumenical paper "Baptism, Eucharist and Ministry". The preface of the document states:
A 1997 document, "Becoming a Christian: The Ecumenical Implications of Our Common Baptism", gave the views of a commission of experts brought together under the aegis of the World Council of Churches. It states:
Those who heard, who were baptized and entered the community's life, were already made witnesses of and partakers in the promises of God for the last days: the forgiveness of sins through baptism in the name of Jesus and the outpouring of the Holy Ghost on all flesh. Similarly, in what may well be a baptismal pattern, 1 Peter testifies that proclamation of the resurrection of Jesus Christ and teaching about new life lead to purification and new birth. This, in turn, is followed by eating and drinking God's food, by participation in the life of the community—the royal priesthood, the new temple, the people of God—and by further moral formation. At the beginning of 1 Peter the writer sets this baptism in the context of obedience to Christ and sanctification by the Spirit. So baptism into Christ is seen as baptism into the Spirit. In the fourth gospel Jesus' discourse with Nicodemus indicates that birth by water and Spirit becomes the gracious means of entry into the place where God rules.
The vast majority of Christian denominations admit the theological idea that baptism is a sacrament, that has actual spiritual, holy and salvific effects. Certain key criteria must be complied with for it to be valid, i.e., to actually have those effects. If these key criteria are met, violation of some rules regarding baptism, such as varying the authorized rite for the ceremony, renders the baptism illicit (contrary to the church's laws) but still valid.
One of the criteria for validity is use of the correct form of words. The Roman Catholic Church teaches that the use of the verb "baptize" is essential. Catholics of the Latin Church, Anglicans and Methodists use the form "I baptize you..." Eastern Orthodox and some Eastern Catholics use a passive voice form "The Servant/(Handmaiden) of God is baptized in the name of..." or "This person is baptized by my hands..."
Use of the Trinitarian formula "in the name of the Father, and of the Son, and of the Holy Spirit" is also considered essential; thus these churches do not accept as valid baptisms of non-Trinitarian churches such as Oneness Pentecostals.
Another essential condition is use of water. A baptism in which some liquid that would not usually be called water, such as wine, milk, soup or fruit juice was used would not be considered valid.
Another requirement is that the celebrant intends to perform baptism. This requirement entails merely the intention "to do what the Church does", not necessarily to have Christian faith, since it is not the person baptizing, but the Holy Spirit working through the sacrament, who produces the effects of the sacrament. Doubt about the faith of the baptizer is thus no ground for doubt about the validity of the baptism.
Some conditions expressly do not affect validity—for example, whether submersion, immersion, affusion or aspersion is used. However, if water is sprinkled, there is a danger that the water may not touch the skin of the unbaptized. As has been stated, "it is not sufficient for the water to merely touch the candidate; it must also flow, otherwise there would seem to be no real ablution. At best, such a baptism would be considered doubtful. If the water touches only the hair, the sacrament has probably been validly conferred, though in practice the safer course must be followed. If only the clothes of the person have received the aspersion, the baptism is undoubtedly void." For many communions, validity is not affected if a single submersion or pouring is performed rather than a triple, but in Orthodoxy this is controversial.
According to the Catholic Church, baptism imparts an indelible "seal" upon the soul of the baptized and therefore a person who has already been baptized cannot be validly baptized again. This teaching was affirmed against the Donatists who practiced rebaptism. The grace received in baptism is believed to operate "ex opere operato" and is therefore considered valid even if administered in heretical or schismatic groups.
The Catholic, Lutheran, Anglican, Presbyterian and Methodist Churches accept baptism performed by other denominations within this group as valid, subject to certain conditions, including the use of the Trinitarian formula. It is only possible to be baptized once, thus people with valid baptisms from other denominations may not be baptized again upon conversion or transfer. For Roman Catholics, this is affirmed in the Canon Law 864, in which it is written that "[e]very person not yet baptized and only such a person is capable of baptism." Such people are accepted upon making a profession of faith and, if they have not yet validly received the sacrament/rite of confirmation or chrismation, by being confirmed. Specifically, "Methodist theologians argued that since God never abrogated a covenant made and sealed with proper intentionality, rebaptism was never an option, unless the original baptism had been defective by not having been made in the name of the Trinity." In some cases it can be difficult to decide if the original baptism was in fact valid; if there is doubt, conditional baptism is administered, with a formula on the lines of "If you are not yet baptized, I baptize you..."
In the still recent past, it was common practice in the Roman Catholic Church to baptize conditionally almost every convert from Protestantism because of a perceived difficulty in judging about the validity in any concrete case. In the case of the major Protestant Churches, agreements involving assurances about the manner in which they administer baptism has ended this practice, which sometimes continues for other groups of Protestants. The Catholic Church has always recognized the validity of baptism in the Churches of Eastern Christianity, but it has explicitly denied the validity of the baptism conferred in The Church of Jesus Christ of Latter-day Saints.
Practice in the Eastern Orthodox Church for converts from other communions is not uniform. However, generally baptisms performed in the name of the Holy Trinity are accepted by the Orthodox Christian Church. If a convert has not received the sacrament (mysterion) of baptism, he or she must be baptised in the name of the Holy Trinity before they may enter into communion with the Orthodox Church. If he has been baptized in another Christian confession (other than Orthodox Christianity) his previous baptism is considered retroactively filled with grace by chrismation or, in rare circumstances, confession of faith alone as long as the baptism was done in the name of the Holy Trinity (Father, Son and Holy Spirit). The exact procedure is dependent on local canons and is the subject of some controversy.
Oriental Orthodox Churches recognise the validity of baptisms performed within the Eastern Orthodox Communion. Some also recognise baptisms performed by Catholic Churches. Any supposed baptism not performed using the Trinitarian formula is considered invalid.
In the eyes of the Catholic Church, all Orthodox Churches, Anglican and Lutheran Churches, the baptism conferred by The Church of Jesus Christ of Latter-day Saints is invalid. An article published together with the official declaration to that effect gave reasons for that judgment, summed up in the following words: "The Baptism of the Catholic Church and that of the Church of Jesus Christ of Latter-day Saints differ essentially, both for what concerns faith in the Father, Son and Holy Spirit, in whose name Baptism is conferred, and for what concerns the relationship to Christ who instituted it."
The Church of Jesus Christ of Latter-day Saints stresses that baptism must be administered by one having proper authority; consequently, the church does not recognize the baptism of any other church as valid.
Jehovah's Witnesses do not recognise any other baptism occurring after 1914 as valid, as they believe that they are now the one true church of Christ, and that the rest of "Christendom" is false religion.
There is debate among Christian churches as to who can administer baptism. Some claim that the examples given in the New Testament only show apostles and deacons administering baptism. Ancient Christian churches interpret this as indicating that baptism should be performed by the clergy except "in extremis", i.e., when the one being baptized is in immediate danger of death. Then anyone may baptize, provided, in the view of the Eastern Orthodox Church, the person who does the baptizing is a member of that Church, or, in the view of the Catholic Church, that the person, even if not baptized, intends to do what the Church does in administering the rite. Many Protestant churches see no specific prohibition in the biblical examples and permit any believer to baptize another.
In the Roman Catholic Church, canon law for the Latin Church lays down that the ordinary minister of baptism is a bishop, priest or deacon, but its administration is one of the functions "especially entrusted to the parish priest". If the person to be baptized is at least fourteen years old, that person's baptism is to be referred to the bishop, so that he can decide whether to confer the baptism himself. If no ordinary minister is available, a catechist or some other person whom the local ordinary has appointed for this purpose may licitly do the baptism; indeed in a case of necessity "any" person (irrespective of that person's religion) who has the requisite intention may confer the baptism By "a case of necessity" is meant imminent danger of death because of either illness or an external threat. "The requisite intention" is, at the minimum level, the intention "to do what the Church does" through the rite of baptism.
In the Eastern Catholic Churches, a deacon is not considered an ordinary minister. Administration of the sacrament is reserved to the Parish Priest or to another priest to whom he or the local hierarch grants permission, a permission that can be presumed if in accordance with canon law. However, "in case of necessity, baptism can be administered by a deacon or, in his absence or if he is impeded, by another cleric, a member of an institute of consecrated life, or by any other Christian faithful; even by the mother or father, if another person is not available who knows how to baptize."
The discipline of the Eastern Orthodox Church, Oriental Orthodoxy and the Assyrian Church of the East is similar to that of the Eastern Catholic Churches. They require the baptizer, even in cases of necessity, to be of their own faith, on the grounds that a person cannot convey what he himself does not possess, in this case membership in the Church. The Latin Catholic Church does not insist on this condition, considering that the effect of the sacrament, such as membership of the Church, is not produced by the person who baptizes, but by the Holy Spirit. For the Orthodox, while Baptism "in extremis" may be administered by a deacon or any lay-person, if the newly baptized person survives, a priest must still perform the other prayers of the Rite of Baptism, and administer the Mystery of Chrismation.
The discipline of Anglicanism and Lutheranism is similar to that of the Latin Catholic Church. For Methodists and many other Protestant denominations, too, the ordinary minister of baptism is a duly ordained or appointed minister of religion.
Newer movements of Protestant Evangelical churches, particularly non-denominational, allow laypeople to baptize.
In The Church of Jesus Christ of Latter-day Saints, only a man who has been ordained to the Aaronic priesthood holding the priesthood office of priest or higher office in the Melchizedek priesthood may administer baptism.
A Jehovah's Witnesses baptism is performed by a "dedicated male" adherent. Only in extraordinary circumstances would a "dedicated" baptizer be unbaptized (see section "Jehovah's Witnesses").
Anabaptists and Baptists recognize only believer's baptism or "adult baptism". Baptism is seen as an act identifying one as having accepted Jesus Christ as savior.
Early Anabaptists were given that name because they re-baptized persons who they felt had not been properly baptized, having received infant baptism, sprinkling.
The traditional form of Anabaptist baptism was pouring or sprinkling, the form commonly used in the West in the early 16th century when they emerged. Since the 18th century immersion and submersion became more widespread. Today all forms of baptism can be found among Anabaptist.
Baptism memorializes the death, burial and resurrection of Jesus. It is considered a covenantal act, signifying entrance into the New Covenant of Christ.
For the majority of Baptists, Christian baptism is the immersion of a believer in water in the name of the Father, the Son, and the Holy Spirit. It is an act of obedience symbolizing the believer's faith in a crucified, buried, and risen Saviour, the believer's death to sin, the burial of the old life, and the resurrection to walk in newness of life in Christ Jesus. It is a testimony to the believer's faith in the final resurrection of the dead. Baptism does not accomplish anything in itself, but is an outward personal sign that the person's sins have already been washed away by the blood of Christ's cross.
For a new convert the general practice is that baptism also allows the person to be a registered member of the local Baptist congregation (though some churches have adopted "new members classes" as a mandatory step for congregational membership).
Regarding rebaptism the general rules are:
Baptism in Churches of Christ is performed only by full bodily immersion, based on the Koine Greek verb "baptizo" which means to dip, immerse, submerge or plunge. Submersion is seen as more closely conforming to the death, burial and resurrection of Jesus than other modes of baptism. Churches of Christ argue that historically immersion was the mode used in the 1st century, and that pouring and sprinkling later emerged as secondary modes when immersion was not possible. Over time these secondary modes came to replace immersion. Only those mentally capable of belief and repentance are baptized (i.e., infant baptism is not practiced because the New Testament has no precedent for it).
Churches of Christ have historically had the most conservative position on baptism among the various branches of the Restoration Movement, understanding baptism by immersion to be a necessary part of conversion. The most significant disagreements concerned the extent to which a correct understanding of the role of baptism is necessary for its validity. David Lipscomb insisted that if a believer was baptized out of a desire to obey God, the baptism was valid, even if the individual did not fully understand the role baptism plays in salvation. Austin McGary contended that to be valid, the convert must also understand that baptism is for the forgiveness of sins. McGary's view became the prevailing one in the early 20th century, but the approach advocated by Lipscomb never totally disappeared. As such, the general practice among churches of Christ is to require rebaptism by immersion of converts, even those who were previously baptized by immersion in other churches.
More recently, the rise of the International Churches of Christ has caused some to reexamine the issue.
Churches of Christ consistently teach that in baptism a believer surrenders his life in faith and obedience to God, and that God "by the merits of Christ's blood, cleanses one from sin and truly changes the state of the person from an alien to a citizen of God's kingdom. Baptism is not a human work; it is the place where God does the work that only God can do." Baptism is a passive act of faith rather than a meritorious work; it "is a confession that a person has nothing to offer God." While Churches of Christ do not describe baptism as a "sacrament", their view of it can legitimately be described as "sacramental." They see the power of baptism coming from God, who chose to use baptism as a vehicle, rather than from the water or the act itself, and understand baptism to be an integral part of the conversion process, rather than just a symbol of conversion. A recent trend is to emphasize the transformational aspect of baptism: instead of describing it as just a legal requirement or sign of something that happened in the past, it is seen as "the event that places the believer 'into Christ' where God does the ongoing work of transformation." There is a minority that downplays the importance of baptism in order to avoid sectarianism, but the broader trend is to "reexamine the richness of the biblical teaching of baptism and to reinforce its central and essential place in Christianity."
Because of the belief that baptism is a necessary part of salvation, some Baptists hold that the Churches of Christ endorse the doctrine of baptismal regeneration. However, members of the Churches of Christ reject this, arguing that since faith and repentance are necessary, and that the cleansing of sins is by the blood of Christ through the grace of God, baptism is not an inherently redeeming ritual. Rather, their inclination is to point to the biblical passage in which Peter, analogizing baptism to Noah's flood, posits that "likewise baptism doth also now save us" but parenthetically clarifies that baptism is ""not" the putting away of the filth of the flesh but the "response of a good conscience" toward God" (1 Peter 3:21). One author from the churches of Christ describes the relationship between faith and baptism this way, ""Faith" is the "reason why" a person is a child of God; "baptism" is the "time at which" one is incorporated into Christ and so becomes a child of God" (italics are in the source). Baptism is understood as a confessional expression of faith and repentance, rather than a "work" that earns salvation.
The Methodist Articles of Religion, with regard to baptism, teach:
While baptism imparts regenerating grace, its permanence is contingent upon repentance and a personal commitment to Jesus Christ.
In the Methodist Churches, baptism is a sacrament of initiation into the visible Church. Wesleyan covenant theology further teaches that baptism is a sign and a seal of the covenant of grace:
Methodists recognize three modes of baptism as being valid—"immersion, sprinkling, or pouring" in the name of the Holy Trinity.
In Reformed baptismal theology, baptism is seen as primarily God's offer of union with Christ and all his benefits to the baptized. This offer is believed to be intact even when it is not received in faith by the person baptized. Reformed theologians believe the Holy Spirit brings into effect the promises signified in baptism. Baptism is held by almost the entire Reformed tradition to effect regeneration, even in infants who are incapable of faith, by effecting faith which would come to fruition later. Baptism also initiates one into the visible church and the covenant of grace. Baptism is seen as a replacement of circumcision, which is considered the rite of initiation into the covenant of grace in the Old Testament.
Reformed Christians believe that immersion is not necessary for baptism to be properly performed, but that pouring or sprinkling are acceptable. Only ordained ministers are permitted to administer baptism in Reformed churches, with no allowance for emergency baptism, though baptisms performed by non-ministers are generally considered valid. Reformed churches, while rejecting the baptismal ceremonies of the Roman Catholic church, accept the validity of baptisms performed with them and do not rebaptize.
In Catholic teaching, baptism is stated to be "necessary for salvation by actual reception or at least by desire". Catholic doctrine holds that the baptism ceremony is ordinarily performed by deacons, priests, or bishops, but in an emergency can be performed by any Catholic. This teaching is based on the Gospel according to John which says that Jesus proclaimed: "Truly, truly, I say to you, unless one is born of water and the Spirit, he cannot enter into the Kingdom of God." It dates back to the teachings and practices of 1st-century Christians, and the connection between salvation and baptism was not, on the whole, an item of major dispute until Huldrych Zwingli denied the necessity of baptism, which he saw as merely a sign granting admission to the Christian community. The Catechism of the Catholic Church states that "Baptism is necessary for salvation for those to whom the Gospel has been proclaimed and who have had the possibility of asking for this sacrament." The Council of Trent also states in the "Decree Concerning Justification" from session six that baptism is necessary for salvation. A person who knowingly, willfully and unrepentantly rejects baptism has no hope of salvation. However, if knowledge is absent, "those also can attain to salvation who through no fault of their own do not know the Gospel of Christ or His Church, yet sincerely seek God and moved by grace strive by their deeds to do His will as it is known to them through the dictates of conscience."
The Catechism of the Catholic Church also states: "Since Baptism signifies liberation from sin and from its instigator the devil, one or more exorcisms are pronounced over the candidate". In the Roman Rite of the baptism of a child, the wording of the prayer of exorcism is: "Almighty and ever-living God, you sent your only Son into the world to cast out the power of Satan, spirit of evil, to rescue man from the kingdom of darkness and bring him into the splendour of your kingdom of light. We pray for this child: set him (her) free from original sin, make him (her) a temple of your glory, and send your Holy Spirit to dwell with him (her). Through Christ our Lord."
In the Catholic Church by baptism all sins are forgiven, original sin and all personal sins. Given once for all, baptism cannot be repeated. Baptism not only purifies from all sins, but also makes the neophyte "a new creature," an adopted son of God, who has become a "partaker of the divine nature," member of Christ and co-heir with him, and a temple of the Holy Spirit. Sanctifying grace, the grace of justification, given by God by baptism, erases the original sin and personal actual sins.
Catholics are baptized in water, by submersion, immersion or affusion, or aspersion (sprinkling), in the name (singular) of the Father and of the Son and of the Holy Spirit—not three gods, but one God subsisting in three Persons. While sharing in the one divine essence, the Father, Son, and Holy Spirit are distinct, not simply three "masks" or manifestations of one divine being. The faith of the Church and of the individual Christian is based on a relationship with these three "Persons" of the one God. Adults can also be baptized through the Rite of Christian Initiation of Adults.
It is claimed that Pope Stephen I, St. Ambrose and Pope Nicholas I declared that baptisms in the name of "Jesus" only as well as in the name of "Father, Son and Holy Spirit" were valid. The correct interpretation of their words is disputed. Current canonical law requires the Trinitarian formula and water for validity.
The Church recognizes two equivalents of baptism with water: "baptism of blood" and "baptism of desire". Baptism of blood is that undergone by unbaptized individuals who are martyred for their faith, while baptism of desire generally applies to catechumens who die before they can be baptized. The Catechism of the Catholic Church describes these two forms:
The Church has always held the firm conviction that those who suffer death for the sake of the faith without having received Baptism are baptized by their death for and with Christ. This "Baptism of blood", like the desire for Baptism, brings about the fruits of Baptism without being a sacrament. (1258)
For catechumens who die before their Baptism, their explicit desire to receive it, together with repentance for their sins, and charity, assures them the salvation that they were not able to receive through the sacrament. (1259)
The Catholic Church holds that those who are ignorant of Christ's Gospel and of the Church, but who seek the truth and do God's will as they understand it, may be supposed to have an implicit desire for baptism and can be saved: "'Since Christ died for all, and since all men are in fact called to one and the same destiny, which is divine, we must hold that the Holy Spirit offers to all the possibility of being made partakers, in a way known to God, of the Paschal mystery.' Every man who is ignorant of the Gospel of Christ and of his Church, but seeks the truth and does the will of God in accordance with his understanding of it, can be saved. It may be supposed that such persons would have desired Baptism explicitly if they had known its necessity." As for unbaptized infants, the Church is unsure of their fate; "the Church can only entrust them to the mercy of God".
In Eastern Orthodoxy, baptism is considered a sacrament and mystery which transforms the old and sinful person into a new and pure one, where the old life, the sins, any mistakes made are gone and a clean slate is given. In Greek and Russian Orthodox traditions, it is taught that through Baptism a person is united to the Body of Christ by becoming an official member of the Orthodox Church. During the service, the Orthodox priest blesses the water to be used . The catechumen (the one baptised) is fully immersed in the water three times in the name of the Trinity. This is considered to be a death of the "old man" by participation in the crucifixion and burial of Christ, and a rebirth into new life in Christ by participation in his resurrection. Properly a new name is given, which becomes the person's name.
Babies of Orthodox families are normally baptized shortly after birth. Older converts to Orthodoxy are usually formally baptized into the Orthodox Church, though exceptions are sometimes made. Those who have left Orthodoxy and adopted a new religion, if they return to their Orthodox roots, are usually received back into the church through Chrismation.
Properly and generally, the Mystery of Baptism is administered by bishops and other priests; however, in emergencies any Orthodox Christian can baptize. In such cases, should the person survive the emergency, it is likely that the person will be properly baptized by a priest at some later date. This is not considered to be a second baptism, nor is it imagined that the person is not already Orthodox, but rather it is a fulfillment of the proper form.
The service of Baptism in Greek Orthodox (and other Eastern Orthodox) churches has remained largely unchanged for over 1500 years. This fact is witnessed to by St. Cyril of Jerusalem (d. 386), who, in his "Discourse on the Sacrament of Baptism", describes the service in much the same way as is currently in use.
The Christian Congregation of Jehovah's Witnesses believes that baptism should be performed by complete immersion (submersion) in water and only when an individual is old enough to understand its significance. They believe that water baptism is an outward symbol that a person has made an unconditional dedication through Jesus Christ to do the will of God. Only after baptism, is a person considered a full-fledged Witness, and an official member of the Christian Congregation. They consider baptism to constitute ordination as a minister.
Prospective candidates for baptism must express their desire to be baptized well in advance of a planned baptismal event, to allow for congregation elders to assess their suitability (regarding true repentance and conversion). Elders approve candidates for baptism if the candidates are considered to understand what is expected of members of the religion and to demonstrate sincere dedication to the faith.
Most baptisms among Jehovah's Witnesses are performed at scheduled assemblies and conventions by elders and ministerial servants, in special pools, or sometimes oceans, rivers, or lakes, depending on circumstances, and rarely occur at local Kingdom Halls. Prior to baptism, at the conclusion of a pre-baptism talk, candidates must affirm two questions:
Only baptized males (elders or ministerial servants) may baptize new members. Baptizers and candidates wear swimsuits or other informal clothing for baptism, but are directed to avoid clothing that is considered undignified or too revealing. Generally, candidates are individually immersed by a single baptizer, unless a candidate has special circumstances such as a physical disability. In circumstances of extended isolation, a qualified candidate's dedication and stated intention to become baptized may serve to identify him as a member of Jehovah's Witnesses, even if immersion itself must be delayed. In rare instances, unbaptized males who had stated such an intention have reciprocally baptized each other, with both baptisms accepted as valid. Individuals who had been baptized in the 1930s and 1940s by female Witnesses due to extenuating circumstances, such as in concentration camps, were later re-baptized but still recognized their original baptism dates.
In The Church of Jesus Christ of Latter-day Saints (LDS Church), baptism is recognized as the first of several ordinances (rituals) of the gospel. In Mormonism, baptism has the main purpose of remitting the sins of the participant. It is followed by confirmation, which inducts the person into membership in the church and constitutes a baptism with the Holy Spirit. Latter-day Saints believe that baptism must be by full immersion, and by a precise ritualized ordinance: if some part of the participant is not fully immersed, or the ordinance was not recited verbatim, the ritual must be repeated. It typically occurs in a baptismal font.
In addition, members of the LDS Church do not believe a baptism is valid unless it is performed by a Latter-day Saint one who has proper authority (a priest or elder). Authority is passed down through a form of apostolic succession. All new converts to the faith must be baptized or re-baptized. Baptism is seen as symbolic both of Jesus' death, burial and resurrection and is also symbolic of the baptized individual discarding their "natural" self and donning a new identity as a disciple of Jesus.
According to Latter-day Saint theology, faith and repentance are prerequisites to baptism. The ritual does not cleanse the participant of original sin, as Latter-day Saints do not believe the doctrine of original sin. Mormonism rejects infant baptism and baptism must occur after the age of accountability, defined in Latter-day Saint scripture as eight years old.
Latter-day Saint theology also teaches baptism for the dead in which deceased ancestors are baptized vicariously by the living, and believe that their practice is what Paul wrote of in Corinthians 15:29. This occurs in Latter-day Saint temples.
Quakers (members of the Religious Society of Friends) do not believe in the baptism of either children or adults with water, rejecting all forms of outward sacraments in their religious life. Robert Barclay's "Apology for the True Christian Divinity" (a historic explanation of Quaker theology from the 17th century), explains Quakers' opposition to baptism with water thus:
Barclay argued that water baptism was only something that happened until the time of Christ, but that now, people are baptised inwardly by the spirit of Christ, and hence there is no need for the external sacrament of water baptism, which Quakers argue is meaningless.
The Salvation Army does not practice water baptism, or indeed other outward sacraments. William Booth and Catherine Booth, the founders of the Salvation Army, believed that many Christians had come to rely on the outward signs of spiritual grace rather than on grace itself. They believed what was important was spiritual grace itself. However, although the Salvation Army does not practice baptism, they are not opposed to baptism within other Christian denominations.
There are some Christians termed "Hyperdispensationalists" (Mid-Acts dispensationalism) who accept only Paul's Epistles as directly applicable for the church today. They do not accept water baptism as a practice for the church since Paul who was God's apostle to the nations was not sent to baptize. Ultradispensationalists (Acts 28 dispensationalism) who do not accept the practice of the Lord's supper, do not practice baptism because these are not found in the Prison Epistles. Both sects believe water baptism was a valid practice for covenant Israel. Hyperdispensationalists also teach that Peter's gospel message was not the same as Paul's. Hyperdispensationalists assert:
Water baptism found early in the Book of Acts is, according to this view, now supplanted by the one baptism foretold by John the Baptist. Others make a distinction between John's prophesied baptism by Christ with the Holy Spirit and the Holy Spirit's baptism of the believer into the body of Christ; the latter being the one baptism for today. The one baptism for today, it is asserted, is the "baptism of the Holy Spirit" of the believer into the Body of Christ church.
Many in this group also argue that John's promised baptism by fire is pending, referring to the destruction of the world by fire.
John, as he said "baptized with water", as did Jesus's disciples to the early, Jewish Christian church. Jesus himself never personally baptized with water, but did so through his disciples. Unlike Jesus' first apostles, Paul, his apostle to the Gentiles, was sent to preach rather than to baptize in contradiction to the Great Commission. But Paul did occasionally still baptize Jews, for instance in Corinth and in Philippi. He also taught the spiritual significance of Spirit baptism in identifying the believer with the atoning death of Christ, his burial, and resurrection. Romans 6 baptism does not mention nor imply water but is rather a real baptism into Christ's death.
Other Hyperdispensationalists believe that baptism was necessary until mid-Acts. The great commission and its baptism was directed to early Jewish believers, not the Gentile believers of mid-Acts or later. Any Jew who believed did not receive salvation or the baptism of the Holy Spirit until they were water baptized. This period ended with the calling of Paul. Peter's reaction when the Gentiles received the Holy Spirit before baptism is seen as proof of transition from water to Spirit baptism. Also significant is the lack of any instructions in the Acts 15 apostolic conference requiring Gentiles to be water baptized.
Most Christian churches see baptism as a once-in-a-lifetime event that can be neither repeated nor undone. They hold that those who have been baptized remain baptized, even if they renounce the Christian faith by adopting a non-Christian religion or by rejecting religion entirely. But some other organizations and individuals are practicing debaptism.
Comparative Summary of Baptisms of Denominations of Christian Influence. (This section does not give a complete listing of denominations, and therefore, it only mentions a fraction of the churches practicing "believer's baptism".)
Many cultures practice or have practiced initiation rites, with or without the use of water, including the ancient Egyptian, the Hebraic/Jewish, the Babylonian, the Mayan, and the Norse cultures. The modern Japanese practice of Miyamairi is such as ceremony that does not use water. In some, such evidence may be archaeological and descriptive in nature, rather than a modern practice.
Apuleius, a 2nd-century Roman writer, described an initiation into the mysteries of Isis. The initiation was preceded by a normal bathing in the public baths and a ceremonial sprinkling by the priest of Isis, after which the candidate was given secret instructions in the temple of the goddess. The candidate then fasted for ten days from meat and wine, after which he was dressed in linen and led at night into the innermost part of the sanctuary, where the actual initiation, the details of which were secret, took place. On the next two days, dressed in the robes of his consecration, he participated in feasting. Apuleius describes also an initiation into the cult of Osiris and yet a third initiation, of the same pattern as the initiation into the cult of Isis, without mention of a preliminary bathing.
The water-less initiations of Lucius, the character in Apuleius's story who had been turned into an ass and changed back by Isis into human form, into the successive degrees of the rites of the goddess was accomplished only after a significant period of study to demonstrate his loyalty and trustworthiness, akin to catechumenal practices preceding baptism in Christianity.
The Ecclesia Gnostica Catholica, or Gnostic Catholic Church (the ecclesiastical arm of Ordo Templi Orientis), offers its Rite of Baptism to any person at least 11 years old. The ceremony is performed before a Gnostic Mass and represents a symbolic birth into the Thelemic community.
The word "baptism" or "christening" is sometimes used to describe the inauguration of certain objects for use.
"Baptism of Ships": at least since the time of the Crusades, rituals have contained a blessing for ships. The priest begs God to bless the vessel and protect those who sail in. The ship is usually sprinkled with holy water.
The name "Baptism of Bells" has been given to the blessing of (musical, especially church) bells, at least in France, since the 11th century. It is derived from the washing of the bell with holy water by the bishop, before he anoints it with the oil of the infirm without and with chrism within; a fuming censer is placed under it and the bishop prays that these sacramentals of the Church may, at the sound of the bell, put the demons to flight, protect from storms, and call the faithful to prayer.
"Baptism of Dolls": the custom of 'dolly dunking' was once a common practice in parts of the United Kingdom, particularly in Cornwall where it has been revived in recent years.
Mandaeans revere John the Baptist and practice frequent baptism (masbuta) as a ritual of purification, not of initiation. They are possibly the earliest people to practice baptism. Mandaeans undergo baptism on Sundays ("Habshaba"), wearing a white sacral robe (Rasta) and consists of a triple full immersion in water, a triple "signing" of the forehead with water and a triple drinking of water. The priest (Rabbi) then removes a ring made of myrtle worn by the baptized and places it on his/her forehead. This is then followed by a handshake (kushta-"hand of truth") with the priest. Living water is a requirement for baptism, therefore can only take place in rivers. All rivers are named "Yardena" (Jordan) and are believed to be nourished by the "World of Light". By the river bank, a Mandaean's forehead is anointed with sesame oil and partakes in a communion of bread and water. Baptism for Mandaeans allows for salvation by connecting with the "World of Light" and for forgiveness of sins. | https://en.wikipedia.org/wiki?curid=4298 |
Beatmatching
Beatmatching or pitch cue is a disc jockey technique of pitch shifting or timestretching an upcoming track to match its tempo to that of the currently playing track, and to adjust them such that the beats (and, usually, the bars) are synchronised — e.g. the kicks and snares in two house records hit at the same time when both records are played simultaneously. Beatmatching is a component of beatmixing which employs beatmatching combined with equalization, attention to phrasing and track selection in an attempt to make a single mix that flows together and has a good structure.
The technique was developed to keep the people from leaving the dancefloor at the end of the song. These days it is considered basic among disc jockeys (DJs) in electronic dance music genres, and it is standard practice in clubs to keep the constant beat through the night, even if DJs change in the middle.
Beatmatching is no longer considered a novelty, and new digital software has made the technique much easier to master.
The beatmatching technique consists of the following steps:
One of the key things to consider when beatmatching is the tempo of both songs, and the musical theory behind the songs. Attempting to beatmatch songs with completely different beats per minute (BPM) will result in one of the songs sounding too fast or too slow.
When beatmatching, a popular technique is to vary the equalization of both tracks. For example, when the kicks are occurring on the same beat, a more seamless transition can occur if the lower frequencies are taken out of one of the songs, and the lower frequencies of the other song is boosted. Doing so creates a smoother transition.
The pitch and tempo of a track are normally linked together: spin a disc 5% faster and both pitch and tempo will be 5% higher. However, some modern DJ software can change pitch and tempo independently using time-stretching and pitch-shifting, allowing harmonic mixing. There is also a feature in modern DJ software which may be called "master tempo" or "key adjust" which changes the tempo while keeping the original pitch.
Beatmatching was invented by Francis Grasso in the late 1960s and early 1970s. Initially he was counting the tempo with a metronome and looking for records with the same tempo. Later a mixer was built for him by Alex Rosner which let him listen to any channel in the headphones independently of what was playing on the speakers; this became the defining feature of DJ mixers. That and turntables with pitch control enabled him to mix tracks with different tempo by changing the pitch of the "cued" (redirected to headphones) track to match its tempo with the track being played by ear. Essentially, the technique he originated hasn't changed since.
These days beat-matching is considered central to DJing, and features making it possible are a requirement for DJ-oriented players. In 1978, the Technics SL-1200MK2 turntable was released, whose comfortable and precise sliding pitch control and high torque direct drive motor made beat-matching easier and it became the standard among DJs. With the advent of the compact disc, DJ-oriented Compact Disc players with pitch control and other features enabling beat-matching (and sometimes scratching), dubbed CDJs, were introduced by various companies. More recently, software with similar capabilities has been developed to allow manipulation of digital audio files stored on computers using turntables with special vinyl records (e.g. Final Scratch, M-Audio Torq, Serato Scratch Live) or computer interface (e.g. Traktor DJ Studio, Mixxx, Virtual DJ). Other software including algorithmic beat-matching is Ableton Live, which allows for realtime music manipulation and deconstruction, or Mixmeister, a DJ Mixset creation tool. Freeware software such as Rapid Evolution can detect the beats per minute and determine the percent BPM difference between songs.
The change from pure hardware to software is on the rise, and big DJs are introducing new equipment to their kits such as the laptop, and dropping the difficulty of carrying hundreds of CDs with them. The creation of the mp3-player allowed DJs to have an alternative tool for DJIng. Limitations with mp3-player DJing equipment has meant that only second generation equipment such as the IDJ2 or the Cortex Dmix-300 have the pitch control that alters tempo and allows for beat-matching on a digital music player. However, recent additions to the Pioneer CDJ family, such as the CDJ-2000, allow mp3-player and other digital storage devices (such as external hard drives, SD cards and USB memory sticks) to be connected to the CDJ device via USB. This allows the DJ to make use of the beat-matching capabilities of the CDJ unit whilst playing digital music files from the mp3-player or other storage device.
Most modern DJ hardware and software now offers a "Sync" feature which automatically adjusts the tempo between tracks being mixed so the DJ no longer needs to spend time and effort matching beats. This has caused some controversy in the DJ industry since almost anyone can beat-match thanks to the new function. | https://en.wikipedia.org/wiki?curid=4301 |
Beltane
Beltane or Beltain () is the Gaelic May Day festival. Most commonly it is held on 1 May, or about halfway between the spring equinox and the summer solstice. Historically, it was widely observed throughout Ireland, Scotland and the Isle of Man. In Irish the name for the festival day is (), in Scottish Gaelic () and in Manx Gaelic /. It is one of the four Gaelic seasonal festivals—along with Samhain, Imbolc and Lughnasadh—and is similar to the Welsh .
Beltane is mentioned in some of the earliest Irish literature and is associated with important events in Irish mythology. Also known as ("first of summer"), it marked the beginning of summer and it was when cattle were driven out to the summer pastures. Rituals were performed to protect the cattle, crops and people, and to encourage growth. Special bonfires were kindled, and their flames, smoke and ashes were deemed to have protective powers. The people and their cattle would walk around or between bonfires, and sometimes leap over the flames or embers. All household fires would be doused and then re-lit from the Beltane bonfire. These gatherings would be accompanied by a feast, and some of the food and drink would be offered to the . Doors, windows, byres and livestock would be decorated with yellow May flowers, perhaps because they evoked fire. In parts of Ireland, people would make a May Bush: typically a thorn bush or branch decorated with flowers, ribbons, bright shells and rushlights. Holy wells were also visited, while Beltane dew was thought to bring beauty and maintain youthfulness. Many of these customs were part of May Day or Midsummer festivals in other parts of Great Britain and Europe.
Beltane celebrations had largely died out by the mid-20th century, although some of its customs continued and in some places it has been revived as a cultural event. Since the late 20th century, Celtic neopagans and Wiccans have observed Beltane, or something based on it, as a religious holiday. Neopagans in the Southern Hemisphere celebrate Beltane on or around 1 November.
Beltane was one of four Gaelic seasonal festivals: Samhain (~1 November), Imbolc (~1 February), Beltane (~1 May), and Lughnasadh (~1 August). Beltane marked the beginning of the pastoral summer season, when livestock were driven out to the summer pastures. Rituals were held at that time to protect them from harm, both natural and supernatural, and this mainly involved the "symbolic use of fire". There were also rituals to protect crops, dairy products and people, and to encourage growth. The (often referred to as spirits or fairies) were thought to be especially active at Beltane (as at Samhain) and the goal of many Beltane rituals was to appease them. Most scholars see the as remnants of the pagan gods and nature spirits. Beltane was a "spring time festival of optimism" during which "fertility ritual again was important, perhaps connecting with the waxing power of the sun".
Beltane (the beginning of summer) and Samhain (the beginning of winter) are thought to have been the most important of the four Gaelic festivals. Sir James George Frazer wrote in "The Golden Bough: A Study in Magic and Religion" that the times of Beltane and Samhain are of little importance to European crop-growers, but of great importance to herdsmen. Thus, he suggests that halving the year at 1 May and 1 November dates from a time when the Celts were mainly a pastoral people, dependent on their herds.
The earliest mention of Beltane is in Old Irish literature from Gaelic Ireland. According to the early medieval texts and , Beltane was held on 1 May and marked the beginning of summer. The texts say that, to protect cattle from disease, the druids would make two fires "with great incantations" and drive the cattle between them.
According to 17th-century historian Geoffrey Keating, there was a great gathering at the hill of Uisneach each Beltane in medieval Ireland, where a sacrifice was made to a god named Beil. Keating wrote that two bonfires would be lit in every district of Ireland, and cattle would be driven between them to protect them from disease. There is no reference to such a gathering in the annals, but the medieval "Dindsenchas" includes a tale of a hero lighting a holy fire on Uisneach that blazed for seven years. Ronald Hutton writes that this may "preserve a tradition of Beltane ceremonies there", but adds "Keating or his source may simply have conflated this legend with the information in "Sanas Chormaic" to produce a piece of pseudo-history." Nevertheless, excavations at Uisneach in the 20th century found evidence of large fires and charred bones, showing it to have been ritually significant.
Beltane is also mentioned in medieval Scottish literature. An early reference is found in the poem 'Peblis to the Play', contained in the Maitland Manuscripts of 15th- and 16th-century Scots poetry, which describes the celebration in the town of Peebles.
From the late 18th century to the mid 20th century, many accounts of Beltane customs were recorded by folklorists and other writers. For example John Jamieson, in his "Etymological Dictionary of the Scottish Language" (1808) describes some of the Beltane customs which persisted in the 18th and early 19th centuries in parts of Scotland, which he noted were beginning to die out. In the 19th century, folklorist Alexander Carmichael (1832–1912), collected the song "Am Beannachadh Bealltain" ("The Beltane Blessing") in his "Carmina Gadelica", which he heard from a crofter in South Uist.
Bonfires continued to be a key part of the festival in the modern era. All hearth fires and candles would be doused before the bonfire was lit, generally on a mountain or hill. Ronald Hutton writes that "To increase the potency of the holy flames, in Britain at least they were often kindled by the most primitive of all means, of friction between wood." In the 19th century, for example, John Ramsay described Scottish Highlanders kindling a need-fire or force-fire at Beltane. Such a fire was deemed sacred. In the 19th century, the ritual of driving cattle between two fires—as described in "Sanas Cormaic" almost 1000 years before—was still practised across most of Ireland and in parts of Scotland. Sometimes the cattle would be driven "around" a bonfire or be made to leap over flames or embers. The people themselves would do likewise. In the Isle of Man, people ensured that the smoke blew over them and their cattle. When the bonfire had died down, people would daub themselves with its ashes and sprinkle it over their crops and livestock. Burning torches from the bonfire would be taken home, where they would be carried around the house or boundary of the farmstead and would be used to re-light the hearth. From these rituals, it is clear that the fire was seen as having protective powers. Similar rituals were part of May Day, Midsummer or Easter customs in other parts of the British Isles and mainland Europe. According to Frazer, the fire rituals are a kind of imitative or sympathetic magic. According to one theory, they were meant to mimic the Sun and to "ensure a needful supply of sunshine for men, animals, and plants". According to another, they were meant to symbolically "burn up and destroy all harmful influences".
Food was also cooked at the bonfire and there were rituals involving it. Alexander Carmichael wrote that there was a feast featuring lamb, and that formerly this lamb was sacrificed. In 1769, Thomas Pennant wrote that, in Perthshire, a caudle made from eggs, butter, oatmeal and milk was cooked on the bonfire. Some of the mixture was poured on the ground as a libation. Everyone present would then take an oatmeal cake, called the "bannoch Bealltainn" or "Beltane bannock". A bit of it was offered to the spirits to protect their livestock (one bit to protect the horses, one bit to protect the sheep, and so forth) and a bit was offered to each of the animals that might harm their livestock (one to the fox, one to the eagle, and so forth). Afterwards, they would drink the caudle.
According to 18th century writers, in parts of Scotland there was another ritual involving the oatmeal cake. The cake would be cut and one of the slices marked with charcoal. The slices would then be put in a bonnet and everyone would take one out while blindfolded. According to one writer, whoever got the marked piece would have to leap through the fire three times. According to another, those present would pretend to throw them into the fire and, for some time afterwards, they would speak of them as if they were dead. This "may embody a memory of actual human sacrifice", or it may have always been symbolic. A similar ritual (i.e. of pretending to burn someone in the fire) was practised at spring and summer bonfire festivals in other parts of Europe.
Yellow flowers such as primrose, rowan, hawthorn, gorse, hazel, and marsh marigold were placed at doorways and windows in 19th century Ireland, Scotland and Mann. Sometimes loose flowers were strewn at the doors and windows and sometimes they were made into bouquets, garlands or crosses and fastened to them. They would also be fastened to cows and equipment for milking and butter making. It is likely that such flowers were used because they evoked fire. Similar May Day customs are found across Europe.
The May Bush and May Bough was popular in parts of Ireland until the late 19th century. This was a small tree or branch—typically hawthorn, rowan, holly or sycamore—decorated with bright flowers, ribbons, painted shells, and so forth. The tree would either be decorated where it stood, or branches would be decorated and placed inside or outside the house. It may also be decorated with candles or rushlights. Sometimes a May Bush would be paraded through the town. In parts of southern Ireland, gold and silver hurling balls known as May Balls would be hung on these May Bushes and handed out to children or given to the winners of a hurling match. In Dublin and Belfast, May Bushes were brought into town from the countryside and decorated by the whole neighbourhood. Each neighbourhood vied for the most handsome tree and, sometimes, residents of one would try to steal the May Bush of another. This led to the May Bush being outlawed in Victorian times. In some places, it was customary to dance around the May Bush, and at the end of the festivities it may be burnt in the bonfire.
Thorn trees were seen as special trees and were associated with the "aos sí". The custom of decorating a May Bush or May Tree was found in many parts of Europe. Frazer believes that such customs are a relic of tree worship and writes: "The intention of these customs is to bring home to the village, and to each house, the blessings which the tree-spirit has in its power to bestow." Emyr Estyn Evans suggests that the May Bush custom may have come to Ireland from England, because it seemed to be found in areas with strong English influence and because the Irish saw it as unlucky to damage certain thorn trees. However, "lucky" and "unlucky" trees varied by region, and it has been suggested that Beltane was the only time when cutting thorn trees was allowed. The practice of bedecking a May Bush with flowers, ribbons, garlands and bright shells is found among the Gaelic diaspora, most notably in Newfoundland, and in some Easter traditions on the East Coast of the United States.
Holy wells were often visited at Beltane, and at the other Gaelic festivals of Imbolc and Lughnasadh. Visitors to holy wells would pray for health while walking sunwise (moving from east to west) around the well. They would then leave offerings; typically coins or clooties (see clootie well). The first water drawn from a well on Beltane was seen as being especially potent, as was Beltane morning dew. At dawn on Beltane, maidens would roll in the dew or wash their faces with it. It would also be collected in a jar, left in the sunlight, and then filtered. The dew was thought to increase sexual attractiveness, maintain youthfulness, and help with skin ailments.
People also took steps specifically to ward-off or appease the "aos sí". Food was left or milk poured at the doorstep or places associated with the , such as 'fairy trees', as an offering. In Ireland, cattle would be brought to 'fairy forts', where a small amount of their blood would be collected. The owners would then pour it into the earth with prayers for the herd's safety. Sometimes the blood would be left to dry and then be burnt. It was thought that dairy products were especially at risk from harmful spirits. To protect farm produce and encourage fertility, farmers would lead a procession around the boundaries of their farm. They would "carry with them seeds of grain, implements of husbandry, the first well water, and the herb vervain (or rowan as a substitute). The procession generally stopped at the four cardinal points of the compass, beginning in the east, and rituals were performed in each of the four directions".
The festival persisted widely up until the 1950s, and in some places the celebration of Beltane continues today.
As a festival, Beltane had largely died out by the mid-20th century, although some of its customs continued and in some places it has been revived as a cultural event. In Ireland, Beltane fires were common until the mid 20th century, but the custom seems to have lasted to the present day only in County Limerick (especially in Limerick itself) and in Arklow, County Wicklow. However, the custom has been revived in some parts of the country. Some cultural groups have sought to revive the custom at Uisneach and perhaps at the Hill of Tara. The lighting of a community Beltane fire from which each hearth fire is then relit is observed today in some parts of the Gaelic diaspora, though in most of these cases it is a cultural revival rather than an unbroken survival of the ancient tradition. In some areas of Newfoundland, the custom of decorating the May Bush is also still extant. The town of Peebles in the Scottish Borders holds a traditional week-long Beltane Fair every year in June, when a local girl is crowned Beltane Queen on the steps of the parish church. Like other Borders festivals, it incorporates a Common Riding.
Since 1988, a Beltane Fire Festival has been held every year during the night of 30 April on Calton Hill in Edinburgh, Scotland. While inspired by traditional Beltane, this festival is a modern arts and cultural event which incorporates myth and drama from a variety of world cultures and diverse literary sources. Two central figures of the Bel Fire procession and performance are the May Queen and the Green Man.
Beltane and Beltane-based festivals are held by some Neopagans. As there are many kinds of Neopaganism, their Beltane celebrations can be very different despite the shared name. Some try to emulate the historic festival as much as possible. Other Neopagans base their celebrations on many sources, the Gaelic festival being only one of them.
Neopagans usually celebrate Beltane on 30 April – 1 May in the Northern Hemisphere and 31 October – 1 November in the Southern Hemisphere, beginning and ending at sunset. Some Neopagans celebrate it at the astronomical midpoint between the spring equinox and summer solstice (or the full moon nearest this point). In the Northern Hemisphere, this midpoint is when the ecliptic longitude of the Sun reaches 45 degrees. In 2014, this was on 5 May.
Celtic Reconstructionists strive to reconstruct the pre-Christian religions of the Celts. Their religious practices are based on research and historical accounts, but may be modified slightly to suit modern life. They avoid modern syncretism and eclecticism (i.e. combining practises from unrelated cultures).
Celtic Reconstructionists usually celebrate when the local hawthorn trees are in bloom. Many observe the traditional bonfire rites, to whatever extent this is feasible where they live. This may involve passing themselves and their pets or livestock between two bonfires, and bringing home a candle lit from the bonfire. If they are unable to make a bonfire or attend a bonfire ceremony, torches or candles may be used instead. They may decorate their homes with a May Bush, branches from blooming thorn trees, or equal-armed rowan crosses. Holy wells may be visited and offerings made to the spirits or deities of the wells. Traditional festival foods may also be prepared.
Wiccans use the name "Beltane" or "Beltain" for their May Day celebrations. It is one of the yearly Sabbats of the Wheel of the Year, following Ostara and preceding Midsummer. Unlike Celtic Reconstructionism, Wicca is syncretic and melds practices from many different cultures. In general, the Wiccan Beltane is more akin to the Germanic/English May Day festival, both in its significance (focusing on fertility) and its rituals (such as maypole dancing). Some Wiccans enact a ritual union of the May Lord and May Lady.
In Irish, the festival is usually called ('day of Beltane') while the month of May is ("month of Beltane"). In Scottish Gaelic, the festival is and the month is or . Sometimes the older Scottish Gaelic spelling is used. The word comes from ('first of summer'), an old alternative name for the festival. The term (Scottish) or (Irish), 'the bright or yellow day of Beltane', means the first of May. In Ireland it is referred to in a common folk tale as ; the first day of the week (Monday/) is added to emphasise the first day of summer.
The name is anglicized as Beltane, Beltain, Beltaine, Beltine and Beltany.
Since the early 20th century it has been commonly accepted that Old Irish is derived from a Common Celtic , meaning 'bright fire'. The element might be cognate with the English word "bale" (as in ) meaning 'white' or 'shining'; compare Old English , and Lithuanian/Latvian /, found in the name of the Baltic; in Slavic languages or also means 'white', as in ('White Rus′' or Belarus) or ('White Sea'). A more recent etymology by Xavier Delamarre would derive it from a Common Celtic , cognate with the name of the Lithuanian goddess of death , the root of both being Proto-Indo-European *"" ("suffering, death").
In Ó Duinnín's Irish dictionary (1904), Beltane is referred to as which it explains is short for meaning 'first (of) summer'. The dictionary also states that is May Day and is the month of May.
There are a number of place names in Ireland containing the word , indicating places where Bealtaine festivities were once held. It is often anglicised as "Beltany". There are three Beltanys in County Donegal, including the Beltany stone circle, and two in County Tyrone. In County Armagh there is a place called Tamnaghvelton/ ('the Beltane field'). Lisbalting/ ('the Beltane ringfort') is in County Tipperary, while Glasheennabaultina/ ('the Beltane stream') is the name of a stream joining the River Galey in County Limerick. | https://en.wikipedia.org/wiki?curid=4306 |
Bethlehem
Bethlehem (; , "House of Meat"; "", , "House of Bread"; ; ; initially named after Canaanite fertility god Lehem) is a city located in the central West Bank, Palestine, about south of Jerusalem. Its population is approximately 25,000 people. It is the capital of the Bethlehem Governorate. The economy is primarily tourist-driven, peaking during the Christmas season, when Christians make pilgrimage to the Church of the Nativity. Rachel's Tomb, an important Jewish holy site, is located at the northern entrance of Bethlehem.
The earliest known mention of Bethlehem was in the Amarna correspondence of 1350–1330 BCE when the town was inhabited by the Canaanites. The Hebrew Bible, which says that the city of Bethlehem was built up as a fortified city by Rehoboam, identifies it as the city David was from and where he was crowned as the king of Israel. The Gospels of Matthew and Luke identify Bethlehem as the birthplace of Jesus. Bethlehem was destroyed by the Emperor Hadrian during the second-century Bar Kokhba revolt; its rebuilding was promoted by Empress Helena, mother of Constantine the Great, who commissioned the building of its great Church of the Nativity in 327 CE. The church was badly damaged by the Samaritans, who sacked it during a revolt in 529, but was rebuilt a century later by Emperor Justinian I.
Bethlehem became part of Jund Filastin following the Muslim conquest in 637. Muslim rule continued in Bethlehem until its conquest in 1099 by a crusading army, who replaced the town's Greek Orthodox clergy with a Latin one. In the mid-13th century, the Mamluks demolished the city's walls, which were subsequently rebuilt under the Ottomans in the early 16th century. Control of Bethlehem passed from the Ottomans to the British at the end of World War I. Bethlehem came under Jordanian rule during the 1948 Arab-Israeli War and was later captured by Israel in the 1967 Six-Day War. Since the 1995 Oslo Accords, Bethlehem has been administered by the Palestinian Authority.
Following an influx of refugees as a result of Israeli advances in the 1967 war, Bethlehem has a Muslim majority, but is still home to a significant Palestinian Christian community. It is now encircled and encroached upon by dozens of Israeli settlements and the Israeli West Bank barrier, which separates both Muslim and Christian communities from their land and livelihoods, and sees a steady exodus of those from both communities being driven out.
The earliest reference to Bethlehem appears in the Amarna correspondence (c. 1400 BCE). In one of his six letters to Pharaoh, Abdi-Heba, the Egyptian-appointed governor of Jerusalem, appeals for aid in retaking "Bit-Laḫmi" in the wake of disturbances by Apiru mercenaries: "Now even a town near Jerusalem, Bit-Lahmi by name, a village which once belonged to the king, has fallen to the enemy ... Let the king hear the words of your servant Abdi-Heba, and send archers to restore the imperial lands of the king!"
It is thought that the similarity of this name to its modern forms indicates that it was originally a settlement of Canaanites who shared a Semitic cultural and linguistic heritage with the later arrivals. "Laḫmu" was the Akkadian god of fertility, worshipped by the Canaanites as "Leḥem". Some time in the third millennium BCE, Canaanites erected a temple on the hill now known as the Hill of the Nativity, probably dedicated to Lehem. The temple, and subsequently the town that formed around it, would then have been known as "Beyt Leḥem", "House (Temple) of Lehem". The Philistines later established a garrison there.
Biblical scholar William F. Albright noted that the pronunciation of the name remained essentially the same for 3,500 years, but has meant different things: "'Temple of the God Lakhmu' in Canaanite, 'House of Bread' in Hebrew and Aramaic, 'House of Meat' in Arabic."
A burial ground discovered in spring 2013, and surveyed in 2015 by a joint Italian-Palestinian team found that the necropolis covered 3 hectares (more than 7 acres) and originally contained more than 100 tombs in use between roughly 2200 B.C. and 650 B.C. The archaeologists were able to identify at least 30 tombs.
Archaeological confirmation of Bethlehem as a city in the Kingdom of Judah was uncovered in 2012 at the archaeological dig at the City of David in the form of a "bulla" (seal impression in dried clay) in ancient Hebrew script that reads "From the town of Bethlehem to the King," indicating that it was used to seal the string closing a shipment of grain, wine, or other goods sent as a tax payment in the 8th or 7th century BCE.
Biblical scholars believe Bethlehem, located in the "hill country" of Judah, may be the same as the Biblical Ephrath, which means "fertile", as there is a reference to it in the Book of Micah as Bethlehem Ephratah. The Bible also calls it Beth-Lehem Judah, and the New Testament describes it as the "City of David". It is first mentioned in the Tanakh and the Bible as the place where the matriarch Rachel died and was buried "by the wayside" (Gen. 48:7). Rachel's Tomb, the traditional grave site, stands at the entrance to Bethlehem. According to the Book of Ruth, the valley to the east is where Ruth of Moab gleaned the fields and returned to town with Naomi. It was the home of Jesse, father of King David of Israel, and the site of David's anointment by the prophet Samuel. It was from the well of Bethlehem that three of his warriors brought him water when he was hiding in the cave of Adullam.
Writing in the 4th century, the Pilgrim of Bordeaux reported that the sepulchers of David, Ezekiel, Asaph, Job, Jesse, and Solomon were located near Bethlehem. There has been no corroboration of this.
The Gospel of Matthew and the Gospel of Luke represent Jesus as having been born in Bethlehem. Modern scholars, however, regard the two accounts as contradictory and the Gospel of Mark, the earliest gospel, mentions nothing about Jesus having been born in Bethlehem, saying only that he came from Nazareth. Nonetheless, the tradition that Jesus was born in Bethlehem was prominent in the early church. In around 155, the apologist Justin Martyr recommended that those who doubted Jesus was really born in Bethlehem could go there and visit the very cave where he was supposed to have been born. The same cave is also referenced by the apocryphal Gospel of James and the fourth-century church historian Eusebius. After the Bar Kokhba revolt ( 132–136 CE) was crushed, the Roman emperor Hadrian converted the Christian site above the Grotto into a shrine dedicated to the Greek god Adonis, to honour his favourite, the Greek youth Antinous.
In around 395 CE, the Church Father Jerome wrote in a letter: "Bethlehem... belonging now to us... was overshadowed by a grove of Tammuz, that is to say, Adonis, and in the cave where once the infant Christ cried, the lover of Venus was lamented." Many scholars have taken this letter as evidence that the cave of the nativity over which the Church of the Nativity was later built had at one point been a shrine to the ancient Near Eastern fertility god Tammuz. Eusebius, however, mentions nothing about the cave having been associated with Tammuz and there are no other Patristic sources that suggest Tammuz had a shrine in Bethlehem. Peter Welten has argued that the cave was never dedicated to Tammuz and that Jerome misinterpreted Christian mourning over the Massacre of the Innocents as a pagan ritual over Tammuz's death. Joan E. Taylor has countered this contention by arguing that Jerome, as an educated man, could not have been so naïve as to mistake Christian mourning over the Massacre of the Innocents as a pagan ritual for Tammuz.
In 326–328, the empress Helena, consort of the emperor Constantius Chlorus, and mother of the emperor Constantine the Great, made a pilgrimage to Syra-Palaestina, in the course of which she visited the ruins of Bethlehem. The Church of the Nativity was built at her initiative over the cave where Jesus was purported to have been born. During the Samaritan revolt of 529, Bethlehem was sacked and its walls and the Church of the Nativity destroyed; they were rebuilt on the orders of the Emperor Justinian I. In 614, the Persian Sassanid Empire, supported by Jewish rebels, invaded Palestina Prima and captured Bethlehem. A story recounted in later sources holds that they refrained from destroying the church on seeing the magi depicted in Persian clothing in a mosaic.
In 637, shortly after Jerusalem was captured by the Muslim armies, 'Umar ibn al-Khattāb, the second Caliph, promised that the Church of the Nativity would be preserved for Christian use. A mosque dedicated to Umar was built upon the place in the city where he prayed, next to the church. Bethlehem then passed through the control of the Islamic caliphates of the Umayyads in the 8th century, then the Abbasids in the 9th century. A Persian geographer recorded in the mid-9th century that a well preserved and much venerated church existed in the town. In 985, the Arab geographer al-Muqaddasi visited Bethlehem, and referred to its church as the "Basilica of Constantine, the equal of which does not exist anywhere in the country-round." In 1009, during the reign of the sixth Fatimid Caliph, al-Hakim bi-Amr Allah, the Church of the Nativity was ordered to be demolished, but was spared by local Muslims, because they had been permitted to worship in the structure's southern transept.
In 1099, Bethlehem was captured by the Crusaders, who fortified it and built a new monastery and cloister on the north side of the Church of the Nativity. The Greek Orthodox clergy were removed from their sees and replaced with Latin clerics. Up until that point the official Christian presence in the region was Greek Orthodox. On Christmas Day 1100, Baldwin I, first king of the Frankish Kingdom of Jerusalem, was crowned in Bethlehem, and that year a Latin episcopate was also established in the town.
In 1187, Saladin, the Sultan of Egypt and Syria who led the Muslim Ayyubids, captured Bethlehem from the Crusaders. The Latin clerics were forced to leave, allowing the Greek Orthodox clergy to return. Saladin agreed to the return of two Latin priests and two deacons in 1192. However, Bethlehem suffered from the loss of the pilgrim trade, as there was a sharp decrease of European pilgrims. William IV, Count of Nevers had promised the Christian bishops of Bethlehem that if Bethlehem should fall under Muslim control, he would welcome them in the small town of Clamecy in present-day Burgundy, France. As a result, the Bishop of Bethlehem duly took up residence in the hospital of Panthenor, Clamecy, in 1223. Clamecy remained the continuous 'in partibus infidelium' seat of the Bishopric of Bethlehem for almost 600 years, until the French Revolution in 1789.
Bethlehem, along with Jerusalem, Nazareth, and Sidon, was briefly ceded to the Crusader Kingdom of Jerusalem by a treaty between Holy Roman Emperor Frederick II and Ayyubid Sultan al-Kamil in 1229, in return for a ten-year truce between the Ayyubids and the Crusaders. The treaty expired in 1239, and Bethlehem was recaptured by the Muslims in 1244. In 1250, with the coming to power of the Mamluks under Rukn al-Din Baibars, tolerance of Christianity declined. Members of the clergy left the city, and in 1263 the town walls were demolished. The Latin clergy returned to Bethlehem the following century, establishing themselves in the monastery adjoining the Basilica of the Nativity. The Greek Orthodox were given control of the basilica and shared control of the Milk Grotto with the Latins and the Armenians.
From 1517, during the years of Ottoman control, custody of the Basilica was bitterly disputed between the Catholic and Greek Orthodox churches. By the end of the 16th century, Bethlehem had become one of the largest villages in the District of Jerusalem, and was subdivided into seven quarters. The Basbus family served as the heads of Bethlehem among other leaders during this period. The Ottoman tax record and census from 1596 indicates that Bethlehem had a population of 1,435, making it the 13th largest village in Palestine at the time. Its total revenue amounted to 30,000 akce.
Bethlehem paid taxes on wheat, barley and grapes. The Muslims and Christians were organized into separate communities, each having its own leader. Five leaders represented the village in the mid-16th century, three of whom were Muslims. Ottoman tax records suggest that the Christian population was slightly more prosperous or grew more grain than grapes (the former being a more valuable commodity).
From 1831 to 1841, Palestine was under the rule of the Muhammad Ali Dynasty of Egypt. During this period, the town suffered an earthquake as well as the destruction of the Muslim quarter in 1834 by Egyptian troops, apparently as a reprisal for the murder of a favored loyalist of Ibrahim Pasha. In 1841, Bethlehem came under Ottoman rule once again and remained so until the end of World War I. Under the Ottomans, Bethlehem's inhabitants faced unemployment, compulsory military service, and heavy taxes, resulting in mass emigration, particularly to South America. An American missionary in the 1850s reported a population of under 4,000, nearly all of whom belonged to the Greek Church. He also noted that a lack of water crippled the town's growth.
Socin found from an official Ottoman village list from about 1870 that Bethlehem had a population of 179 Muslims in 59 houses, 979 "Latins" in 256 houses, 824 "Greeks" in 213 houses, and 41 Armenians in 11 houses, a total of 539 houses. The population count only included men. Hartmann found that Bethlehem had 520 houses.
Bethlehem was administered by the British Mandate from 1920 to 1948. In the United Nations General Assembly's 1947 resolution to partition Palestine, Bethlehem was included in the international enclave of Jerusalem to be administered by the United Nations. Jordan captured the city during the 1948 Arab-Israeli War. Many refugees from areas captured by Israeli forces in 1947–48 fled to the Bethlehem area, primarily settling in what became the official refugee camps of 'Azza (Beit Jibrin) and 'Aida in the north and Dheisheh in the south. The influx of refugees significantly transformed Bethlehem's Christian majority into a Muslim one.
Jordan retained control of the city until the Six-Day War in 1967, when Bethlehem was captured by Israel, along with the rest of the West Bank. Following the Six-Day War, Israel took control of the city.
On December 21, 1995, Israeli troops withdrew from Bethlehem, and three days later the city came under the administration and military control of the Palestinian National Authority in accordance with the Interim Agreement on the West Bank and the Gaza Strip. During the Second Palestinian Intifada in 2000–2005, Bethlehem's infrastructure and tourism industry were damaged. In 2002, it was a primary combat zone in Operation Defensive Shield, a major military counteroffensive by the Israeli Defense Forces (IDF). The IDF besieged the Church of the Nativity, where dozens of Palestinian militants had sought refuge. The siege lasted 39 days. Several militants were killed. It ended with an agreement to exile 13 of the militants to foreign countries.
Today, the city is surrounded by two bypass roads for Israeli settlers, leaving the inhabitants squeezed between thirty-seven Jewish enclaves, where a quarter of all West Bank settlers, roughly 170,000, live; the gap between the two roads is closed by the 8-metre high Israeli West Bank barrier, which cuts Bethlehem off from its sister city Jerusalem.
Christian families that have lived in Bethlehem for hundreds of years are being forced to leave as land in Bethlehem is seized, and homes bulldozed, for construction of thousands of new Israeli homes. Land seizures for Israeli settlements have also prevented construction of a new hospital for the inhabitants of Bethlehem, as well as the barrier separating dozens of Palestinian families from their farmland and Christian communities from their places of worship.
Bethlehem is located at an elevation of about above sea level, higher than nearby Jerusalem. Bethlehem is situated on the southern portion in the Judean Mountains.
The city is located northeast of Gaza City and the Mediterranean Sea, west of Amman, Jordan, southeast of Tel Aviv, Israel and south of Jerusalem. Nearby cities and towns include Beit Safafa and Jerusalem to the north, Beit Jala to the northwest, Husan to the west, al-Khadr and Artas to the southwest, and Beit Sahour to the east. Beit Jala and the latter form an agglomeration with Bethlehem. The Aida and Azza refugee camps are located within the city limits.
In the center of Bethlehem is its old city. The old city consists of eight quarters, laid out in a mosaic style, forming the area around the Manger Square. The quarters include the Christian an-Najajreh, al-Farahiyeh, al-Anatreh, al-Tarajmeh, al-Qawawsa and Hreizat quarters and al-Fawaghreh — the only Muslim quarter. Most of the Christian quarters are named after the Arab Ghassanid clans that settled there. Al-Qawawsa Quarter was formed by Arab Christian emigrants from the nearby town of Tuqu' in the 18th century. There is also a Syriac quarter outside of the old city, whose inhabitants originate from Midyat and Ma'asarte in Turkey. The total population of the old city is about 5,000.
Bethlehem has a Mediterranean climate, with hot and dry summers and mild, wetter winters. Winter temperatures (mid-December to mid-March) can be cool and rainy. January is the coldest month, with temperatures ranging from 1 to 13 degree Celsius (33–55 °F). From May through September, the weather is warm and sunny. August is the hottest month, with a high of 30 degrees Celsius (86 °F). Bethlehem receives an average of of rainfall annually, 70% between November and January.
Bethlehem's average annual relative humidity is 60% and reaches its highest rates between January and February. Humidity levels are at their lowest in May. Night dew may occur in up to 180 days per year. The city is influenced by the Mediterranean Sea breeze that occurs around mid-day. However, Bethlehem is affected also by annual waves of hot, dry, sandy and dust "Khamaseen" winds from the Arabian Desert, during April, May and mid-June.
According to Ottoman tax records, Christians made up roughly 60% of the population in the early 16th century, while the Christian and Muslim population became equal by the mid-16th century. However, there were no Muslim inhabitants counted by the end of the century, with a recorded population of 287 adult male tax-payers. Christians, like all non-Muslims throughout the Ottoman Empire, were required to pay the jizya tax. In 1867, an American visitor describes the town as having a population of 3,000 to 4,000; of whom about 100 were Protestants, 300 were Muslims and "the remainder belonging to the Latin and Greek Churches with a few Armenians." Another report from the same year puts the Christian population at 3,000, with an additional 50 Muslims. An 1885 source put the population at approximately 6,000 of "principally Christians, Latins and Greeks" with no Jewish inhabitants.
In 1948, the religious makeup of the city was 85% Christian, mostly of the Greek Orthodox and Roman Catholic denominations, and 13% Muslim. In the 1967 census taken by Israel authorities, the town of Bethlehem proper numbered 14,439 inhabitants, its 7,790 Muslim inhabitants represented 53.9% of the population, while the Christians of various denominations numbered 6,231 or 46.1%.
In the PCBS's 1997 census, the city had a population of 21,670, including a total of 6,570 refugees, accounting for 30.3% of the city's population. In 1997, the age distribution of Bethlehem's inhabitants was 27.4% under the age of 10, 20% from 10 to 19, 17.3% from 20–29, 17.7% from 30 to 44, 12.1% from 45–64 and 5.3% above the age of 65. There were 11,079 males and 10,594 females. In the 2007 PCBS census, Bethlehem had a population of 25,266, of which 12,753 were males and 12,513 were females. There were 6,709 housing units, of which 5,211 were households. The average household consisted of 4.8 family members.
After the Muslim conquest of the Levant in the 630s, the local Christians were Arabized even though large numbers were ethnically Arabs of the Ghassanid clans. Bethlehem's two largest Arab Christian clans trace their ancestry to the Ghassanids, including al-Farahiyyah and an-Najajreh. The former have descended from the Ghassanids who migrated from Yemen and from the Wadi Musa area in present-day Jordan and an-Najajreh descend from Najran. Another Bethlehem clan, al-Anatreh, also trace their ancestry to the Ghassanids.
The percentage of Christians in the town has been in a steady decline since the mid-twentieth century. In 1947, Christians made up 85% of the population, but by 1998, the figure had declined to 40%. In 2005, the mayor of Bethlehem, Victor Batarseh explained that "due to the stress, either physical or psychological, and the bad economic situation, many people are emigrating, either Christians or Muslims, but it is more apparent among Christians, because they already are a minority." The Palestinian Authority is officially committed to equality for Christians, although there have been incidents of violence against them by the Preventive Security Service and militant factions. In 2006, the Palestinian Centre for Research and Cultural Dialogue conducted a poll among the city's Christians according to which 90% said they had had Muslim friends, 73.3% agreed that the PNA treated Christian heritage in the city with respect and 78% attributed the exodus of Christians to the Israeli blockade. The only mosque in the Old City is the Mosque of Omar, located in the Manger Square. By 2016, the Christian population of Bethlehem had declined to only 16%.
A study by Pew Research Center concluded that the decline in the Arab Christian population of the area was partially a result of a lower birth rate among Christians than among Muslims, but also partially due to the fact that Christians were more likely to emigrate from the region than any other religious group. Amon Ramnon, a researcher at the Jerusalem Institute for Policy Research, stated that the reason why more Christians were emigrating than Muslims is because it is easier for Arab Christians to integrate into western communities than for Arab Muslims, since many of them attend church-affiliated schools, where they are taught European languages. A higher percentage of Christians in the region are urban-dwellers, which also makes it easier for them to emigrate and assimilate into western populations. A statistical analysis of the Christian exodus cited lack of economic and educational opportunity, especially due to the Christians' middle-class status and higher education. Since the Second Intifada, 10% of the Christian population have left the city. However, it is likely that there are many other factors, most of which are shared with the Palestinian population as a whole.
Shopping is a major attraction, especially during the Christmas season. The city's main streets and old markets are lined with shops selling Palestinian handicrafts, Middle Eastern spices, jewelry and oriental sweets such as baklawa. Olive wood carvings are the item most purchased by tourists visiting Bethlehem. Religious handicrafts include ornaments handmade from mother-of-pearl, as well as olive wood statues, boxes, and crosses. Other industries include stone and marble-cutting, textiles, furniture and furnishings. Bethlehem factories also produce paints, plastics, synthetic rubber, pharmaceuticals, construction materials and food products, mainly pasta and confectionery.
Cremisan Wine, founded in 1885, is a winery run by monks in the Monastery of Cremisan. The grapes are grown mainly in the al-Khader district. In 2007, the monastery's wine production was around 700,000 liters per year.
In 2008, Bethlehem hosted the largest economic conference to date in the Palestinian territories. It was initiated by Palestinian Prime Minister and former Finance Minister Salam Fayyad to convince more than a thousand businessmen, bankers and government officials from throughout the Middle East to invest in the West Bank and Gaza Strip. A total of 1.4 billion US dollars was secured for business investments in the Palestinian territories.
Tourism is Bethlehem's main industry. Unlike other Palestinian localities prior to 2000, the majority of the employed residents did not have jobs in Israel. More than 20% of the working population is employed in the industry. Tourism accounts for approximately 65% of the city's economy and 11% of the Palestinian National Authority. The city has more than two million visitors every year. Tourism in Bethlehem ground to a halt for over a decade after the Second Intifada, but gradually began to pick back up in the early 2010s.
The Church of the Nativity is one of Bethlehem's major tourist attractions and a magnet for Christian pilgrims. It stands in the center of the city — a part of the Manger Square — over a grotto or cave called the Holy Crypt, where Jesus is believed to have been born. Nearby is the Milk Grotto where the Holy Family took refuge on their Flight to Egypt and next door is the cave where St. Jerome spent thirty years creating the Vulgate, the dominant Latin version of the Bible until the Reformation.
There are over thirty hotels in Bethlehem. Jacir Palace, built in 1910 near the church, is one of Bethlehem's most successful hotels and its oldest. It was closed down in 2000 due to the Israeli-Palestinian conflict, but reopened in 2005 as the Jacir Palace InterContinental at Bethlehem.
In the New Testament, the Gospel of Luke says that Jesus' parents traveled from Nazareth to Bethlehem, where Jesus was born. The Gospel of Matthew mentions Bethlehem as the place of birth, and adds that King Herod was told that a 'King of the Jews' had been born in the town, prompting Herod to order the killing of all the boys who were two years old or under in the town and surrounding area. Joseph, warned of Herod's impending action by an angel of the Lord, decided to flee to Egypt with his family and then later settled in Nazareth after Herod's death.
Early Christian traditions describe Jesus as being born in Bethlehem: in one account, a verse in the Book of Micah is interpreted as a prophecy that the Messiah would be born there. The second century Christian apologist Justin Martyr stated in his "Dialogue with Trypho" (written c. 155–161) that the Holy Family had taken refuge in a cave outside of the town and then placed Jesus in a manger. Origen of Alexandria, writing around the year 247, referred to a cave in the town of Bethlehem which local people believed was the birthplace of Jesus. This cave was possibly one which had previously been a site of the cult of Tammuz. The Gospel of Mark and the Gospel of John do not include a nativity narrative, but refer to him only as being from Nazareth. In a 2005 article in "Archaeology" magazine, archaeologist Aviram Oshri points to an absence of evidence for the settlement of Bethlehem near Jerusalem at the time when Jesus was born, and postulates that Jesus was born in Bethlehem of Galilee. In a 2011 article in "Biblical Archaeology Review" magazine, Jerome Murphy-O'Connor argues for the traditional position that Jesus was born in Bethlehem near Jerusalem.
Christmas rites are held in Bethlehem on three different dates: December 25 is the traditional date by the Roman Catholic and Protestant denominations, but Greek, Coptic and Syrian Orthodox Christians celebrate Christmas on January 6 and Armenian Orthodox Christians on January 19. Most Christmas processions pass through Manger Square, the plaza outside the Basilica of the Nativity. Roman Catholic services take place in St. Catherine's Church and Protestants often hold services at Shepherds' Fields.
Bethlehem celebrates festivals related to saints and prophets associated with Palestinian folklore. One such festival is the annual Feast of Saint George (al-Khadr) on 5–6 May. During the celebrations, Greek Orthodox Christians from the city march in procession to the nearby town of al-Khader to baptize newborns in the waters around the Monastery of St. George and sacrifice a sheep in ritual. The Feast of St. Elijah is commemorated by a procession to Mar Elias, a Greek Orthodox monastery north of Bethlehem.
The women embroiderers of Bethlehem were known for their bridalwear. Bethlehem embroidery was renowned for its "strong overall effect of colors and metallic brilliance." Less formal dresses were made of indigo fabric with a sleeveless coat ("bisht") from locally woven wool worn over top. Dresses for special occasions were made of striped silk with winged sleeves with a short "taqsireh" jacket known as the Bethlehem jacket. The taqsireh was made of velvet or broadcloth, usually with heavy embroidery.
Bethlehem work was unique in its use of couched gold or silver cord, or silk cord onto the silk, wool, felt or velvet used for the garment, to create stylized floral patterns with free or rounded lines. This technique was used for "royal" wedding dresses ("thob malak"), taqsirehs and the "shatwehs" worn by married women. It has been traced by some to Byzantium, and by others to the formal costumes of the Ottoman Empire's elite. As a Christian village, local women were also exposed to the detailing on church vestments with their heavy embroidery and silver brocade.
The art of mother-of-pearl carving is said to have been a Bethlehem tradition since the 15th century when it was introduced by Franciscan friars from Italy. A constant stream of pilgrims generated a demand for these items, which also provided jobs for women. The industry was noted by Richard Pococke, who visited Bethlehem in 1727.
Bethlehem is home to the Palestinian Heritage Center, established in 1991. The center aims to preserve and promote Palestinian embroidery, art and folklore. The International Center of Bethlehem is another cultural center that concentrates primarily on the culture of Bethlehem. It provides language and guide training, woman's studies and arts and crafts displays, and training.
The Bethlehem branch of the Edward Said National Conservatory of Music has about 500 students. Its primary goals are to teach children music, train teachers for other schools, sponsor music research, and the study of Palestinian folklore music.
Bethlehem has four museums: The Crib of the Nativity Theatre and Museum offers visitors 31 three-dimensional models depicting the significant stages of the life of Jesus. Its theater presents a 20-minute animated show. The Badd Giacaman Museum, located in the Old City of Bethlehem, dates back to the 18th century and is primarily dedicated to the history and process of olive oil production. Baituna al-Talhami Museum, established in 1972, contains displays of Bethlehem culture. The International Museum of Nativity was built by United Nations Educational, Scientific and Cultural Organization (UNESCO) to exhibit "high artistic quality in an evocative atmosphere".
Bethlehem is the "muhfaza" (seat) or district capital of the Bethlehem Governorate.
Bethlehem held its first municipal elections in 1876, after the "mukhtars" ("heads") of the quarters of Bethlehem's Old City (excluding the Syriac Quarter) made the decision to elect a local council of seven members to represent each clan in the town. A Basic Law was established so that if the victor for mayor was a Catholic, his deputy should be of the Greek Orthodox community.
Throughout, Bethlehem's rule by the British and Jordan, the Syriac Quarter was allowed to participate in the election, as were the Ta'amrah Bedouins and Palestinian refugees, hence ratifying the number of municipal members in the council to 11. In 1976, an amendment was passed to allow women to vote and become council members and later the voting age was increased from 21 to 25.
There are several branches of political parties on the council, including Communist, Islamist, and secular. The leftist factions of the Palestine Liberation Organization (PLO) such as the Popular Front for the Liberation of Palestine (PFLP) and the Palestinian People's Party (PPP) usually dominate the reserved seats. Hamas gained the majority of the open seats in the 2005 Palestinian municipal elections.
In the October 2012 municipal elections, Fatah member Vera Baboun won, becoming the first female mayor of Bethlehem.
According to the Palestinian Central Bureau of Statistics (PCBS), in 1997, approximately 84% of Bethlehem's population over the age of 10 was literate. Of the city's population, 10,414 were enrolled in schools (4,015 in primary school, 3,578 in secondary and 2,821 in high school). About 14.1% of high school students received diplomas. There were 135 schools in the Bethlehem Governorate in 2006; 100 run the Education Ministry of the Palestinian National Authority, seven by the United Nations Relief and Works Agency (UNRWA) and 28 were private.
Bethlehem is home to Bethlehem University, a Catholic Christian co-educational institution of higher learning founded in 1973 in the Lasallian tradition, open to students of all faiths. Bethlehem University is the first university established in the West Bank, and can trace its roots to 1893 when the De La Salle Christian Brothers opened schools throughout Palestine and Egypt.
Bethlehem has three bus stations owned by private companies which offer service to Jerusalem, Beit Jala, Beit Sahour, Hebron, Nahalin, Battir, al-Khader, al-Ubeidiya and Beit Fajjar. There are two taxi stations that make trips to Beit Sahour, Beit Jala, Jerusalem, Tuqu' and Herodium. There are also two car rental departments: Murad and 'Orabi. Buses and taxis with West Bank licenses are not allowed to enter Israel, including Jerusalem, without a permit.
The Israeli construction of the West Bank barrier has affected Bethlehem politically, socially, and economically. The barrier is located along the northern side of the town's built-up area, within m of houses in 'Aida refugee camp on one side, and the Jerusalem municipality on the other. Most entrances and exits from the Bethlehem agglomeration to the rest of the West Bank are currently subjected to Israeli checkpoints and roadblocks. The level of access varies based on Israeli security directives. Travel for Bethlehem's Palestinian residents from the West Bank into Jerusalem is regulated by a permit-system. Palestinians require a permit to enter the Jewish holy site of Rachel's Tomb. Israeli citizens are barred from entering Bethlehem and the nearby biblical Solomon's Pools.
Bethlehem is twinned with: | https://en.wikipedia.org/wiki?curid=4312 |
Benjamin
Benjamin () was the last-born of Jacob's thirteen children (12 sons and 1 daughter), and the second and last son of Rachel in Jewish, Christian and Islamic tradition. He was the progenitor of the Israelite Tribe of Benjamin. In the Hebrew Bible unlike Rachel's first son, Joseph, Benjamin was born in Canaan.
In the Samaritan Pentateuch, Benjamin's name appears as "Binyaamem" (). In the Quran, Benjamin is referred to as righteous young child, who remained with Jacob when the older brothers plotted against Joseph. Later rabbinic traditions name him as one of four ancient Israelites who died without sin, the other three being Chileab, Jesse and Amram.
According to the Hebrew Bible, Benjamin's name arose when Jacob deliberately changed the name "Benoni", the original name of Benjamin, since Benoni was an allusion to Rachel's dying just after she had given birth, as it means "son of my pain". Textual scholars regard these two names as fragments of naming narratives coming from different sources - one being the Jahwist and the other being the Elohist.
Unusually for one of the 12 tribes of Israel, the Bible does not explain the etymology of Benjamin's name. Medieval commentator Rashi gives two different explanations, based on Midrashic sources. "Son of the south", with south derived from the word for the right hand side, referring to the birth of Benjamin in Canaan, as compared with the birth of all the other sons of Jacob in Aram. Modern scholars have proposed that "son of the south" / "right" is a reference to the tribe being subordinate to the more dominant tribe of Ephraim. Alternatively, Rashi suggests it means "son of days", meaning a son born in Jacob's old age. The Samaritan Pentateuch consistently spells his name "בן ימים", with a terminal mem, ("Binyamim"), which could be translated literally as "spirit man" but is in line with the interpretation that the name was a reference to the advanced age of Jacob when Benjamin was born.
According to classical rabbinical sources, Benjamin was only born after Rachel had fasted for a long time, as a religious devotion with the hope of a new child as a reward. By then Jacob had become over 100 years old. Benjamin is treated as a young child in most of the Biblical narrative, but at one point is abruptly described as the father of ten sons. Textual scholars believe that this is the result of the genealogical passage, in which his children are named, being from a much later source than the Jahwist and Elohist narratives, which make up most of the Joseph narrative, and which consistently describe Benjamin as a child.
The Torah's Joseph narrative, at a stage when Joseph is unrecognised by his brothers, describes Joseph as testing whether his brothers have reformed by secretly planting a silver cup in Benjamin's bag. Then, publicly searching the bags for it, and after "finding" it in Benjamin's possession, demanding that Benjamin become his slave as a punishment.
The narrative goes on to state that when Judah (on behalf of the other brothers) begged Joseph not to enslave Benjamin and instead enslave him, since enslavement of Benjamin would break Jacob's heart. This caused Joseph to recant and reveal his identity. The midrashic book of Jasher argues that prior to revealing his identity, Joseph asked Benjamin to find his missing brother (i.e. Joseph) via astrology, using an astrolabe-like tool. It continues by stating that Benjamin divined that the "man on the throne" was Joseph, so Joseph identified himself to Benjamin (but not the other brothers), and revealed his scheme (as in the Torah) to test how fraternal the other brothers were.
Some classical rabbinical sources argue that Joseph identified himself for other reasons. In these sources, Benjamin swore an oath, on the memory of Joseph, that he was innocent of theft, and, when challenged about how believable the oath would be, explained that remembering Joseph was so important to him that he had named his sons in Joseph's honour. These sources go on to state that Benjamin's oath touched Joseph so deeply that Joseph was no longer able to pretend to be a stranger.
In the narrative, just prior to this test, when Joseph had first met all of his brothers (but not identified himself to them), he had held a feast for them; the narrative heavily implies that Benjamin was Joseph's favorite brother, since he is overcome with tears when he first meets Benjamin in particular, and he gives Benjamin five times as much food as he apportions to the others. According to textual scholars, this is really the Jahwist's account of the reunion after Joseph identifies himself, and the account of the threat to enslave Benjamin is just the Elohist's version of the same event, with the Elohist being more terse about Joseph's emotions towards Benjamin, merely mentioning that Benjamin was given five times as many gifts as the others.
Biblical scholars believe, due to their geographic overlap and their treatment in older passages, that Ephraim and Manasseh were originally considered one tribe, that of "Joseph". According to several biblical scholars, Benjamin was also originally part of this single tribe, but the biblical account of Joseph as his father became lost.
The description of Benjamin being born after the arrival in Canaan is thought by some scholars to refer to the tribe of Benjamin coming into existence by branching from the Joseph group after the tribe had settled in Canaan. A number of biblical scholars suspect that the distinction of the "Joseph tribes" (including Benjamin) is that they were the only Israelites which went to Egypt and returned, while the main Israelite tribes simply emerged as a subculture from the Canaanites and had remained in Canaan throughout.
According to this view, the story of Jacob's visit to Laban to obtain a wife originated as a metaphor for this migration, with the property and family which were gained from Laban representing the gains of the Joseph tribes by the time they returned from Egypt. According to textual scholars, the Jahwist version of the Laban narrative only mentions the Joseph tribes, and Rachel, and does not mention the other tribal matriarchs whatsoever.
According to , Benjamin had ten sons: Bela, Becher, Ashbel, Gera, Naaman, Ehi, Rosh, Muppim, Huppim, and Ard. The name of his wife/wives are not given, but the Book of Jubilees calls his wife Ijasaka and the Book of Jasher mentions two wives, Mechalia the daughter of Aram and Aribath the daughter of Shomron. The classical rabbinical tradition adds that each son's name honors Joseph:
There is a disparity between the list given in Genesis 46 and that in , where the sons of Benjamin are listed along with the tribes they are the progenitors of.
Becher, Gera, Ehi, and Rosh are omitted from the second list. Ard and Naaman, who are the sons of Benjamin according to Numbers 26, are listed as the sons of Belah and are the progenitors of the Ardites and the Naamites respectively.
Though not named in the Quran, Benjamin is referred to as the righteous youngest son of Jacob, in the narrative of Joseph in Islamic tradition. Apart from that, however, Islamic tradition does not provide much detail regarding Benjamin's life, and refers to him as being born from Jacob's wife Rachel. As with Jewish tradition, it also further links a connection between the names of Benjamin's children and Joseph. | https://en.wikipedia.org/wiki?curid=4313 |
Black Sabbath
Black Sabbath were an English rock band formed in Birmingham in 1968 by guitarist Tony Iommi, drummer Bill Ward, bassist Geezer Butler and vocalist Ozzy Osbourne. They are often cited as pioneers of heavy metal music. The band helped define the genre with releases such as "Black Sabbath" (1970), "Paranoid" (1970), and "Master of Reality" (1971). The band had multiple line-up changes following Osbourne's departure in 1979, with Iommi being the only constant member throughout its history.
After previous iterations of the group called the Polka Tulk Blues Band and Earth, the band settled on the name Black Sabbath in 1969. They distinguished themselves through occult themes with horror-inspired lyrics and tuned-down guitars. Signing to Philips Records in November 1969, they released their first single, "Evil Woman" in January 1970. Their debut album, "Black Sabbath", was released the following month. Though it received a negative critical response, the album was a commercial success, leading to a follow-up record, "Paranoid", later in 1970. The band's popularity grew, and by 1973's "Sabbath Bloody Sabbath", critics were starting to respond favourably. Osbourne's regular use of drugs and alcohol led to his firing in 1979. He was replaced by former Rainbow vocalist Ronnie James Dio. Following two albums with Dio, Black Sabbath endured many personnel changes in the 1980s and 1990s that included vocalists Ian Gillan, Glenn Hughes, Ray Gillen and Tony Martin, as well as several drummers and bassists. In 1991, Iommi and Butler rejoined Dio and drummer Vinny Appice to record "Dehumanizer" (1992). The original line-up reunited with Osbourne in 1997 and released a live album "Reunion". Black Sabbath's final studio album and nineteenth overall, "13" (2013), features all of the original members except Ward. During their farewell tour, the band played their final concert in their home city of Birmingham on 4 February 2017.
Black Sabbath were ranked by MTV as the "Greatest Metal Band" of all time, and placed second in VH1's "100 Greatest Artists of Hard Rock" list. "Rolling Stone" magazine ranked them number 85 on their "100 Greatest Artists of All Time". They have sold over 70 million records worldwide. Black Sabbath were inducted into the UK Music Hall of Fame in 2005 and the Rock and Roll Hall of Fame in 2006. They have also won two Grammy Awards for Best Metal Performance, and in 2019 the band were presented a Grammy Lifetime Achievement Award.
Following the break-up of their previous band Mythology in 1968, guitarist Tony Iommi and drummer Bill Ward sought to form a heavy blues rock band in Aston, Birmingham. They enlisted bassist Geezer Butler and vocalist Ozzy Osbourne, who had played together in a band called Rare Breed, Osbourne having placed an advertisement in a local music shop: "OZZY ZIG Needs Gig – has own PA". The new group was initially named the Polka Tulk Blues Band, the name taken either from a brand of talcum powder or an Indian/Pakistani clothing shop; the exact origin is confused. The Polka Tulk Blues Band included slide guitarist Jimmy Phillips, a childhood friend of Osbourne's, and saxophonist Alan "Aker" Clarke. After shortening the name to Polka Tulk, the band again changed their name to Earth (which Osbourne hated) and continued as a four-piece without Phillips and Clarke. Iommi became concerned that Phillips and Clarke lacked the necessary dedication and were not taking the band seriously. Rather than asking them to leave, they instead decided to break up and then quietly reformed the band as a four-piece. While the band was performing under the Earth title, they recorded several demos written by Norman Haines such as "The Rebel", "Song for Jim", and "When I Came Down". The demo titled "Song for Jim" was in reference to Jim Simpson. Simpson was a manager for the bands Bakerloo Blues Line and Tea & Symphony, as well as being trumpet player for the group Locomotive. Simpson had recently started a new club named Henry's Blueshouse at The Crown Hotel in Birmingham and offered to let Earth play there after they agreed to waive the usual support band fee in return for free t-shirts. The audience response was positive and Simpson agreed to manage Earth.
In December 1968, Iommi abruptly left Earth to join Jethro Tull. Although his stint with the band would be short-lived, Iommi made an appearance with Jethro Tull on "The Rolling Stones Rock and Roll Circus" TV show. Unsatisfied with the direction of Jethro Tull, Iommi returned to Earth in January 1969. "It just wasn't right, so I left", Iommi said. "At first I thought Tull were great, but I didn't much go for having a leader in the band, which was Ian Anderson's way. When I came back from Tull, I came back with a new attitude altogether. They taught me that to get on, you got to work for it."
While playing shows in England in 1969, the band discovered they were being mistaken for another English group named Earth. They decided to change their name again. A cinema across the street from the band's rehearsal room was showing the 1963 horror film "Black Sabbath" starring Boris Karloff and directed by Mario Bava. While watching people line up to see the film, Butler noted that it was "strange that people spend so much money to see scary movies". Following that, Osbourne and Butler wrote the lyrics for a song called "Black Sabbath", which was inspired by the work of horror and adventure-story writer Dennis Wheatley, along with a vision that Butler had of a black silhouetted figure standing at the foot of his bed. Making use of the musical tritone, also known as "the Devil's Interval", the song's ominous sound and dark lyrics pushed the band in a darker direction, a stark contrast to the popular music of the late 1960s, which was dominated by flower power, folk music, and hippie culture. Judas Priest frontman Rob Halford has called the track "probably the most evil song ever written". Inspired by the new sound, the band changed their name to Black Sabbath in August 1969, and made the decision to focus on writing similar material, in an attempt to create the musical equivalent of horror films.
The band's first show as Black Sabbath took place on 30 August 1969, in Workington, England. They were signed to Philips Records in November 1969, and released their first single, "Evil Woman" (a cover of a song by the band Crow), recorded at Trident Studios, through Philips subsidiary Fontana Records in January 1970. Later releases were handled by Philips' newly formed progressive rock label, Vertigo Records.
Black Sabbath's first major exposure came when the band appeared on John Peel's "Top Gear" radio show in 1969, performing "Black Sabbath", "N.I.B.", "Behind the Wall of Sleep", and "Sleeping Village" to a national audience in Great Britain shortly before recording of their first album commenced. Although the "Evil Woman" single failed to chart, the band were afforded two days of studio time in November to record their debut album with producer Rodger Bain. Iommi recalls recording live: "We thought 'We have two days to do it and one of the days is mixing.' So we played live. Ozzy was singing at the same time, we just put him in a separate booth and off we went. We never had a second run of most of the stuff."
"Black Sabbath" was released on Friday the 13th, February 1970, and reached number 8 in the UK Albums Chart. Following its U.S. and Canadian release in May 1970 by Warner Bros. Records, the album reached number 23 on the "Billboard" 200, where it remained for over a year. The album was given negative reviews by many critics. Lester Bangs dismissed it in a "Rolling Stone" review as "discordant jams with bass and guitar reeling like velocitised speedfreaks all over each other's musical perimeters, yet never quite finding synch". It sold in substantial numbers despite being panned, giving the band their first mainstream exposure. It has since been certified platinum in both U.S. by the Recording Industry Association of America (RIAA) and in the UK by British Phonographic Industry (BPI), and is now generally accepted as the first heavy metal album.
The band returned to the studio in June 1970, just four months after "Black Sabbath" was released. The new album was initially set to be named "War Pigs" after the song "War Pigs", which was critical of the Vietnam War; however, Warner changed the title of the album to "Paranoid". The album's lead-off single, "Paranoid", was written in the studio at the last minute. Ward explains: "We didn't have enough songs for the album, and Tony just played the [Paranoid] guitar lick and that was it. It took twenty, twenty-five minutes from top to bottom." The single was released in September 1970 and reached number four on the UK Singles Chart, remaining Black Sabbath's only top ten hit. The album followed in the UK in October 1970, where, pushed by the success of the "Paranoid" single, it made number one in the charts.
The U.S. release was held off until January 1971, as the "Black Sabbath" album was still on the charts at the time of "Paranoid"s UK release. Black Sabbath subsequently toured the United States for the first time and played their first U.S. show at a club called Ungano's at 210 West 70th Street in New York City. The album reached No. 12 in the U.S. in March 1971, and would go on to sell four million copies in the U.S., with virtually no radio airplay. Like "Black Sabbath", the album was panned by rock critics of the era, but modern-day reviewers such as AllMusic's Steve Huey cite "Paranoid" as "one of the greatest and most influential heavy metal albums of all time", which "defined the sound and style of heavy metal more than any other record in rock history". The album was ranked at No. 131 on "Rolling Stone" magazine's list of The 500 Greatest Albums of All Time. "Paranoid"s chart success allowed the band to tour the U.S. for the first time in October 1970, which spawned the release of the album's second single "Iron Man". Although the single failed to reach the top 40, "Iron Man" remains one of Black Sabbath's most popular songs, as well as the band's highest charting U.S. single until 1998's "Psycho Man".
In February 1971, after a one-off performance at the Myponga Pop Festival in Australia, Black Sabbath returned to the studio to begin work on their third album. Following the chart success of "Paranoid", the band were afforded more studio time, along with a "briefcase full of cash" to buy drugs. "We were getting into coke, big time", Ward explained. "Uppers, downers, Quaaludes, whatever you like. It got to the stage where you come up with ideas and forget them, because you were just so out of it."
Production completed in April 1971, in July the band released "Master of Reality", just six months after the U.S. release of "Paranoid". The album reached the top ten in the U.S. and the United Kingdom, and was certified gold in less than two months, eventually receiving platinum certification in the 1980s and Double Platinum in the early 21st century. It contained Sabbath's first acoustic songs, alongside fan favourites such as "Children of the Grave" and "Sweet Leaf". Critical response of the era was generally unfavourable, with Lester Bangs delivering an ambivalent review of "Master of Reality" in "Rolling Stone", describing the closing "Children of the Grave" as "naïve, simplistic, repetitive, absolute doggerel – but in the tradition [of rock'n'roll]... The only criterion is excitement, and Black Sabbath's got it". (In 2003, "Rolling Stone" would place the album at number 300 on their 500 Greatest Albums of All Time list.)
Following the "Master of Reality" world tour in 1972, Sabbath took its first break in three years. As Ward explained: "The band started to become very fatigued and very tired. We'd been on the road non-stop, year in and year out, constantly touring and recording. I think "Master of Reality" was kind of like the end of an era, the first three albums, and we decided to take our time with the next album."
In June 1972, the band reconvened in Los Angeles to begin work on their next album at the Record Plant. With more time in the studio, the album saw the band experimenting with new textures, such as strings, piano, orchestration and multi-part songs. Recording was plagued with problems, many as a result of substance abuse issues. Struggling to record the song "Cornucopia" after "sitting in the middle of the room, just doing drugs", Ward was nearly fired. "I hated the song, there were some patterns that were just... horrible," the drummer said. "I nailed it in the end, but the reaction I got was the cold shoulder from everybody. It was like 'Well, just go home, you're not being of any use right now.' I felt like I'd blown it, I was about to get fired". Butler thought that the end product "was very badly produced, as far as I was concerned. Our then-manager insisted on producing it, so he could claim production costs."
The album was originally titled "Snowblind" after the song of the same name, which deals with cocaine abuse. The record company changed the title at the last minute to "Black Sabbath Vol. 4". Ward observed, "There was no Volume 1, 2 or 3, so it's a pretty stupid title really". "Vol. 4" was released in September 1972 and, while critics were dismissive, it achieved gold status in less than a month, and was the band's fourth consecutive release to sell a million in the U.S. "Tomorrow's Dream" was released as a single – the band's first since "Paranoid" – but failed to chart.
Following an extensive tour of the U.S., in 1973 the band travelled again to Australia, followed by a tour for the first time to New Zealand, before moving onto mainland Europe. "The band were definitely in their heyday," recalled Ward, "in the sense that nobody had burnt out quite yet."
Following the "Volume 4" world tour, Black Sabbath returned to Los Angeles to begin work on their next release. Pleased with the "Volume 4" album, the band sought to recreate the recording atmosphere, and returned to the Record Plant studio in Los Angeles. With new musical innovations of the era, the band were surprised to find that the room they had used previously at the Record Plant was replaced by a "giant synthesiser". The band rented a house in Bel Air and began writing in the summer of 1973, but in part because of substance issues and fatigue, they were unable to complete any songs. "Ideas weren't coming out the way they were on "Volume 4" and we really got discontent" Iommi said. "Everybody was sitting there waiting for me to come up with something. I just couldn't think of anything. And if I didn't come up with anything, nobody would do anything."
After a month in Los Angeles with no results, the band opted to return to England. They rented Clearwell Castle in The Forest of Dean. "We rehearsed in the dungeons and it was really creepy but it had some atmosphere, it conjured up things, and stuff started coming out again." While working in the dungeon, Iommi stumbled onto the main riff of "Sabbath Bloody Sabbath", which set the tone for the new material. Recorded at Morgan Studios in London by Mike Butcher and building off the stylistic changes introduced on "Volume 4", new songs incorporated synthesisers, strings, and complex arrangements. Yes keyboardist Rick Wakeman was brought in as a session player, appearing on "Sabbra Cadabra".
In November 1973, Black Sabbath began to receive positive reviews in the mainstream press after the release of "Sabbath Bloody Sabbath", with Gordon Fletcher of "Rolling Stone" calling the album "an extraordinarily gripping affair", and "nothing less than a complete success." Later reviewers such as AllMusic's Eduardo Rivadavia cite the album as a "masterpiece, essential to any heavy metal collection", while also displaying "a newfound sense of finesse and maturity." The album marked the band's fifth consecutive platinum selling album in the U.S., reaching number four on the United Kingdom charts, and number eleven in the U.S.
The band began a world tour in January 1974, which culminated at the California Jam festival in Ontario, California, on 6 April 1974. Attracting over 200,000 fans, Black Sabbath appeared alongside popular 1970s rock and pop bands Deep Purple, Eagles, Emerson, Lake & Palmer, Rare Earth, Seals & Crofts, Black Oak Arkansas, and Earth, Wind & Fire. Portions of the show were telecast on ABC Television in the U.S., exposing the band to a wider American audience. In the same year, the band shifted management, signing with notorious English manager Don Arden. The move caused a contractual dispute with Black Sabbath's former management, and while on stage in the U.S., Osbourne was handed a subpoena that led to two years of litigation.
Black Sabbath began work on their sixth album in February 1975, again in England at Morgan Studios in Willesden, this time with a decisive vision to differ the sound from "Sabbath, Bloody Sabbath". "We could've continued and gone on and on, getting more technical, using orchestras and everything else which we didn't particularly want to. We took a look at ourselves, and we wanted to do a rock album – "Sabbath, Bloody Sabbath" wasn't a rock album, really." Produced by Black Sabbath and Mike Butcher, "Sabotage" was released in July 1975. As with its precursor, the album initially saw favourable reviews, with "Rolling Stone" stating ""Sabotage" is not only Black Sabbath's best record since "Paranoid", it might be their best ever", although later reviewers such as AllMusic noted that "the magical chemistry that made such albums as "Paranoid" and "Volume 4" so special was beginning to disintegrate".
"Sabotage" reached the top 20 in both the U.S. and the United Kingdom, but was the band's first release not to achieve Platinum status in the U.S., only achieving Gold certification. Although the album's only single "Am I Going Insane (Radio)" failed to chart, "Sabotage" features fan favourites such as "Hole in the Sky", and "Symptom of the Universe". Black Sabbath toured in support of "Sabotage" with openers Kiss, but were forced to cut the tour short in November 1975, following a motorcycle accident in which Osbourne ruptured a muscle in his back. In December 1975, the band's record companies released a greatest hits album without input from the band, titled "We Sold Our Soul for Rock 'n' Roll". The album charted throughout 1976, eventually selling two million copies in the U.S.
Black Sabbath began work for their next album at Criteria Studios in Miami, Florida, in June 1976. To expand their sound, the band added keyboard player Gerald Woodroffe, who also had appeared to a lesser extent on "Sabotage". During the recording of "Technical Ecstasy", Osbourne admits that he began losing interest in Black Sabbath and began to consider the possibility of working with other musicians. Recording of "Technical Ecstasy" was difficult; by the time the album was completed Osbourne was admitted to Stafford County Asylum in Britain. It was released on 25 September 1976 to mixed reviews, and (for the first time) later music critics gave the album less favourable retrospective reviews; two decades after its release AllMusic gave the album two stars, and noted that the band was "unravelling at an alarming rate". The album featured less of the doomy, ominous sound of previous efforts, and incorporated more synthesisers and uptempo rock songs. "Technical Ecstasy" failed to reach the top 50 in the U.S., and was the band's second consecutive release not to achieve platinum status, although it was later certified gold in 1997. The album included "Dirty Women", which remains a live staple, as well as Ward's first lead vocal on the song "It's Alright". Touring in support of "Technical Ecstasy" began in November 1976, with openers Boston and Ted Nugent in the U.S., and completed in Europe with AC/DC in April 1977.
In late 1977, while in rehearsal for their next album, and just days before the band was set to enter the studio, Osbourne abruptly quit the band. Iommi called vocalist Dave Walker, a longtime friend of the band, who had previously been a member of Fleetwood Mac and Savoy Brown, and informed him that Osbourne had left the band. Walker, who was at that time fronting a band called Mistress, flew to Birmingham from California in late 1977 to write material and rehearse with Black Sabbath. On 8 January 1978, Black Sabbath made their only live performance with Walker on vocals, playing an early version of the song "Junior's Eyes" on the BBC Television programme "Look! Hear!" Walker later recalled that while in Birmingham he had bumped into Osbourne in a pub and came to the conclusion that Osbourne was not fully committed to leaving Black Sabbath. "The last Sabbath albums were just very depressing for me", Osbourne said. "I was doing it for the sake of what we could get out of the record company, just to get fat on beer and put a record out." Walker has said that he wrote a lot of lyrics during his brief time in the band but none of them were ever used. If any recordings of this version of the band other than the "Look! Hear!" footage still exist, Walker says that he is not aware of them.
Osbourne initially set out to form a solo project featuring former Dirty Tricks members John Frazer-Binnie, Terry Horbury, and Andy Bierne. As the new band were in rehearsals in January 1978, Osbourne had a change of heart and rejoined Black Sabbath. "Three days before we were due to go into the studio, Ozzy wanted to come back to the band", Iommi explained. "He wouldn't sing any of the stuff we'd written with the other guy (Walker), so it made it very difficult. We went into the studio with basically no songs. We'd write in the morning so we could rehearse and record at night. It was so difficult, like a conveyor belt, because you couldn't get time to reflect on stuff. 'Is this right? Is this working properly?' It was very difficult for me to come up with the ideas and putting them together that quick."
The band spent five months at Sounds Interchange Studios in Toronto, Ontario, Canada, writing and recording what would become "Never Say Die!". "It took quite a long time", Iommi said. "We were getting really drugged out, doing a lot of dope. We'd go down to the sessions, and have to pack up because we were too stoned, we'd have to stop. Nobody could get anything right, we were all over the place, everybody's playing a different thing. We'd go back and sleep it off, and try again the next day." The album was released in September 1978, reaching number twelve in the United Kingdom, and number 69 in the U.S. Press response was unfavourable and did not improve over time with Eduardo Rivadavia of AllMusic stating two decades after its release that the album's "unfocused songs perfectly reflected the band's tense personnel problems and drug abuse." The album featured the singles "Never Say Die" and "Hard Road", both of which cracked the top 40 in the United Kingdom. The band also made their second appearance on the BBC's "Top of the Pops", performing "Never Say Die". It took nearly 20 years for the album to be certified Gold in the U.S.
Touring in support of "Never Say Die!" began in May 1978 with openers Van Halen. Reviewers called Black Sabbath's performance "tired and uninspired", a stark contrast to the "youthful" performance of Van Halen, who were touring the world for the first time. The band filmed a performance at the Hammersmith Odeon in June 1978, which was later released on DVD as "Never Say Die". The final show of the tour, and Osbourne's last appearance with the band (until later reunions) was in Albuquerque, New Mexico on 11 December.
Following the tour, Black Sabbath returned to Los Angeles and again rented a house in Bel Air, where they spent nearly a year working on new material for the next album. The entire band were abusing both alcohol and other drugs, but Iommi says Osbourne "was on a totally different level altogether". The band would come up with new song ideas but Osbourne showed little interest and would refuse to sing them. Pressure from the record label and frustrations with Osbourne's lack of input coming to a head, Iommi made the decision to fire Osbourne in 1979. Iommi believed the only options available were to fire Osbourne or break the band up completely. "At that time, Ozzy had come to an end", Iommi said. "We were all doing a lot of drugs, a lot of coke, a lot of everything, and Ozzy was getting drunk so much at the time. We were supposed to be rehearsing and nothing was happening. It was like 'Rehearse today? No, we'll do it tomorrow.' It really got so bad that we didn't do anything. It just fizzled out." Drummer Ward, who was close with Osbourne, was chosen by Tony to break the news to the singer on 27 April 1979. "I hope I was professional, I might not have been, actually. When I'm drunk I am horrible, I am horrid", Ward said. "Alcohol was definitely one of the most damaging things to Black Sabbath. We were destined to destroy each other. The band were toxic, very toxic."
Sharon Arden (later Sharon Osbourne), daughter of Black Sabbath manager Don Arden, suggested former Rainbow vocalist Ronnie James Dio to replace Ozzy Osbourne in 1979. Don Arden was at this point still trying to convince Osbourne to rejoin the band, as he viewed the original line-up as the most profitable. Dio officially joined in June, and the band began writing their next album. With a notably different vocal style from Osbourne's, Dio's addition to the band marked a change in Black Sabbath's sound. "They were totally different altogether", Iommi explains. "Not only voice-wise, but attitude-wise. Ozzy was a great showman, but when Dio came in, it was a different attitude, a different voice and a different musical approach, as far as vocals. Dio would sing "across" the riff, whereas Ozzy would follow the riff, like in "Iron Man". Ronnie came in and gave us another angle on writing."
Geezer Butler temporarily left the band in September 1979 for personal reasons. According to Dio, the band initially hired Craig Gruber (with whom Dio had previously played while in Elf) on bass to assist with writing the new album. Gruber was soon replaced by Geoff Nicholls of Quartz. The new line-up returned to Criteria Studios in November to begin recording work, with Butler returning to the band in January 1980, and Nicholls moving to keyboards. Produced by Martin Birch, "Heaven and Hell" was released on 25 April 1980, to critical acclaim. Over a decade after its release AllMusic said the album was "one of Sabbath's finest records, the band sounds reborn and re-energised throughout". "Heaven and Hell" peaked at number 9 in the United Kingdom, and number 28 in the U.S., the band's highest charting album since "Sabotage". The album eventually sold a million copies in the U.S., and the band embarked on an extensive world tour, making their first live appearance with Dio in Germany on 17 April 1980.
Black Sabbath toured the U.S. throughout 1980 with Blue Öyster Cult on the "Black and Blue" tour, with a show at Nassau Coliseum in Uniondale, New York filmed and released theatrically in 1981 as "Black and Blue". On 26 July 1980, the band played to 75,000 fans at a sold-out Los Angeles Memorial Coliseum with Journey, Cheap Trick, and Molly Hatchet. The next day, the band appeared at the 1980 Day on the Green at Oakland Coliseum. While on tour, Black Sabbath's former label in England issued a live album culled from a seven-year-old performance, titled "Live at Last" without any input from the band. The album reached number five on the British charts, and saw the re-release of "Paranoid" as a single, which reached the top 20.
On 18 August 1980, after a show in Minneapolis, Ward quit the band. "It was intolerable for me to get on the stage without Ozzy. And I drank 24 hours a day, my alcoholism accelerated". Geezer Butler stated that after the show, Ward came in drunk, talking about the things where "He might as well be a Martian". Ward then got angry, and decided to pack his things, and get on a bus to leave. The group then brought in drummer Vinny Appice to replace Ward.
The band completed the "Heaven and Hell" world tour in February 1981, and returned to the studio to begin work on their next album. Black Sabbath's second studio album produced by Martin Birch and featuring Ronnie James Dio as vocalist "Mob Rules" was released in October 1981, to be well received by fans, but less so by the critics. "Rolling Stone" reviewer J. D. Considine gave the album one star, claiming ""Mob Rules" finds the band as dull-witted and flatulent as ever". Like most of the band's earlier work, time helped to improve the opinions of the music press, a decade after its release, AllMusic's Eduardo Rivadavia called "Mob Rules" "a magnificent record". The album was certified gold, and reached the top 20 on the UK charts. The album's title track "The Mob Rules", which was recorded at John Lennon's old house in England, also featured in the 1981 animated film "Heavy Metal", although the film version is an alternate take, and differs from the album version.
Unhappy with the quality of 1980's "Live at Last", the band recorded another live album—titled "Live Evil"—during the "Mob Rules" world tour, across the United States in Dallas, San Antonio, and Seattle, in 1982. During the mixing process for the album, Iommi and Butler had a falling out with Dio. Misinformed by their then-current mixing engineer, Iommi and Butler accused Dio of sneaking into the studio at night to raise the volume of his vocals. In addition, Dio was not satisfied with the pictures of him in the artwork. Butler also accused Dio and Appice of working on a solo album during the album's mixing without telling the other members of Black Sabbath. "Ronnie wanted more say in things," Iommi said. "And Geezer would get upset with him and that is where the rot set in. "Live Evil" is when it all fell apart. Ronnie wanted to do more of his own thing, and the engineer we were using at the time in the studio didn't know what to do, because Ronnie was telling him one thing and we were telling him another. At the end of the day, we just said, 'That's it, the band is over'". "When it comes time for the vocal, nobody tells me what to do. Nobody! Because they're not as good as me, so I do what I want to do," Dio later said. "I refuse to listen to "Live Evil", because there are too many problems. If you look at the credits, the vocals and drums are listed off to the side. Open up the album and see how many pictures there are of Tony, and how many there are of me and Vinny".
Ronnie James Dio left Black Sabbath in November 1982 to start his own band, and took drummer Vinny Appice with him. "Live Evil" was released in January 1983, but was overshadowed by Ozzy Osbourne's platinum selling album "Speak of the Devil".
The remaining original members, Iommi and Butler, began auditioning singers for the band's next release. Deep Purple and Whitesnake's David Coverdale, Samson's Nicky Moore and Lone Star's John Sloman were all considered and Iommi states in his autobiography that Michael Bolton auditioned. The band settled on former Deep Purple vocalist Ian Gillan to replace Dio in December 1982. The project was initially not to be called Black Sabbath, but pressure from the record label forced the group to retain the name. The band entered The Manor Studios in Shipton-on-Cherwell, Oxfordshire, in June 1983 with a returned and newly sober Bill Ward on drums. "That was the very first album that I ever did clean and sober," Ward recalled. "I only got drunk "after" I finished all my work on the album – which wasn't a very good idea... Sixty to seventy per cent of my energy was taken up on learning how to get through the day without taking a drink and learning how to do things without drinking, and thirty per cent of me was involved in the album."
"Born Again" (7 August 1983) was panned on release by critics. Despite this negative reception, it reached number four in the UK, and number 39 in the U.S. Even three decades after its release, AllMusic's Eduardo Rivadavia called the album "dreadful", noting that "Gillan's bluesy style and humorous lyrics were completely incompatible with the lords of doom and gloom".
Unable to tour because of the pressures of the road, Ward quit the band. "I fell apart with the idea of touring," he later explained. "I got so much fear behind touring, I didn't talk about the fear, I drank behind the fear instead and that was a big mistake." He was replaced by former Electric Light Orchestra drummer Bev Bevan for the "Born Again" '83–'84 world tour, (often unofficially referred to as the 'Feighn Death Sabbath '83–'84' World Tour) which began in Europe with Diamond Head, and later in the U.S. with Quiet Riot and Night Ranger. The band headlined the 1983 Reading Festival in England, adding Deep Purple's "Smoke on the Water" to their encore.
The tour in support of "Born Again" included a giant set of the Stonehenge monument. In a move later parodied in the mockumentary "This Is Spinal Tap", the band made a mistake in ordering the set piece. Butler explained:
Following the completion of the "Born Again" tour in March 1984, vocalist Ian Gillan left Black Sabbath to re-join Deep Purple, which was reforming after a long hiatus. Bevan left at the same time, and Gillan remarked that he and Bevan were made to feel like "hired help" by Iommi. The band then recruited an unknown Los Angeles vocalist named David Donato and Ward once again rejoined the band. The new line-up wrote and rehearsed throughout 1984, and eventually recorded a demo with producer Bob Ezrin in October. Unhappy with the results, the band parted ways with Donato shortly after. Disillusioned with the band's revolving line-up, Ward left shortly after stating "This isn't Black Sabbath". Butler would quit Sabbath next in November 1984 to form a solo band. "When Ian Gillan took over that was the end of it for me," he said. "I thought it was just a joke and I just totally left. When we got together with Gillan it was not supposed to be a Black Sabbath album. After we had done the album we gave it to Warner Bros. and they said they were going to put it out as a Black Sabbath album and we didn't have a leg to stand on. I got really disillusioned with it and Gillan was really pissed off about it. That lasted one album and one tour and then that was it."
Following both Ward's and Butler's exits, sole remaining original member Iommi put Sabbath on hiatus, and began work on a solo album with long-time Sabbath keyboardist Geoff Nicholls. While working on new material, the original Sabbath line-up agreed to a spot at Bob Geldof's Live Aid, performing at the Philadelphia show on 13 July 1985. This event – which also featured reunions of The Who and Led Zeppelin – marked the first time the original line-up had appeared on stage since 1978. "We were all drunk when we did Live Aid," recalled Geezer Butler, "but we'd all got drunk separately."
Returning to his solo work, Iommi enlisted bassist Dave Spitz (ex-Great White), drummer Eric Singer and initially intended to use multiple singers, including Rob Halford of Judas Priest, former Deep Purple and Trapeze vocalist Glenn Hughes, and former Sabbath vocalist Ronnie James Dio. This plan didn't work as he forecasted. "We were going to use different vocalists on the album, guest vocalists, but it was so difficult getting it together and getting releases from their record companies. Glenn Hughes came along to sing on one track and we decided to use him on the whole album."
The band spent the remainder of the year in the studio, recording what would become "Seventh Star" (1986). Warner Bros. refused to release the album as a Tony Iommi solo release, instead insisting on using the name Black Sabbath. Pressured by the band's manager, Don Arden, the two compromised and released the album as "Black Sabbath featuring Tony Iommi" in January 1986. "It opened up a whole can of worms," Iommi explained. "If we could have done it as a solo album, it would have been accepted a lot more." "Seventh Star" sounded little like a Sabbath album, incorporating instead elements popularised by the 1980s Sunset Strip hard rock scene. It was panned by the critics of the era, although later reviewers such as AllMusic gave album verdicts, calling the album "often misunderstood and underrated".
The new line-up rehearsed for six weeks preparing for a full world tour, although the band were eventually forced to use the Sabbath name. "I was into the 'Tony Iommi project', but I wasn't into the Black Sabbath moniker," Hughes said. "The idea of being in Black Sabbath didn't appeal to me "whatsoever". Glenn Hughes singing in Black Sabbath is like James Brown singing in Metallica. It wasn't gonna work." Just four days before the start of the tour, Hughes got into a bar fight with the band's production manager John Downing which splintered the singer's orbital bone. The injury interfered with Hughes' ability to sing, and the band brought in vocalist Ray Gillen to continue the tour with W.A.S.P. and Anthrax, although nearly half of the U.S. dates would be cancelled because of poor ticket sales.
One vocalist whose status is disputed, both inside and outside Sabbath, is Christian evangelist and former Joshua frontman Jeff Fenholt. Fenholt insists he was a singer in Sabbath between January and May 1985. Iommi has never confirmed this. Fenholt gives a detailed account in Garry Sharpe-Young's book "Sabbath Bloody Sabbath: The Battle for Black Sabbath".
Black Sabbath began work on new material in October 1986 at Air Studios in Montserrat with producer Jeff Glixman. The recording was fraught with problems from the beginning, as Glixman left after the initial sessions to be replaced by producer Vic Coppersmith-Heaven. Bassist Dave Spitz quit over "personal issues", and former Rainbow and Ozzy Osbourne bassist Bob Daisley was brought in. Daisley re-recorded all of the bass tracks, and wrote the album's lyrics, but before the album was complete, he left to join Gary Moore's backing band, taking drummer Eric Singer with him. After problems with second producer Coppersmith-Heaven, the band returned to Morgan Studios in England in January 1987 to work with new producer Chris Tsangarides. While working in the United Kingdom, new vocalist Ray Gillen abruptly left Black Sabbath to form Blue Murder with guitarist John Sykes (ex-Tygers of Pan Tang, Thin Lizzy, Whitesnake). The band enlisted heavy metal vocalist Tony Martin to re-record Gillen's tracks, and former Electric Light Orchestra drummer Bev Bevan to complete a few percussion overdubs. Before the release of the new album Black Sabbath accepted an offer to play six shows at Sun City, South Africa during the apartheid era. The band drew criticism from activists and artists involved with Artists United Against Apartheid, who had been boycotting South Africa since 1985. Drummer Bev Bevan refused to play the shows, and was replaced by Terry Chimes, formerly of the Clash.
After nearly a year in production, "The Eternal Idol" was released on 8 December 1987 and ignored by contemporary reviewers. On-line internet era reviews were mixed. AllMusic said that "Martin's powerful voice added new fire" to the band, and the album contained "some of Iommi's heaviest riffs in years." "Blender" gave the album two stars, claiming the album was "Black Sabbath in name only". The album would stall at No. 66 in the United Kingdom, while peaking at 168 in the U.S. The band toured in support of "Eternal Idol" in Germany, Italy and for the first time, Greece. In part due to a backlash from promoters over the South Africa incident, other European shows were cancelled. Bassist Dave Spitz left the band shortly before the tour, and was replaced by Jo Burt, formerly of Virginia Wolf.
Following the poor commercial performance of "The Eternal Idol", Black Sabbath were dropped by both Vertigo Records and Warner Bros. Records, and signed with I.R.S. Records. The band took time off in 1988, returning in August to begin work on their next album. As a result of the recording troubles with "Eternal Idol", Tony Iommi opted to produce the band's next album himself. "It was a completely new start", Iommi said. "I had to rethink the whole thing, and decided that we needed to build up some credibility again". Iommi enlisted former Rainbow drummer Cozy Powell, long-time keyboardist Nicholls and session bassist Laurence Cottle, and rented a "very cheap studio in England".
Black Sabbath released "Headless Cross" in April 1989, and it was also ignored by contemporary reviewers, although AllMusic contributor Eduardo Rivadavia gave the album four stars and called it "the finest non-Ozzy or Dio Black Sabbath album". Anchored by the number 62 charting single "Headless Cross", the album reached number 31 on the UK charts, and number 115 in the U.S. Queen guitarist Brian May, a good friend of Iommi's, played a guest solo on the song "When Death Calls". Following the album's release the band added touring bassist Neil Murray, formerly of Colosseum II, National Health, Whitesnake, Gary Moore's backing band, and Vow Wow.
The unsuccessful "Headless Cross" U.S. tour began in May 1989 with openers Kingdom Come and Silent Rage, but because of poor ticket sales, the tour was cancelled after just eight shows. The European leg of the tour began in September, where the band were enjoying chart success. After a string of Japanese shows the band embarked on a 23 date Russian tour with Girlschool. Black Sabbath was one of the first bands to tour Russia, after Mikhail Gorbachev opened the country to western acts for the first time in 1989.
The band returned to the studio in February 1990 to record "Tyr", the follow-up to "Headless Cross". While not technically a concept album, some of the album's lyrical themes are loosely based on Norse mythology. "Tyr" was released on 6 August 1990, reaching number 24 on the UK albums chart, but was the first Black Sabbath release not to break the "Billboard" 200 in the U.S. The album would receive mixed internet-era reviews, with AllMusic noting that the band "mix myth with metal in a crushing display of musical synthesis", while "Blender" gave the album just one star, claiming that "Iommi continues to besmirch the Sabbath name with this unremarkable collection". The band toured in support of "Tyr" with Circus of Power in Europe, but the final seven United Kingdom dates were cancelled because of poor ticket sales. For the first time in their career, the band's touring cycle did not include U.S. dates.
While on his "Lock Up the Wolves" U.S. tour in August 1990, former Sabbath vocalist Ronnie James Dio was joined onstage at the Roy Wilkins Auditorium by Geezer Butler to perform "Neon Knights". Following the show, the two expressed interest in rejoining Sabbath. Butler convinced Iommi, who in turn broke up the current lineup, dismissing vocalist Tony Martin and bassist Neil Murray. "I do regret that in a lot of ways," Iommi said. "We were at a good point then. We decided to [reunite with Dio] and I don't even know why, really. There's the financial aspect, but that wasn't it. I seemed to think maybe we could recapture something we had."
Dio and Butler joined Iommi and Cozy Powell in autumn 1990 to begin the next Sabbath release. While rehearsing in November, Powell suffered a broken hip when his horse died and fell on the drummer's legs. Unable to complete the album, Powell was replaced by former drummer Vinny Appice, reuniting the "Mob Rules" lineup, and the band entered the studio with producer Reinhold Mack. The year-long recording was plagued with problems, primarily stemming from writing tension between Iommi and Dio. Songs were rewritten multiple times. "It was just hard work," Iommi said. "We took too long on it, that album cost us a million dollars, which is bloody ridiculous." Dio recalled the album as difficult, but worth the effort: "It was something we had to really wring out of ourselves, but I think that's why it works. Sometimes you need that kind of tension, or else you end up making the Christmas album".
The resulting "Dehumanizer" was released on 22 June 1992. In the U.S., the album was released on 30 June 1992 by Reprise Records, as Dio and his namesake band were still under contract to the label at the time. While the album received mixed , it was the band's biggest commercial success in a decade. Anchored by the top 40 rock radio single "TV Crimes", the album peaked at number 44 on the "Billboard 200." The album also featured "Time Machine", a version of which had been recorded for the 1992 film "Wayne's World". Additionally, the perception among fans of a return of some semblance of the "real" Sabbath provided the band with much needed momentum.
Sabbath began touring in support of "Dehumanizer" in July 1992 with Testament, Danzig, Prong, and Exodus. While on tour, former vocalist Ozzy Osbourne announced his first retirement, and invited Sabbath to open for his solo band at the final two shows of his "No More Tours" tour in Costa Mesa, California. The band agreed, aside from Dio, who told Iommi, "I'm not doing that. I'm not supporting a clown." Dio spoke of the situation years later:
Dio quit Sabbath following a show in Oakland, California on 13 November 1992, one night before the band were set to appear at Osbourne's retirement show. Judas Priest vocalist Rob Halford stepped in at the last minute, performing two nights with the band. Iommi and Butler joined Osbourne and former drummer Ward on stage for the first time since 1985's Live Aid concert, performing a brief set of Sabbath songs. This set the stage for a longer-term reunion of the original lineup, though that plan proved short-lived. "Ozzy, Geezer, Tony and Bill announced the reunion of Black Sabbath – again," remarked Dio. "And I thought that it was a great idea. But I guess Ozzy didn't think it was such a great idea… I'm never surprised when it comes to whatever happens with them. Never at all. They are very predictable. They don't talk."
Drummer Vinny Appice left the band following the reunion show to rejoin Ronnie James Dio's solo band, later appearing on Dio's "Strange Highways" and "Angry Machines". Iommi and Butler enlisted former Rainbow drummer Bobby Rondinelli, and reinstated former vocalist Tony Martin. The band returned to the studio to work on new material, although the project was not originally intended to be released under the Black Sabbath name. As Geezer Butler explains:
Under pressure from their record label, the band released their seventeenth studio album, "Cross Purposes", on 8 February 1994, under the Black Sabbath name. The album received mixed reviews, with "Blender" giving the album two stars, calling Soundgarden's 1994 album "Superunknown" "a far better Sabbath album than this by-the-numbers potboiler". AllMusic's Bradley Torreano called "Cross Purposes" "the first album since "Born Again" that actually sounds like a real Sabbath record". The album just missed the Top 40 in the UK reaching number 41, and also reached 122 on the "Billboard" 200 in the U.S. "Cross Purposes" contained the song "Evil Eye", which was co-written by Van Halen guitarist Eddie Van Halen, although uncredited because of record label restrictions. Touring in support of "Cross Purposes" began in February with Morbid Angel and Motörhead in the U.S. The band filmed a live performance at the Hammersmith Apollo on 13 April 1994, which was released on VHS accompanied by a CD, titled "Cross Purposes Live". After the European tour with Cathedral and Godspeed in June 1994, drummer Bobby Rondinelli quit the band and was replaced by original Black Sabbath drummer Ward for five shows in South America.
Following the touring cycle for "Cross Purposes", bassist Geezer Butler quit the band for the second time. "I finally got totally disillusioned with the last Sabbath album, and I much preferred the stuff I was writing to the stuff Sabbath were doing". Butler formed a solo project called GZR, and released "Plastic Planet" in 1995. The album contained the song "Giving Up the Ghost", which was critical of Tony Iommi for carrying on with the Black Sabbath name, with the lyrics: "You plagiarised and parodied / the magic of our meaning / a legend in your own mind / left all your friends behind / you can't admit that you're wrong / the spirit is dead and gone" ("I heard it's something about me..." said Iommi. "I had the album given to me a while back. I played it once, then somebody else had it, so I haven't really paid any attention to the lyrics... It's nice to see him doing his own thing – getting things off his chest. I don't want to get into a rift with Geezer. He's still a friend."
Following Butler's departure, newly returned drummer Ward once again left the band. Iommi reinstated former members Neil Murray on bass and Cozy Powell on drums, effectively reuniting the 1990 "Tyr" line-up. The band enlisted Body Count guitarist Ernie C to produce the new album, which was recorded in London in autumn of 1994. The album featured a guest vocal on "Illusion of Power" by Body Count vocalist Ice-T. The resulting "Forbidden" was released on 8 June 1995, but failed to chart in the U.S. The album was widely panned by critics; AllMusic's Bradley Torreano said "with boring songs, awful production, and uninspired performances, this is easily avoidable for all but the most enthusiastic fan"; while "Blender" magazine called "Forbidden" "an embarrassment... the band's worst album".
Black Sabbath embarked on a world tour in July 1995 with openers Motörhead and Tiamat, but two months into the tour, drummer Cozy Powell left the band, citing health issues, and was replaced by former drummer Bobby Rondinelli. "The members I had in the last lineup – Bobby Rondinelli, Neil Murray – they're great, great characters..." Iommi told Sabbath fanzine "Southern Cross". "That, for me, was an ideal lineup. I wasn't sure vocally what we should do, but Neil Murray and Bobby Rondinelli I really got on well with."
After completing Asian dates in December 1995, Tony Iommi put the band on hiatus, and began work on a solo album with former Black Sabbath vocalist Glenn Hughes, and former Judas Priest drummer Dave Holland. The album was not officially released following its completion, although a widely traded bootleg called "Eighth Star" surfaced soon after. The album was officially released in 2004 as "The 1996 DEP Sessions", with Holland's drums re-recorded by session drummer Jimmy Copley.
In 1997, Tony Iommi disbanded the current line-up to officially reunite with Ozzy Osbourne and the original Black Sabbath line-up. Vocalist Tony Martin claimed that an original line-up reunion had been in the works since the band's brief reunion at Ozzy Osbourne's 1992 Costa Mesa show, and that the band released subsequent albums to fulfill their record contract with I.R.S. Records. Martin later recalled "Forbidden" (1995) as a "filler album that got the band out of the label deal, rid of the singer, and into the reunion. However I wasn't privy to that information at the time". I.R.S. Records released a compilation album in 1996 to fulfill the band's contract, titled "The Sabbath Stones", which featured songs from "Born Again" (1983) to "Forbidden" (1995).
In the summer of 1997, Iommi, Butler and Osbourne reunited to coheadline the Ozzfest tour alongside Osbourne's solo band. The line-up featured Osbourne's drummer Mike Bordin filling in for Ward. "It started off with me going off to join Ozzy for a couple of numbers," explained Iommi, "and then it got into Sabbath doing a short set, involving Geezer. And then it grew as it went on… We were concerned in case Bill couldn't make it – couldn't do it – because it was a lot of dates, and important dates… The only rehearsal that we had to do was for the drummer. But I think if Bill had come in, it would have took a lot more time. We would have had to focus a lot more on him."
In December 1997, the group was joined by Ward, marking the first reunion of the original quartet since Osbourne's 1992 "retirement show". This lineup recorded two shows at the Birmingham NEC, released as the double album "Reunion" on 20 October 1998. The album reached number eleven on the "Billboard" 200, went platinum in the U.S. and spawned the single "Iron Man", which won Sabbath their first Grammy Award in 2000 for Best Metal Performance, 30 years after the song was originally released. "Reunion" featured two new studio tracks, "Psycho Man" and "Selling My Soul", both of which cracked the top 20 of the "Billboard Mainstream Rock Tracks" chart.
Shortly before a European tour in the summer of 1998, Ward suffered a heart attack and was temporarily replaced by former drummer Vinny Appice. Ward returned for a U.S. tour with openers Pantera, which began in January 1999 and continued through the summer, headlining the annual Ozzfest tour. Following these appearances, the band was put on hiatus while members worked on solo material. Iommi released his first official solo album, "Iommi", in 2000, while Osbourne continued work on "Down to Earth" (2001).
Sabbath returned to the studio to work on new material with all four original members and producer Rick Rubin in the spring of 2001, but the sessions were halted when Osbourne was called away to finish tracks for his solo album in the summer. "It just came to an end…" Iommi said. "It's a shame because [the songs] were really Iommi commented on the difficulty getting all the members together to work:
In March 2002, Osbourne's Emmy-winning reality show "The Osbournes" debuted on MTV, and quickly became a worldwide hit. The show introduced Osbourne to a broader audience and to capitalise, the band's back catalogue label, Sanctuary Records released a double live album "Past Lives" (2002), which featured concert material recorded in the 1970s, including the "Live at Last" (1980) album. The band remained on hiatus until the summer of 2004 when they returned to headline Ozzfest 2004 and 2005. In November 2005, Black Sabbath were inducted into the UK Music Hall of Fame, and in March 2006, after eleven years of eligibility, the band were inducted into the U.S. Rock and Roll Hall of Fame. At the awards ceremony Metallica played two Sabbath songs, "Hole in the Sky" and "Iron Man" in tribute.
While Ozzy Osbourne was working on new solo album material in 2006, Rhino Records released "", a compilation of songs culled from the four Black Sabbath releases featuring Ronnie James Dio. For the release, Iommi, Butler, Dio, and Appice reunited to write and record three new songs as Black Sabbath. "The Dio Years" was released on 3 April 2007, reaching number 54 on the "Billboard" 200, while the single "The Devil Cried" reached number 37 on the "Mainstream Rock Tracks" chart. Pleased with the results, Iommi and Dio decided to reunite the Dio era line-up for a world tour. While the line-up of Osbourne, Butler, Iommi, and Ward was still officially called Black Sabbath, the new line-up opted to call themselves Heaven & Hell, after the album of the same title, to avoid confusion. When asked about the name of the group, Iommi stated "it really is Black Sabbath, whatever we do... so everyone knows what they're getting [and] so people won't expect to hear 'Iron Man' and all those songs. We've done them for so many years, it's nice to do just all the stuff we did with Ronnie again."
Ward was initially set to participate, but dropped out before the tour began due to musical differences with "a couple of the band members". He was replaced by former drummer Vinny Appice, effectively reuniting the line-up that had featured on the "Mob Rules" (1981) and "Dehumanizer" (1992) albums.
Heaven & Hell toured the U.S. with openers Megadeth and Machine Head, and recorded a live album and DVD in New York on 30 March 2007, titled "Live from Radio City Music Hall". In November 2007, Dio confirmed that the band had plans to record a new studio album, which was recorded in the following year. In April 2008 the band announced the upcoming release of a new box set and their participation in the Metal Masters Tour, alongside Judas Priest, Motörhead and Testament. The box set, "The Rules of Hell", featuring remastered versions of all the Dio fronted Black Sabbath albums, was supported by the Metal Masters Tour. In 2009, the band announced the title of their debut studio album, "The Devil You Know", released on 28 April.
On 26 May 2009, Osbourne filed suit in a federal court in New York against Iommi alleging that he illegally claimed the band name. Iommi noted that he has been the only constant band member for its full 41-year career and that his bandmates relinquished their rights to the name in the 1980s, therefore claiming more rights to the name of the band. Although in the suit, Osbourne was seeking 50% ownership of the trademark, he said that he hoped the proceedings would lead to equal ownership among the four original members.
In March 2010, Black Sabbath announced that along with Metallica they would be releasing a limited edition single together to celebrate Record Store Day. It was released on 17 April 2010. Ronnie James Dio died on 16 May 2010 from stomach cancer. In June 2010, the legal battle between Ozzy Osbourne and Tony Iommi over the trademarking of the Black Sabbath name ended, but the terms of the settlement have not been disclosed.
In a January 2010 interview while promoting his biography "I Am Ozzy", Osbourne stated that although he would not rule it out, he was doubtful there would be a reunion with all four original members of the band. Osbourne stated: "I'm not gonna say I've written it out forever, but right now I don't think there's any chance. But who knows what the future holds for me? If it's my destiny, fine." In July, Butler said that there would be no reunion in 2011, as Osbourne was already committed to touring with his solo band. However, by that August they had already met up to rehearse together, and continued to do so through the autumn.
On 11 November 2011, Iommi, Butler, Osbourne, and Ward announced that they were reuniting to record a new album with a full tour in support beginning in 2012. Guitarist Iommi was diagnosed with lymphoma on 9 January 2012, which forced the band to cancel all but two shows (Download Festival, and Lollapalooza Festival) of a previously booked European tour. It was later announced that an intimate show would be played in their hometown Birmingham. It was the first concert since the reunion and the only indoors concerts that year. In February 2012, drummer Ward announced that he would not participate further in the band's reunion until he was offered a "signable contract".
On 21 May 2012, at the O2 Academy in Birmingham, Black Sabbath played their first concert since 2005, with Tommy Clufetos playing the drums. In June, they performed at Download Festival at Donington Park in Leicestershire, England, followed by the last concert of the short tour at Lollapalooza Festival in Chicago. Later that month, the band started recording an album.
On 13 January 2013, the band announced that the album would be released in June under the title "13". Brad Wilk of Rage Against the Machine was chosen as the drummer, and Rick Rubin was chosen as the producer. Mixing of the album commenced in February. On 12 April 2013, the band released the album's track listing. The standard version of the album features eight new tracks, and the deluxe version features three bonus tracks.
The band's first single from "13", "God Is Dead?", was released on 19 April 2013. On 20 April 2013, Black Sabbath commenced their first Australia/New Zealand tour in 40 years followed by a North American Tour in Summer 2013. The second single of the album, "End of the Beginning", debuted on 15 May in a "" episode, where all three members appeared. In June 2013, "13" topped both the UK Albums Chart and the U.S. "Billboard" 200. In 2014, Black Sabbath received their first Grammy Award since 2000 with "God Is Dead?" winning Best Metal Performance.
In July 2013, Black Sabbath embarked on a North American Tour (for the first time since July 2001), followed by a Latin American tour in October 2013. In November 2013, the band started their European tour which lasted until December 2013. In March and April 2014, they made 12 stops in North America (mostly in Canada) as the second leg of their North American Tour before embarking in June 2014 on the second leg of their European tour, which ended with a concert at London's Hyde Park.
On 29 September 2014, Osbourne told "Metal Hammer" that Black Sabbath would begin work on their twentieth studio album in early 2015 with producer Rick Rubin, followed by a final tour in 2016. In an April 2015 interview, however, Osbourne said that these plans "could change", and added, "We all live in different countries and some of them want to work and some of them don't want to, I believe. But we are going to do another tour together."
On 3 September 2015, it was announced that Black Sabbath would embark on their final tour, titled The End, from January 2016 to February 2017. Numerous dates and locations across the U.S., Canada, Europe, Australia and New Zealand were announced. The final shows of The End tour took place at the Genting Arena in their home city of Birmingham, England on 2 and 4 February 2017. On 26 October 2015, it was announced the band consisting of Osbourne, Iommi and Butler would be returning to the Download Festival on 11 June 2016. Despite earlier reports that they would enter the studio before their farewell tour, Osbourne stated that there would not be another Black Sabbath studio album. However, an 8-track CD entitled "The End" was sold at dates on the tour. Along with some live recordings, the CD includes four unused tracks from the "13" sessions.
On 4 March 2016, Iommi discussed future re-releases of the Tony Martin-era catalogue. He explained: "We've held back on the reissues of those albums because of the current Sabbath thing with Ozzy Osbourne, but they will certainly be happening... I'd like to do a couple of new tracks for those releases with Tony Martin... I'll also be looking at working on "Cross Purposes" and "Forbidden"." Martin had suggested that this could coincide with the 30th anniversary of "The Eternal Idol", in 2017. In an interview that August, Martin added "[Iommi] still has his cancer issues of course and that may well stop it all from happening but if he wants to do something I am ready." On 10 August 2016, Iommi revealed that his cancer was in remission.
Asked in November 2016 about his plans after Black Sabbath's final tour, Iommi replied, "I'll be doing some writing. Maybe I'll be doing something with the guys, maybe in the studio, but no touring." The band played their final concert on 4 February 2017 in Birmingham. The final song was streamed live on the band's Facebook page and fireworks went off as the band took their final bow. The band's final tour was not an easy one, as longstanding tensions between Osbourne and Iommi returned to the surface. Iommi stated that he would not rule out the possibility of one-off shows, "I wouldn't write that off, if one day that came about. That's possible. Or even doing an album, 'cause then, again, you're in one place. But I don't know if that would happen." In an April 2017 interview, Butler revealed that Black Sabbath considered making a blues album as the follow-up to "13", but added that, "the tour got in the way."
On 7 March 2017, Black Sabbath announced their disbandment through posts made on their official social media accounts.
In a June 2018 interview with ITV News, Osbourne expressed interest in reuniting with Black Sabbath for a performance at the 2022 Commonwealth Games which is due to be held in their home city Birmingham. Iommi said that performing at the event as Black Sabbath would be "a great thing to do to help represent Birmingham. I'm up for it. Let's see what happens." He also did not rule out the possibility for the band to reform only for a one-off performance rather than a full-length tour.
Black Sabbath were a heavy metal band, whose music has also been described as psychedelic rock, and acid rock. The band have also been cited as a key influence on genres including stoner rock, grunge, doom metal, and sludge metal. Early on Black Sabbath were influenced by Cream, the Beatles, Fleetwood Mac, Jimi Hendrix, John Mayall & the Bluesbreakers, Led Zeppelin, and Jethro Tull.
Although Black Sabbath went through many line-ups and stylistic changes, their core sound focuses on ominous lyrics and doomy music, often making use of the musical tritone, also called the "devil's interval". While their Ozzy-era albums such as "Sabbath Bloody Sabbath" (1973) had slight compositional similarities to the progressive rock genre that was growing in popularity at the time, standing in stark contrast to popular music of the early 1970s Black Sabbath's dark sound was dismissed by rock critics of the era. Much like many of their early heavy metal contemporaries, the band received virtually no airplay on rock radio.
As the band's primary songwriter, Tony Iommi wrote the majority of Black Sabbath's music, while Osbourne would write vocal melodies, and bassist Geezer Butler would write lyrics. The process was sometimes frustrating for Iommi, who often felt pressured to come up with new material: "If I didn't come up with anything, nobody would do anything." On Iommi's influence, Osbourne later said:
Beginning with their third album, "Master of Reality" (1971), Black Sabbath began to feature tuned-down guitars. In 1965, before forming Black Sabbath, guitarist Tony Iommi suffered an accident while working in a sheet metal factory, losing the tips of two fingers on his right hand. Iommi almost gave up music, but was urged by the factory manager to listen to Django Reinhardt, a jazz guitarist who lost the use of two fingers in a fire. Inspired by Reinhardt, Iommi created two thimbles made of plastic and leather to cap off his missing fingertips. The guitarist began using lighter strings, and detuning his guitar, to better grip the strings with his prosthesis. Early in the band's history Iommi experimented with different dropped tunings, including C tuning, or 3 semitones down, before settling on E/D tuning, or a half-step down from standard tuning.
Black Sabbath has sold over 70 million records worldwide, including a RIAA-certified 15 million in the U.S. They are one of the most influential heavy metal bands of all time. The band helped to create the genre with ground-breaking releases such as "Paranoid" (1970), an album that "Rolling Stone" magazine said "changed music forever", and called the band "the Beatles of heavy metal". "Time" magazine called "Paranoid" "the birthplace of heavy metal", placing it in their Top 100 Albums of All Time.
MTV placed Black Sabbath at number one on their Top Ten Heavy Metal Bands and VH1 placed them at number two on their list of the 100 Greatest Artists of Hard Rock. VH1 ranked Black Sabbath's "Iron Man" the number one song on their 40 Greatest Metal Songs countdown. "Rolling Stone" magazine ranked the band number 85 in their list of the "100 Greatest Artists of All Time". AllMusic's William Ruhlmann said:
According to "Rolling Stone" Holly George-Warren, "Black Sabbath was the heavy metal king of the 1970s." Although initially "despised by rock critics and ignored by radio programmers", the group sold more than 8 million albums by the end of that decade. ""The" heavy metal band…" marvelled Ronnie James Dio. "A band that didn't apologise for coming to town; it just stepped on buildings when it came to town."
Black Sabbath have influenced many acts including Iron Maiden, Slayer, Metallica, Nirvana, Korn, Mayhem, Venom, Judas Priest, Guns N' Roses, Soundgarden, Body Count, Alice in Chains, Anthrax, Disturbed, Death, Opeth, Pantera, Megadeth, the Smashing Pumpkins, Slipknot, Foo Fighters, Fear Factory, Candlemass, Godsmack, and Van Halen. Two gold selling tribute albums have been released, "Nativity in Black Volume 1 & 2", including covers by Sepultura, White Zombie, Type O Negative, Faith No More, Machine Head, Primus, System of a Down, and Monster Magnet.
Metallica's Lars Ulrich, who, along with bandmate James Hetfield inducted Black Sabbath into the Rock and Roll Hall of Fame in 2006, said "Black Sabbath is and always will be synonymous with heavy metal", while Hetfield said "Sabbath got me started on all that evil-sounding shit, and it's stuck with me. Tony Iommi is the king of the heavy riff." Guns N' Roses guitarist Slash said of the "Paranoid" album: "There's just something about that whole record that, when you're a kid and you're turned onto it, it's like a whole different world. It just opens up your mind to another dimension..."Paranoid" is the whole Sabbath experience; very indicative of what Sabbath meant at the time. Tony's playing style—doesn't matter whether it's off "Paranoid" or if it's off "Heaven and Hell"—it's very distinctive." Anthrax guitarist Scott Ian said "I always get the question in every interview I do, 'What are your top five metal albums?' I make it easy for myself and always say the first five Sabbath albums."
Lamb of God's Chris Adler said: "If anybody who plays heavy metal says that they weren't influenced by Black Sabbath's music, then I think that they're lying to you. I think all heavy metal music was, in some way, influenced by what Black Sabbath did." Judas Priest vocalist Rob Halford commented: "They were and still are a groundbreaking band..you can put on the first Black Sabbath album and it still sounds as fresh today as it did 30-odd years ago. And that's because great music has a timeless ability: To me, Sabbath are in the same league as the Beatles or Mozart. They're on the leading edge of something extraordinary." On Black Sabbath's standing, Rage Against the Machine guitarist Tom Morello states: "The heaviest, scariest, coolest riffs and the apocalyptic Ozzy wail are without peer. You can hear the despair and menace of the working-class Birmingham streets they came from in every kick-ass, evil groove. Their arrival ground hippy, flower-power psychedelia to a pulp and set the standard for all heavy bands to come." Phil Anselmo of Pantera and Down stated that "Only a fool would leave out what Black Sabbath brought to the heavy metal genre".
According to Tracii Guns of L.A. Guns and former member of Guns N' Roses, the main riff of "Paradise City" by Guns N' Roses, from "Appetite for Destruction" (1987), was influenced by the song "Zero the Hero" from the "Born Again" album. King Diamond guitarist Andy LaRocque affirmed that the clean guitar part of "Sleepless Nights" from "Conspiracy" (1989) is inspired by Tony Iommi's playing on "Never Say Die!".
In addition to being pioneers of heavy metal, they also have been credited for laying the foundations for heavy metal subgenres stoner rock, sludge metal, thrash metal, black metal and doom metal as well as for alternative rock subgenre grunge. According to the critic Bob Gulla, the band's sound "shows up in virtually all of grunge's most popular bands, including Nirvana, Soundgarden, and Alice in Chains".
Tony Iommi has been credited as the pioneer of lighter gauge guitar strings. The tips of his fingers were severed in a steel factory, and while using thimbles (artificial finger tips) he found that standard guitar strings were too difficult to bend and play. He found that there was only one size of strings available, so after years with Sabbath he had strings custom made.
Culturally, Black Sabbath have exerted a huge influence in both television and literature and have in many cases become synonymous with heavy metal. In the film "Almost Famous", Lester Bangs gives the protagonist an assignment to cover the band (plot point one) with the immortal line: 'Give me 500 words on Black Sabbath'. Contemporary music and arts publication "Trebuchet Magazine" has put this to practice by asking all new writers to write a short piece (500 words) on Black Sabbath as a means of proving their creativity and voice on a well documented subject.
Original and classic line-up | https://en.wikipedia.org/wiki?curid=4314 |
Buffalo Bills
The Buffalo Bills are a professional American football team based in the Buffalo–Niagara Falls metropolitan area. The Bills compete in the National Football League (NFL) as a member club of the league's American Football Conference (AFC) East division. The team plays their home games at New Era Field in Orchard Park, New York. The Bills are the only NFL team that plays its home games in the state of New York. The Bills conduct summer training camp at St. John Fisher College in Pittsford, New York, an eastern suburb of Rochester.
The Bills began play as an original franchise of the American Football League (AFL) in 1960. The club joined the NFL as a result of the AFL–NFL merger for the 1970 season. The 1964 and 1965 Bills were the only teams representing Buffalo that won major league professional sports championships ("back-to-back" American Football League Championships). The Bills are the only team to win four consecutive conference championships and are the only NFL team to lose four consecutive Super Bowl games. The team was owned by Ralph Wilson from the team's founding in 1960, until his death in 2014 at the age of 95. After his death, Wilson's estate reached an agreement to sell the team to Terry and Kim Pegula, which was approved by the other NFL team owners on October 8, 2014. The Bills formerly possessed the longest active playoff drought in any of the four major professional sports in North America: they did not qualify to play in the NFL playoffs from 1999 until 2017 and were the last NFL team (and last team in the major North American professional sports leagues overall) to qualify for the playoffs for the first time in the 21st century. Since then, the Bills have reached the playoffs in two of the past three NFL seasons and also clinched their first 10-win season in two decades in December 2019.
The Bills began competitive play in 1960 as a charter member of the American Football League led by head coach Buster Ramsey and joined the NFL as part of the AFL–NFL merger in 1970. The Bills won two consecutive American Football League titles in 1964 and 1965 with quarterback Jack Kemp and coach Lou Saban, but the club has yet to win a league championship since.
Once the AFL–NFL merger took effect, the Bills became the second NFL team to represent the city; they followed the Buffalo All-Americans, a charter member of the league. Buffalo had been left out of the league since the All-Americans (by that point renamed the Bisons) folded in 1929; the Bills were no less than the third professional non-NFL team to compete in the city before the merger, following the Indians/Tigers of the early 1940s and an earlier team named the Bills, originally the Bisons, in the late 1940s in the All-America Football Conference (AAFC).
Following the AFL–NFL merger, the Bills were generally mediocre in the 1970s, but featured All-Pro running back O. J. Simpson. After being pushed to the brink of failure in the mid-1980s, the collapse of the United States Football League and a series of highly drafted players such as Jim Kelly (who initially played for the USFL instead of the Bills), Thurman Thomas, Bruce Smith and Darryl Talley allowed the Bills to rebuild into a perennial contender in the late 1980s through the mid-1990s, a period in which the team won four consecutive AFC Championships; the team nevertheless lost all four subsequent Super Bowls, records in both categories that still stand.
The rise of the division rival New England Patriots under Tom Brady, along with numerous failed attempts at rebuilding in the 2000s and 2010s, prevented the Bills from reaching the playoffs in seventeen consecutive seasons between 2000 and 2016, a 17-year drought that was the longest active playoff drought in all major professional sports at the time. It was broken when the Bills secured a wild-card berth on December 31, 2017. On October 8, 2014, Buffalo Sabres owners Terry and Kim Pegula received unanimous approval to acquire the Bills during the NFL owners' meetings, becoming the second ownership group of the team after team founder Ralph Wilson.
In 1947 a contest was held to rename the AAFC Bisons, which was owned by James Breuil of the Frontier Oil Company. The winning entry suggested "Bills", reflecting on the famous western frontiersman, Buffalo Bill Cody. Carrying the "frontier" theme further, the winning contestant offered the team was being supported by Frontier Oil and was "opening a new frontier in sports in Western New York." When Buffalo joined the new American Football League in 1960, the name of the city's earlier pro football entry was adopted.
The Bills' uniforms in its first two seasons were based on those of the Detroit Lions at the time.
The team's original colors were Honolulu blue, silver and white, and the helmets were silver with no striping. There was no logo on the helmet, which displayed the players' numbers on each side.
In 1962, the standing red bison was designated as the logo and took its place on a white helmet. In 1962, the team's colors also changed to red, white, and blue. The team switched to blue jerseys with red and white shoulder stripes similar to those worn by the Buffalo Bisons AHL hockey team of the same era. The helmets were white with a red center stripe. The jerseys again saw a change in 1964 when the shoulder stripes were replaced by a distinctive stripe pattern on the sleeves consisting of four stripes, two thicker inner stripes and two thinner outer stripes all bordered by red piping. By 1965, red and blue center stripes were put on the helmets.
The Bills introduced blue pants worn with the white jerseys in 1973, the last year of the standing buffalo helmet. The blue pants remained through 1985. The face mask on the helmet was blue from 1974 through 1986 before changing to white.
The standing bison logo was replaced by a blue charging one with a red slanting stripe streaming from its horn. The newer emblem, which is still the primary one used by the franchise, was designed by aerospace designer Stevens Wright in 1974.
In 1984, the helmet's shell color was changed from white to red, primarily to help Bills quarterback Joe Ferguson distinguish them more readily from three of their division rivals at that time, the Baltimore Colts, the Miami Dolphins, and the New England Patriots, who all also wore white helmets at that point. Ferguson said "Everyone we played had white helmets at that time. Our new head coach Kay Stephenson just wanted to get more of a contrast on the field that may help spot a receiver down the field." (The Patriots have worn silver helmets since 1993, the Colts have since been realigned to the AFC South, and the New York Jets have since switched back to green-colored helmets after 20 years with white ones in 2019.)
In 2002, under the direction of general manager Tom Donahoe, the Bills' uniforms went through radical changes. A darker shade of blue was introduced as the main jersey color, and nickel gray was introduced as an accent color. Both the blue and white jerseys featured red side panels. The white jerseys included a dark blue shoulder yoke and royal blue numbers. The helmet remained primarily red with one navy blue, two nickel, two royal blue, two white stripes, and white face mask. A new logo, a stylized "B" consisting of two bullets and a more detailed buffalo head on top, was proposed and had been released (it can be seen on a few baseball caps that were released for sale), but fan backlash led to the team retaining the running bison logo. The helmet logo adopted in 1974—a charging royal blue bison, with a red streak, white horn and eyeball—remained unchanged.
In 2005, the Bills revived the standing bison helmet and uniform of the mid-1960s as a throwback uniform.
The Bills usually wore the all-blue combination at home and the all-white combination on the road when not wearing the throwback uniforms. They stopped wearing blue-on-white after 2006, while the white-on-blue was not worn after 2007.
For the 2011 season, the Bills unveiled a new uniform design, an updated rendition of the 1975–83 design. This change includes a return to the white helmets with "charging buffalo" logo, and a return to royal blue instead of navy.
Buffalo sporadically wore white at home in the 1980s, but stopped doing so before their Super Bowl years. On November 6, 2011, against the New York Jets, the Bills wore white at home for the first time since 1986. Since 2011, the Bills have worn white for a home game either with their primary uniform or a throwback set.
The Bills' uniform received minor alterations as part of the league's new uniform contract with Nike. The new Nike uniform was unveiled on April 3, 2012.
On November 12, 2015, the Bills and the New York Jets became the first two teams to participate in the NFL's Color Rush uniform initiative, with Buffalo wearing an all-red combination for the first time in team history.
A notable use of the Bills' uniforms outside of football was in the 2018 World Junior Ice Hockey Championships, when the United States men's national junior ice hockey team wore Bills-inspired uniforms in their outdoor game against Team Canada on December 29, 2017.
The Bills have rivalries with their three AFC East opponents, and also have had occasional or historical rivalries with other teams such as the Baltimore/Indianapolis Colts (a former divisional rival), Kansas City Chiefs, Houston Oilers/Tennessee Titans, Cleveland Browns, and Dallas Cowboys. They also play an annual preseason game against the Detroit Lions.
This is often considered Buffalo's most famous rivalry. Though the Bills and Dolphins both originated in the American Football League, the Dolphins did not start playing until 1966 as an expansion team while the Bills were one of the original eight teams. The rivalry first gained prominence when the Dolphins won every match-up against the Bills in the 1970s for an NFL-record 20 straight wins against a single opponent (the Bills defeated the Dolphins in their first matchup of the 1980s). Fortunes changed in the following decades with the rise of Jim Kelly as Buffalo's franchise quarterback, and though Kelly and Dolphins quarterback Dan Marino shared a competitive rivalry in the 1980s and 1990s, the Bills became dominant in the 1990s. Things have since cooled down after the retirements of Kelly and Marino and the rise of the New England Patriots, but Miami remains a fierce rival of the Bills, coming in second place in a recent poll of Buffalo's primary rival, and the two teams have typically been close to each other in win-loss records. Miami leads the overall series 61–48–1 as of 2018, but Buffalo has the advantage in the playoffs at 3–1, including a win in the 1992 AFC Championship Game.
The rivalry with the New England Patriots first started when both teams were original franchises in the American Football League prior to the NFL-AFL merger. After the rise of head coach Bill Belichick and quarterback Tom Brady in New England, the Patriots have dominated the AFC East, including the Bills. The Bills-Patriots rivalry in particular has become lopsided as the Patriots are 31–5 against the Bills since Belichick became head coach in 2000. This has led many fans and players in the 2000s and beyond to replace the Dolphins with the Patriots as Buffalo's most hated rival. Overall, the Patriots lead the series 76–43–1 as of 2019.
The rivalry is also notable as numerous players, including Drew Bledsoe, Doug Flutie, Lawyer Milloy, Brandon Spikes, Scott Chandler, Chris Hogan, Mike Gillislee and Stephon Gilmore have played for both teams at some point in their careers.
The Bills and Jets were both original AFL teams, and both represent the state of New York, though the Jets (since 1984) actually play their games in East Rutherford, New Jersey. While the rivalry represents the differences between New York City and Western New York, it has historically not been as intense as the Bills' rivalries with the Dolphins and Patriots, and the teams' fanbases either have grudging respect or low-key annoyance for each other when the teams are not playing one another. Oftentimes the Bills-Jets rivalry has become characterized by ugly games and shared mediocrity, but it has had a handful of competitive moments. The series heated up recently when former Jets head coach Rex Ryan became the Bills' head coach for two seasons, and has become notable again as Bills quarterback Josh Allen and Jets quarterback Sam Darnold, both drafted in the same year, maintain a friendly rivalry with one another. Buffalo leads the series 62–55 as of 2018, including a playoff win in 1981.
The Tennessee Titans (formerly the Houston Oilers) share an extended history with the Bills, both teams being original AFL clubs in 1960 and rivals in that league's East Division before the AFL-NFL merger. Matchups were intense in the 1990s with quarterback Warren Moon leading the Oilers against Jim Kelly's Bills. Memorable playoff moments between the teams include The Comeback, in which the Frank Reich-led Bills overcame a 35–3 deficit to stun the Oilers 41–38 in 1992, and the Music City Miracle, in which the now-Titans scored on a near-last-minute kickoff return with a controversial lateral pass to stun the Bills 22–16 in 1999. The Music City Miracle was notable for being Buffalo's last playoff appearance until 2017. The Titans currently lead the series 28–19.
Though the Bills and Cleveland Browns are in different divisions and did not start playing each other until after 1972, match-ups between the two teams occasionally get heated up due to the proximity and similarity between the cities of Buffalo and Cleveland, though the teams' fanbases otherwise share a mutual respect. As with the Bills–Jets rivalry, the Bills and Browns often share bad luck and have seen their share of ugly games, including a 6–3 Browns win in which the winning quarterback only completed 2 of 17 passes. However, there have been other occasions when both teams have been competitive such as in the 1980s and most recently in 2007 and 2014. The rivalry also gained heat when former Bills safety Donte Whitner was with the Browns. The Browns currently lead the series 13–9, including a playoff win in 1990, though the Bills have outscored the Browns in the series.
The Browns shared a rivalry with the Bills' predecessors in the All-America Football Conference, playing them twice in the AAFC playoffs before becoming one of three AAFC teams to join the NFL. The Bills were not selected to join the NFL and folded with the rest of the AAFC, leaving Buffalo without professional football until the current Bills were formed in 1959.
The Bills and the Kansas City Chiefs were also original teams in the AFL and have had a long history against each other, despite never being in the same division. This rivalry heated up recently as the Bills and Chiefs met in consecutive years from 2008 to 2015, and again in 2017. The teams have played three playoff games against each other, including the AFL Championship game that determined the AFL's (later AFC) representative in the first Super Bowl, with Kansas City winning and going on to face the Green Bay Packers in the Super Bowl. However, Buffalo defeated Kansas City in the 1993 AFC championship game to advance to its fourth straight Super Bowl appearance. Buffalo currently leads the series 26-21-1.
In recent years, bitterness has emerged between the Bills and the Jacksonville Jaguars, who had handed Buffalo its first playoff loss in New Era Field in 1996. Both teams occupy two of the smallest media markets in the NFL. After years of concurrent bottom feeding in the late 2000s and early 2010s, this rivalry has emerged after former Bills head coach Doug Marrone, who had quit on the team after the 2014 season, was hired as a coaching assistant for Jacksonville and rose to become the Jaguars' head coach. Since then, the series has featured a Bills loss to the Jaguars in London, an ugly, low-scoring playoff game in 2017, trash talk from Jaguars players such as Jalen Ramsey and a brawl between the teams in 2018.
The Buffalo Bills have retired three numbers in franchise history: No. 12 for Jim Kelly, No. 34 for Thurman Thomas and No. 78 for Bruce Smith. Despite the fact that the Bills have only retired three jersey numbers in franchise history, the team has other numbers no longer issued to any player or in reduced circulation.
Since the earliest days of the team, the number 31 was not supposed to be issued to any other player. The Bills had stationery and various other team merchandise showing a running player wearing that number, and it was not supposed to represent any specific person, but the 'spirit of the team.' In the first three decades of the team's existence, the number 31 was only seen once: in 1969, when reserve running back Preston Ridlehuber damaged his number 36 jersey during a game, equipment manager Tony Marchitte gave him the number 31 jersey to wear while repairing the number 36. The number 31 was not issued again until 1990 when first round draft choice James (J.D.) Williams wore it for his first two seasons; it has since been returned to general circulation, with safety Dean Marlowe wearing the number in 2019.
Number 32 had been withdrawn from circulation, but not retired, after O. J. Simpson. Former owner Ralph Wilson insisted on not reissuing the number, even after Simpson's highly publicized murder case and later robbery conviction. The number was placed back into circulation in 2019 with Senorise Perry wearing the number that year.
Number 15 was historically only issued sparingly after the retirement of Jack Kemp, but was later returned to general circulation. Receiver John Brown wears the number as of 2019.
Number 1 has also only rarely been used, for reasons never explained. Kicker Mike Hollis, who played one season for the Bills in 2002, was the most recent to wear the number in the regular season; it went 17 years without being reissued before David Sills was assigned it during the 2019 preseason.
The Buffalo Bills Radio Network is flagshipped at WGR, AM 550 in Buffalo, owned by Entercom. John Murphy is the team's current play-by-play announcer; he was a color commentator alongside, and eventually succeeded, longtime voice Van Miller after Miller's retirement at the end of the 2003 NFL season. Mark Kelso serves as the color analyst. The Bills radio network has approximately seventeen affiliates in upstate New York and one affiliate, CJCL 590AM (The Fan) in Toronto. As of early 2012, it is composed mostly of WGR, Entercom's sister stations WCMF (96.5 FM) and WROC-AM 950 in Rochester, and a fleet of independent AM and FM stations across upstate New York from Jamestown east to Albany. Previous flagship Citadel Broadcasting was purchased by Cumulus Media, who in turn ceased carrying Bills games at the end of the 2011 season, leaving the network without affiliates in Syracuse, Binghamton, and Erie. (The Syracuse affiliations were later picked up by Galaxy Communications.)
In 2018, the team signed an agreement with Nexstar Media Group to carry Bills preseason games across its network of stations in the region. As of 2020, WIVB-TV serves as the flagship station of the network, which includes WJET-TV in Erie, WROC-TV in Rochester, WSYR-TV in Syracuse, WUTR in Utica, WETM-TV in Elmira and WIVT in Binghamton. Steve Tasker does color commentary on these games; the play-by-play position is rotated between Andrew Catalon and Rob Stone. WROC-TV reporter Thad Brown is the sideline reporter. Since 2008, preseason games have been broadcast in high definition.
Beginning in the 2016 season, as per a new rights deal which covers rights to the team as well as its sister NHL franchise, the Buffalo Sabres, most team-related programming, including studio programming and the coach's show, was re-located to MSG Western New York—a joint venture of MSG and the team ownership. Preseason games will continue to air in simulcast on WKBW.
In the event regular season games are broadcast by ESPN, in accordance with the league's television policies, a local Buffalo station simulcasts the game. From 2014 to 2017, WKBW-TV held the broadcast rights to that contest, with the station having won back the rights to cable games after WBBZ-TV held the rights for 2012 and 2013.
Source:
The Bills' official mascot is Billy Buffalo, an eight-foot tall, anthropomorphic blue American bison who wears the jersey "number" BB.
The Bills currently do not have cheerleaders. The Bills operated a cheerleading squad named the Buffalo Jills from 1967 to 1985; from 1986 to 2013, the Jills operated as an independent organization sponsored by various companies, most recently by Citadel Broadcasting. The Jills suspended operations prior to the 2014 season due to legal actions. The Bills and Jills are currently involved in a legal battle, in which the Jills allege they were employees, not independent contractors, and are seeking back pay. Complicating matters is that Citadel's buyer, Cumulus Media, declared bankruptcy and sought to discharge its remaining Bills-related debts in January 2018.
The Bills are one of six teams in the NFL to designate an official marching band or drumline (the others being the Baltimore Ravens, Washington Redskins, New York Jets, Carolina Panthers and Seattle Seahawks). Since the last game of the 2013 season, this position has been served by the Stampede Drumline, known outside of Buffalo as Downbeat Percussion. The Bills have also used the full marching bands from Attica High School, the University of Pittsburgh and Syracuse University at home games in recent years.
The Bills have several theme songs associated with them. The most popular is a variation of the Isley Brothers hit "Shout", recorded by Scott Kemper, which served as the Bills' official promotional song from 1987 through 1990s. It was officially replaced circa 2000 with "The Power of the Bills", although "Shout" remains in usage today. It can be heard at every Bills home game following a field goal or touchdown and at the end of the game if the Bills win. The Bills' unofficial fight song, "Go Bills", was penned by Bills head coach Marv Levy in the mid-1990s on a friendly wager with his players that he will write the song if the team won a particular game.
The Bills Backers are the official fan organization of the Buffalo Bills. It has over 200 chapters across North America, Europe and Oceania. Also notable is the Bills Mafia, a collection of Bills fans organized via Twitter beginning in 2010; the phrase "Bills Mafia" had by 2017 grown to unofficially represent the broad community surrounding and encompassing the team as a whole, and players who join the Bills often speak of joining the Bills Mafia. Outsiders often treat the Bills' fan base in derogatory terms, especially since the 2010s, in part because of negative press coverage of select fans' wilder antics. Bills fans have been noted as much for their boisterous behavior as they have for their generosity; after the Bills received help in breaking their 17-year playoff drought on a last-minute Cincinnati Bengals victory, Bills fans crowdfunded the charities of Bengals players Andy Dalton and Tyler Boyd with hundreds of thousands of dollars as a gesture of thanks.
The Bills are one of the favorite teams of ESPN announcer Chris Berman, who picked the Bills to reach the Super Bowl nearly every year in the 1990s. Berman often uses the catchphrase "No one circles the wagons like the Buffalo Bills!" Berman gave the induction speech for Bills owner Ralph Wilson when Wilson was inducted into the Pro Football Hall of Fame in 2009.
The Bills were also the favorite team of late NBC political commentator Tim Russert, a South Buffalo native, who often referred to the Bills on his Sunday morning talk show, "Meet the Press". (His son, Luke, is also a notable fan of the team.) CNN's Wolf Blitzer, also a Buffalo native, has proclaimed he is also a fan, as has "CBS Evening News" lead anchor and Tonawanda native Jeff Glor and DNC Chairman Tom Perez.
ESPN anchor Kevin Connors is also a noted Bills fan, dating to his time attending Ithaca College. Actor Nick Bakay, a Buffalo native, is also a well-known Bills fan; he has discussed the team in segments of "NFL Top 10". Character actor William Fichtner, raised in Cheektowaga, is a fan, and did a commercial for the team in 2014. In 2015, Fichtner also narrated the ESPN "30 for 30" documentary on the Bills' four Super Bowl appearances, "Four Falls of Buffalo". Former Olympic swimmer Summer Sanders (an in-law to former Bills kicker Todd Schlopy) has professed her fandom of the team. Actor Christopher McDonald, who was raised in Romulus, New York, is a fan of the team.
Bills fans are particularly well known for their wearing of Zubaz zebra-printed sportswear; so much is the association between Bills fans and Zubaz that when a revival of the company opened their first brick-and-mortar storefront, it chose Western New York as its first location.
Bills Fans are also well known for jumping off of elevated surfaces (often cars or RVs) into folding tables during the pre game Tailgate.
Persons notable almost entirely for their Bills fandom include Ken "Pinto Ron" Johnson, whose antics while appearing at every Bills home and away game since 1994 earned enough scrutiny that his tailgate parties were banned from stadium property on order of the league; John Lang, an Elvis impersonator who carries a large guitar that he uses as a billboard; Marc Miller, whose professional wrestling promo-style interview with WGRZ prior to Super Bowl XXVII (distinguished by the line "Dallas is going down, Gary!" and picked up at the time by "The George Michael Sports Machine") was rediscovered in 2019; and Ezra Castro, also known as "Pancho Billa," a resident of the Dallas-Fort Worth Metroplex who wore a large sombrero and lucha mask in Bills colors. Castro was diagnosed with a spinal tumor that had metastasized in 2017; he was invited on stage during the 2018 NFL Draft to read one of the Bills' selections. Castro died May 14, 2019.
Several former Buffalo Bills players have earned a name in politics after their playing careers had ended, almost always as members of the Republican Party. The most famous of these was quarterback Jack Kemp, who was elected to the U.S. House of Representatives from Western New York in 1971—two years after his playing career ended and remained there for nearly two decades, serving as the Republican Party nominee for Vice President of the United States under Bob Dole in 1996. Kemp's backup, Ed Rutkowski, served as county executive of Erie County from 1979 to 1987. Former tight end Jay Riemersma, defensive tackle Fred Smerlas and defensive end Phil Hansen have all run for Congress, though all three either lost or withdrew from their respective races. Quarterback Jim Kelly and running back Thurman Thomas have also both been mentioned as potential candidates for political office, although both have declined all requests to date. | https://en.wikipedia.org/wiki?curid=4315 |
Big Dig
The Central Artery/Tunnel Project (CA/T), commonly known as the Big Dig, was a megaproject in Boston that rerouted the Central Artery of Interstate 93 (I-93), the chief highway through the heart of the city, into the 1.5-mile (2.4 km) tunnel named the Thomas P. O'Neill Jr. Tunnel. The project also included the construction of the Ted Williams Tunnel (extending I-90 to Logan International Airport), the Leonard P. Zakim Bunker Hill Memorial Bridge over the Charles River, and the Rose Kennedy Greenway in the space vacated by the previous I-93 elevated roadway. Initially, the plan was also to include a rail connection between Boston's two major train terminals. Planning began in 1982; the construction work was carried out between 1991 and 2006; and the project concluded on December 31, 2007 when the partnership between the program manager and the Massachusetts Turnpike Authority ended.
The Big Dig was the most expensive highway project in the US, and was plagued by cost overruns, delays, leaks, design flaws, charges of poor execution and use of substandard materials, criminal arrests, and the death of one motorist. The project was originally scheduled to be completed in 1998 at an estimated cost of $2.8 billion (in 1982 dollars, US$7.4 billion adjusted for inflation ). However, the project was completed in December 2007 at a cost of over $8.08 billion (in 1982 dollars, $21.5 billion adjusted for inflation, meaning a cost overrun of about 190%) . "The Boston Globe" estimated that the project will ultimately cost $22 billion, including interest, and that it would not be paid off until 2038. As a result of a death, leaks, and other design flaws, Bechtel and Parsons Brinckerhoff—the consortium that oversaw the project—agreed to pay $407 million in restitution and several smaller companies agreed to pay a combined sum of approximately $51 million.
The Rose Fitzgerald Kennedy Greenway is a roughly series of parks and public spaces, which were the final part of the Big Dig after Interstate 93 was put underground. The Greenway was named in honor of Kennedy family matriarch Rose Fitzgerald Kennedy, and was officially dedicated on July 26, 2004.
This project was developed in response to traffic congestion on Boston's historically tangled streets which were laid out long before the advent of the automobile. As early as 1930 the city's Planning Board recommended a raised express highway running north–south through the downtown district in order to draw traffic off the city streets. Commissioner of Public Works William Callahan promoted plans for an elevated expressway which eventually was constructed between the downtown area and the waterfront. Governor John Volpe interceded in the 1950s to change the design of the last section of the Central Artery putting it underground through the Dewey Square Tunnel. While traffic moved somewhat better, the other problems remained. There was chronic congestion on the Central Artery (I-93), an elevated six-lane highway through the center of downtown Boston, which was, in the words of Pete Sigmund, "like a funnel full of slowly-moving, or stopped, cars (and swearing motorists)."
In 1959, the 1.5-mile-long (2.4 km) road section carried approximately 75,000 vehicles a day, but by the 1990s, this had grown to 190,000 vehicles a day. Traffic jams of 16 hours were predicted for 2010.
The expressway had tight turns, an excessive number of entrances and exits, entrance ramps without merge lanes, and as the decades passed, had continually escalating vehicular traffic that was well beyond its design capacity. Local businesses again wanted relief, city leaders sought a reuniting of the waterfront with the city, and nearby residents desired removal of the matte green-painted elevated road which mayor Thomas Menino called Boston's "other Green Monster". MIT engineers Bill Reynolds and (eventual state Secretary of Transportation) Frederick P. Salvucci envisioned moving the whole expressway underground.
Another important motivation for the final form of the Big Dig was the abandonment of the Massachusetts Department of Public Works' intended expressway system through and around Boston. The Central Artery, as part of Mass. DPW's Master Plan of 1948, was originally planned to be the downtown Boston stretch of Interstate 95, and was signed as such; a bypass road called the Inner Belt, was subsequently renamed Interstate 695. (The law establishing the Interstate highway system was enacted in 1956.) The Inner Belt District was to pass to the west of the downtown core, through the neighborhood of Roxbury and the cities of Brookline, Cambridge, and Somerville. Earlier controversies over impact of the Boston extension of the Massachusetts Turnpike, particularly on the heavily populated neighborhood of Brighton, and the additional large amount of housing that would have had to be destroyed led to massive community opposition to both the Inner Belt and the Boston section of I-95.
Building demolition and land clearances for I-95 through the neighborhoods of Roxbury, Jamaica Plain, and Roslindale led to secession threats by Hyde Park, Boston's youngest and southernmost neighborhood. By 1972, with only a minimum of work done on the I-95 right of way and none on the potentially massively disruptive Inner Belt, Governor Francis Sargent put a moratorium on highway construction within the Route 128 corridor, except for the final short stretch of Interstate 93. In 1974, the remainder of the Master Plan was canceled, leaving Boston with a severely overstressed expressway system for the existing traffic.
With ever-increasing traffic volumes funneled onto I-93 alone, the Central Artery became chronically gridlocked. The Sargent moratorium led to the rerouting of I-95 away from Boston around the Route 128 beltway and the conversion of the cleared land in the southern part of the city into the Southwest Corridor linear park, as well as a new right-of-way for the Orange Line subway and Amtrak. Parts of the planned I-695 right-of-way remain unused and under consideration for future mass-transit projects.
The original 1948 Master Plan included a Third Harbor Tunnel plan that was hugely controversial in its own right, because it would have disrupted the Maverick Square area of East Boston. It was never built.
A major reason for the all-day congestion was that the Central Artery carried not only north–south traffic, but it also carried east–west traffic. Boston's Logan Airport lies across Boston Harbor in East Boston; and before the Big Dig, the only access to the airport from downtown was through the paired Callahan and Sumner tunnels. Traffic on the major highways from west of Boston—the Massachusetts Turnpike and Storrow Drive—mostly traveled on portions of the Central Artery to reach these tunnels. Getting between the Central Artery and the tunnels involved short diversions onto city streets, increasing local congestion.
A number of public transportation projects were included as part of an environmental mitigation for the Big Dig. The most expensive was the building of the Phase II Silver Line tunnel under Fort Point Channel, done in coordination with Big Dig construction. Silver Line buses now use this tunnel and the Ted Williams Tunnel to link South Station and Logan Airport.
, promised projects to extend the Green Line beyond Lechmere, to connect the Red and Blue subway lines, and to restore the Green Line streetcar service to the Arborway in Jamaica Plain have not been completed. Construction of the extension beyond Lechmere has begun. The Red and Blue subway line connection underwent initial design, but no funding has been designated for the project. The Arborway Line restoration has been abandoned, following a final court decision in 2011.
The original Big Dig plan also included the North-South Rail Link, which would have connected North and South Stations (the major passenger train stations in Boston), but this aspect of the project was ultimately dropped by the state transportation administration early in the Dukakis administration. Negotiations with the federal government had led to an agreement to widen some of the lanes in the new harbor tunnel, and accommodating these would require the tunnel to be deeper and mechanically-vented; this left no room for the rail lines, and having diesel trains (then in use) passing through the tunnel would have substantially increased the cost of the ventilation system.
The project was conceived in the 1970s by the Boston Transportation Planning Review to replace the rusting elevated six-lane Central Artery. The expressway separated downtown from the waterfront, and was increasingly choked with bumper-to-bumper traffic. Business leaders were more concerned about access to Logan Airport, and pushed instead for a third harbor tunnel. In their second terms, Michael Dukakis (governor) and Fred Salvucci (secretary of transportation) came up with the strategy of tying the two projects together—thereby combining the project that the business community supported with the project that they and the City of Boston supported.
Planning for the Big Dig as a project officially began in 1982, with environmental impact studies starting in 1983. After years of extensive lobbying for federal dollars, a 1987 public works bill appropriating funding for the Big Dig was passed by the US Congress, but it was vetoed by President Ronald Reagan for being too expensive. When Congress overrode the veto, the project had its green light and ground was first broken in 1991.
In 1997, the state legislature created the Metropolitan Highway System and transferred responsibility for the Central Artery and Tunnel "CA/T" Project from the Massachusetts Highway Department and the Massachusetts Governor's Office to the Massachusetts Turnpike Authority (MTA).
The MTA, which had little experience in managing an undertaking of the scope and magnitude of the CA/T Project, hired a joint venture to provide preliminary designs, manage design consultants and construction contractors, track the project's cost and schedule, advise MTA on project decisions, and (in some instances) act as the MTA's representative. Eventually, MTA combined some of its employees with joint venture employees in an integrated project organization. This was intended to make management more efficient, but it hindered MTA's ability to independently oversee project activities because MTA and the joint venture had effectively become partners in the project.
In addition to political and financial difficulties, the project faced several environmental and engineering obstacles.
The downtown area through which the tunnels were to be dug was largely land fill, and included existing Red Line and Blue Line subway tunnels as well as innumerable pipes and utility lines that would have to be replaced or moved. Tunnel workers encountered many unexpected geological and archaeological barriers, ranging from glacial debris to foundations of buried houses and a number of sunken ships lying within the reclaimed land.
The project received approval from state environmental agencies in 1991, after satisfying concerns including release of toxins by the excavation and the possibility of disrupting the homes of millions of rats, causing them to roam the streets of Boston in search of new housing. By the time the federal environmental clearances were delivered in 1994, the process had taken some seven years, during which time inflation greatly increased the project's original cost estimates.
Reworking such a busy corridor without seriously restricting traffic flow required a number of state-of-the-art construction techniques. Because the old elevated highway (which remained in operation throughout the construction process) rested on pylons located throughout the designated dig area, engineers first utilized slurry wall techniques to create concrete walls upon which the highway could rest. These concrete walls also stabilized the sides of the site, preventing cave-ins during the continued excavation process.
The multi-lane Interstate highway also had to pass under South Station's seven railway tracks, which carried over 40,000 commuters and 400 trains per day. To avoid multiple relocations of train lines while the tunneling advanced, as had been initially planned, a specially designed jack was constructed to support the ground and tracks to allow the excavation to take place below. Construction crews also used ground freezing (an artificial induction of permafrost) to help stabilize surrounding ground as they excavated the tunnel. This was the largest tunneling project undertaken beneath railway lines anywhere in the world. The ground freezing enabled safer, more efficient excavation, and also assisted in environmental issues, as less contaminated fill needed to be exported than if a traditional cut-and-cover method had been applied.
Other challenges included existing subway tunnels crossing the path of the underground highway. To build slurry walls past these tunnels, it was necessary to dig beneath the tunnels and to build an underground concrete bridge to support the tunnels' weight, without interrupting rail service.
The project was managed by the Massachusetts Turnpike Authority, with the Big Dig and the Turnpike's Boston Extension from the 1960s being financially and legally joined by the legislature as the Metropolitan Highway System. Design and construction was supervised by a joint venture of Bechtel Corporation and Parsons Brinckerhoff. Because of the enormous size of the project—too large for any company to undertake alone—the design and construction of the Big Dig was broken up into dozens of smaller subprojects with well-defined interfaces between contractors. Major heavy-construction contractors on the project included Jay Cashman, Modern Continental, Obayashi Corporation, Perini Corporation, Peter Kiewit Sons' Incorporated, J. F. White, and the Slattery division of Skanska USA. (Of those, Modern Continental was awarded the greatest gross value of contracts, joint ventures included.)
The nature of the Charles River crossing had been a source of major controversy throughout the design phase of the project. Many environmental advocates preferred a river crossing entirely in tunnels, but this, along with 27 other plans, was rejected as too costly. Finally, with a deadline looming to begin construction on a separate project that would connect the Tobin Bridge to the Charles River crossing, Salvucci overrode the objections and chose a variant of the plan known as "Scheme Z". This plan was considered to be reasonably cost-effective, but had the drawback of requiring highway ramps stacked up as high as immediately adjacent to the Charles River.
The city of Cambridge objected to the visual impact of the chosen Charles River crossing design. The city sued to revoke the project's environmental certificate and forced the project planners to redesign the river crossing again.
Swiss engineer Christian Menn took over the design of the bridge. He suggested a cradle cable-stayed bridge that would carry ten lanes of traffic. The plan was accepted and construction began on the Leonard P. Zakim Bunker Hill Memorial Bridge. The bridge employed an asymmetrical design and a hybrid of steel and concrete was used to construct it. The distinctive bridge is supported by two forked towers connected to the span by cables and girders. It was the first bridge in the country to employ this method and it was, at the time, the widest cable-stayed bridge in the world, having since been surpassed by the Eastern span replacement of the San Francisco–Oakland Bay Bridge, although that structure is actually a self-anchored suspension bridge, not a cable-stayed bridge.
Meanwhile, construction continued on the Tobin Bridge approach. By the time all parties agreed on the I-93 design, construction of the Tobin connector (today known as the "City Square Tunnel" for a Charlestown area it bypasses) was far along, significantly adding to the cost of constructing the US Route 1 interchange and retrofitting the tunnel.
Boston blue clay and other soils extracted from the path of the tunnel were used to cap many local landfills, fill in the Granite Rail Quarry in Quincy, and restore the surface of Spectacle Island in the Boston Harbor Islands National Recreation Area.
The Storrow Drive Connector, a companion bridge to the Zakim, began carrying traffic from I-93 to Storrow Drive in 1999. The project had been under consideration for years, but was opposed by the wealthy residents of the Beacon Hill neighborhood. However, it finally was accepted because it would funnel traffic bound for Storrow Drive and downtown Boston away from the mainline roadway. The Connector ultimately used a pair of ramps that had been constructed for Interstate 695, enabling the mainline I-93 to carry more traffic that would have used I-695 under the original Master Plan.
When construction began, the project cost, including the Charles River crossing, was estimated at $5.8 billion. Eventual cost overruns were so high that the chairman of the Massachusetts Turnpike Authority, James Kerasiotes, was fired in 2000. His replacement had to commit to an $8.55 billion cap on federal contributions. The total expenses eventually passed $15 billion. Interest brought this cost to $21.93 billion.
Several unusual engineering challenges arose during the project, requiring unusual solutions and methods to address them.
At the beginning of the project, engineers had to figure out the safest way to build the tunnel without endangering the existing elevated highway above. Eventually, they created horizontal braces as wide as the tunnel, then cut away the elevated highway's struts, and lowered it onto the new braces.
On January 18, 2003, the opening ceremony was held for the I-90 Connector Tunnel, extending the Massachusetts Turnpike (Interstate 90) east into the Ted Williams Tunnel, and onwards to Boston Logan International Airport. The Ted Williams tunnel had been completed and was in limited use for commercial traffic and high-occupancy vehicles since late 1995. The westbound lanes opened on the afternoon of January 18 and the eastbound lanes on January 19.
The next phase, moving the elevated Interstate 93 underground, was completed in two stages: northbound lanes opened in March 2003 and southbound lanes (in a temporary configuration) on December 20, 2003. A tunnel underneath Leverett Circle connecting eastbound Storrow Drive to I-93 North and the Tobin Bridge opened December 19, 2004, easing congestion at the circle. All southbound lanes of I-93 opened to traffic on March 5, 2005, including the left lane of the Zakim Bridge, and all of the refurbished Dewey Square Tunnel.
By the end of December 2004, 95% of the Big Dig was completed. Major construction remained on the surface, including construction of final ramp configurations in the North End and in the South Bay interchange, and reconstruction of the surface streets.
The final ramp downtown—exit 20B from I-93 south to Albany Street—opened January 13, 2006.
In 2006, the two Interstate 93 tunnels were dedicated as the Thomas P. O'Neill Jr. Tunnel, after the former Democratic speaker of the House of Representatives from Massachusetts who pushed to have the Big Dig funded by the federal government.
The Commonwealth of Massachusetts was required under the Federal Clean Air Act to mitigate air pollution generated by the highway improvements. Secretary of Transportation Fred Salvucci signed an agreement with the Conservation Law Foundation in 1990 enumerating 14 specific projects the state agreed to build. This list was affirmed in a 1992 lawsuit settlement.
Projects which have been completed include:
As of 2014, several mitigation projects were incomplete:
Some projects, such as restoration of Green Line "E" Arborway service, were removed from the list of mitigation projects and replaced with other projects with similar air-quality improvements.
Some surface treatments that were part of the original project plan, were dropped due to the massive cost overruns on the highway portion of the project.
$99.1 million was allocated for mitigating improvements to the Charles River Basin, including the construction of North Point Park in Cambridge and Paul Revere Park in Charlestown. The North Bank Bridge, providing pedestrian and bicycle connectivity between the parks, was not funded until the American Recovery and Reinvestment Act of 2009. Nashua Street Park on the Boston side was completed in 2003, by McCourt Construction with $7.9 million in funding from MassDOT. As of 2017, $30.5 million had been transferred to the Massachusetts Department of Conservation and Recreation to complete five projects. Another incomplete but required project is the South Bank Bridge over the MBTA Commuter Rail tracks at North Station (connecting Nashua Street Park to the proposed South Bank Park, which is currently a parking lot under the Zakim Bridge at the Charles River locks).
Improvements in the lower Charles River Basin include the new walkway at Lovejoy Wharf (constructed by the developer of 160 North Washington Street, the new headquarters of Converse), the Lynch Family Skate Park (constructed in 2015 by the Charles River Conservancy), rehabilitation of historic operations buildings for the Charles River Dam and lock, a maintenance facility, and a planned pedestrian walkway across the Charles River next to the MBTA Commuter Rail drawbridge at North Station (connecting Nashua Street Park and North Point Park). MassDOT is funding the South Bank Park, and replacement of the North Washington Street Bridge (construction Aug 2018-2023). EF Education is funding public greenspace improvements as part of its three-phase expansion at North Point. Remaining funding may be used to construct the North Point Inlet pedestrian bridge, and a pedestrian walkway over Leverett Circle. Before being replaced with surface access during the reconstruction of the Science Park MBTA Green Line station, Leverett Circle had pedestrian bridges with stairs that provided elevated access between the station, the Charles River Parks, and the sidewalk to the Boston Museum of Science. The replacement ramps would comply with Americans with Disabilities Act requirements and allow easy travel by wheelchair or bicycle over the busy intersection.
While not a legally mandated requirement, public art was part of the urban design planning process (and later design development work) through the Artery Arts Program. The intent of the program was to integrate public art into highway infrastructure (retaining walls, fences, and lighting) and the essential elements of the pedestrian environment (walkways, park landscape elements, and bridges). As overall project costs increased, the Artery Arts Program was seen as a potential liability, even though there was support and interest from the public and professional arts organizations in the area.
At the beginning of the highway design process, a temporary arts program was initiated, and over 50 proposals were selected. However, development began on only a few projects before funding for the program was cut. Permanent public art that was funded includes: super graphic text and facades of former West End houses cast into the concrete elevated highway abutment support walls near North Station by artist Sheila Levrant de Bretteville; Harbor Fog, a sensor-activated mist, light and sound sculptural environment by artist Ross Miller in parcel 17; a historical sculpture celebrating the 18th and 19th century shipbuilding industry and a bust of shipbuilder Donald McKay in East Boston; blue interior lighting of the Zakim Bridge; and the Miller's River Littoral Way walkway and lighting under the loop ramps north of the Charles River.
Extensive landscape planting, as well as a maintenance program to support the plantings, was requested by many community members during public meetings.
The Big Dig untangled the co-mingled traffic from the Massachusetts Turnpike and the Sumner and Callahan tunnels. While only one net lane in each direction was added to the north–south I-93, several new east–west lanes became available. East–west traffic on the Massachusetts Turnpike/I-90 now proceeds directly through the Ted Williams Tunnel to Logan Airport and Route 1A beyond. Traffic between Storrow Drive and the Callahan and Sumner Tunnels still uses a short portion of I-93, but additional lanes and direct connections are provided for this traffic.
The result was a 62% reduction in vehicle hours of travel on I-93, the airport tunnels, and the connection from Storrow Drive, from an average 38,200 hours per day before construction (1994–1995) to 14,800 hours per day in 2004–2005, after the project was largely complete. The savings for travelers was estimated at $166 million annually in the same 2004–2005 time frame. Travel times on the Central Artery northbound during the afternoon peak hour were reduced 85.6%.
A 2008 "Boston Globe" report asserted that waiting time for the majority of trips actually increased as a result of demand induced by the increased road capacity. Because more drivers were opting to use the new roads, traffic bottlenecks were only pushed outward from the city, not reduced or eliminated (although some trips are now faster). The report states, "Ultimately, many motorists going to and from the suburbs at peak rush hours are spending more time stuck in traffic, not less." The "Globe" also asserted that their analysis provides a fuller picture of the traffic situation than a state-commissioned study done two years earlier, in which the Big Dig was credited with helping to save at least $167 million a year by increasing economic productivity and decreasing motor vehicle operating costs. That study did not look at highways outside the Big Dig construction area and did not take into account new congestion elsewhere.
As part of the project, an elaborate Operations Control Center (OCC) control room was constructed in South Boston. Staffed on a "24/7/365" basis, this center monitors and reports on traffic congestion, and responds to emergencies. Continuous video surveillance is provided by hundreds of cameras, and thousands of sensors monitor traffic speed and density, air quality, water levels, temperatures, equipment status, and other conditions inside the tunnel. The OCC can activate emergency ventilation fans, change electronic display signs, and dispatch service crews when necessary.
As far back as 2001, Turnpike Authority officials and contractors knew of thousands of leaks in ceiling and wall fissures, extensive water damage to steel supports and fireproofing systems, and overloaded drainage systems. Many of the leaks were a result of Modern Continental and other subcontractors failing to remove gravel and other debris before pouring concrete. This information was not made public, until engineers at MIT (volunteer students and professors) performed several experiments and found serious problems with the tunnel.
On September 15, 2004, a major leak in the Interstate 93 north tunnel forced the closure of the tunnel while repairs were conducted. This also forced the Turnpike Authority to release information regarding its non-disclosure of prior leaks. A follow-up reported on "extensive" leaks that were more severe than state authorities had previously acknowledged. The report went on to state that the $14.6 billion tunnel system was riddled with more than 400 leaks. A "Boston Globe" report, however, countered that by stating there were nearly 700 leaks in a single section of tunnel beneath South Station. Turnpike officials also stated that the number of leaks being investigated was down from 1,000 to 500.
The problem of leaks is further aggravated by the fact that many of them involve corrosive salt water. This is caused by the proximity of Boston Harbor and the Atlantic Ocean, causing a mix of salt and fresh water leaks in the tunnel. The situation is made worse by road salt spread in the tunnel to melt ice during freezing weather, or brought in by vehicles passing through. Salt water and salt spray are well-known issues that must be dealt with in any marine environment. It has been reported that "hundreds of thousands of gallons of salt water are pumped out monthly" in the Big Dig, and a map has been prepared showing "hot spots" where water leakage is especially serious. Salt-accelerated corrosion has caused ceiling light fixtures to fail (see below), but can also cause rapid deterioration of embedded rebar and other structural steel reinforcements holding the tunnel walls and ceiling in place.
The much larger than expected volume of water that must be continuously pumped consumes a correspondingly larger amount of electrical power, and will cause the pumps to wear out much sooner than originally estimated.
Massachusetts State Police searched the offices of Aggregate Industries, the largest concrete supplier for the underground portions of the project, in June 2005. They seized evidence that Aggregate delivered concrete that did not meet contract specifications. In March 2006 Massachusetts Attorney General Tom Reilly announced plans to sue project contractors and others because of poor work on the project. Over 200 complaints were filed by the state of Massachusetts as a result of leaks, cost overruns, quality concerns, and safety violations. In total, the state has sought approximately $100 million from the contractors ($1 for every $141 spent).
In May 2006, six employees of the company were arrested and charged with conspiracy to defraud the United States. In July 2007, Aggregate Industries settled the case with an agreement to pay $50 million. $42 million of the settlement went to civil cases and $8 million was paid in criminal fines. The company will provide $75 million in insurance for maintenance as well as pay $500,000 toward routine checks on areas suspected to contain substandard concrete.
A fatal accident raised safety questions and closed part of the project for most of the summer of 2006. On July 10, 2006, concrete ceiling panels and debris weighing and measuring fell on a car traveling on the two-lane ramp connecting northbound I-93 to eastbound I-90 in South Boston, killing Milena Del Valle, who was a passenger, and injuring her husband, Angel Del Valle, who was driving. Immediately following the fatal ceiling collapse, Governor Mitt Romney ordered a "stem-to-stern" safety audit conducted by the engineering firm of Wiss, Janney, Elstner Associates, Inc. to look for additional areas of risk. Said Romney: "We simply cannot live in a setting where a project of this scale has the potential of threatening human life, as has already been seen". The collapse and closure of the tunnel greatly snarled traffic in the city. The resulting traffic jams are cited as contributing to the death of another person, a heart attack victim who died en route to Boston Medical Center when his ambulance was caught in one such traffic jam two weeks after the collapse. On September 1, 2006, one eastbound lane of the connector tunnel was re-opened to traffic.
Following extensive inspections and repairs, Interstate 90 east- and westbound lanes reopened in early January 2007. The final piece of the road network, a high occupancy vehicle lane connecting Interstate 93 north to the Ted Williams Tunnel, reopened on June 1, 2007.
On July 10, 2007, after a lengthy investigation, the National Transportation Safety Board found that epoxy glue used to hold the roof in place during construction was not appropriate for long-term bonding. This was determined to be the cause of the roof collapse. The Power-Fast Epoxy Adhesive used in the installation was designed for short-term loading, such as wind or earthquake loads, not long-term loading, such as the weight of a panel.
Powers Fasteners, the makers of the adhesive, revised their product specifications on May 15, 2007, to increase the safety factor from 4 to 10 for all of their epoxy products intended for use in overhead applications. The safety factor on Power-Fast Epoxy was increased from 4 to 16. On December 24, 2007, the Del Valle family announced they had reached a settlement with Powers Fasteners that would pay the family $6 million. In December 2008, Power Fasteners agreed to pay $16 million to the state to settle manslaughter charges.
Public safety workers have called the walkway safety handrails in the Big Dig tunnels "ginsu guardrails", because the squared-off edges of the support posts have caused mutilations and deaths of passengers ejected from crashed vehicles. After an eighth reported death involving the safety handrails, MassDOT officials announced plans to cover or remove the allegedly dangerous fixtures, but only near curves or exit ramps. This partial removal of hazards has been criticized by a safety specialist, who suggests that the handrails are just as dangerous in straight sections of the tunnel.
In March 2011, it became known that senior MassDOT officials had failed to disclose an issue with the lighting fixtures in the O'Neill tunnel. In early February 2011, a maintenance crew found a fixture lying in the middle travel lane in the northbound tunnel. Assuming it to be simple road debris, the maintenance team picked it up and brought it back to its home facility. The next day, a supervisor passing through the yard realized that the fixture was not road debris but was in fact one of the fixtures used to light the tunnel itself. Further investigation revealed that the fixture's mounting apparatus had failed, due to galvanic corrosion of incompatible metals, caused by having aluminum in direct contact with stainless steel, in the presence of salt water. The electrochemical potential difference between stainless steel and aluminum is in the range of 0.5 to 1.0V, depending on the exact alloys involved, and can cause considerable corrosion within months under unfavorable conditions.
After the discovery of the reason why the fixture had failed, a comprehensive inspection of the other fixtures in the tunnel revealed that numerous other fixtures were also in the same state of deterioration. Some of the worst fixtures were temporarily shored up with plastic ties. Moving forward with temporary repairs, members of the MassDOT administration team decided not to let the news of the systemic failure and repair of the fixtures be released to the public or to Governor Deval Patrick's administration.
, it appeared that all of the 25,000 light fixtures would have to be replaced, at an estimated cost of $54 million. The replacement work was mostly done at night, and required lane closures or occasional closing of the entire tunnel for safety, and was estimated to take up to 2 years to complete. , replacement of the light fixtures was still ongoing. | https://en.wikipedia.org/wiki?curid=4318 |
Binary search tree
In computer science, a binary search tree (BST), also called an ordered or sorted binary tree, is a rooted binary tree whose internal nodes each store a key greater than all the keys in the node's left subtree and less than those in its right subtree. A binary tree is a type of data structure for storing data such as numbers in an organized way. Binary search trees allow binary search for fast lookup, addition and removal of data items, and can be used to implement dynamic sets and lookup tables. The order of nodes in a BST means that each comparison skips about half of the remaining tree, so the whole lookup takes time proportional to the binary logarithm of the number of items stored in the tree. This is much better than the linear time required to find items by key in an (unsorted) array, but slower than the corresponding operations on hash tables. Several variants of the binary search tree have been studied.
A binary search tree is a rooted binary tree, whose internal nodes each store a key (and optionally, an associated value), and each has two distinguished sub-trees, commonly denoted "left" and "right". The tree additionally satisfies the binary search property: the key in each node is greater than or equal to any key stored in the left sub-tree, and less than or equal to any key stored in the right sub-tree. The leaves (final nodes) of the tree contain no key and have no structure to distinguish them from one another.
Often, the information represented by each node is a record rather than a single data element. However, for sequencing purposes, nodes are compared according to their keys rather than any part of their associated records. The major advantage of binary search trees over other data structures is that the related sorting algorithms and search algorithms such as in-order traversal can be very efficient; they are also easy to code.
Binary search trees are a fundamental data structure used to construct more abstract data structures such as sets, multisets, and associative arrays.
Binary search requires an order relation by which every element (item) can be compared with every other element in the sense of a total preorder. The part of the element which effectively takes place in the comparison is called its "key". Whether duplicates, i. e. different elements with the same key, shall be allowed in the tree or not, does not depend on the order relation, but on the application only.
For a search function supporting and handling duplicates in a tree, see section Searching with duplicates allowed.
In the context of binary search trees, a total preorder is realized most flexibly by means of a three-way comparison subroutine.
Binary search trees support three main operations: insertion of elements, deletion of elements, and lookup (checking whether a key is present).
Searching in a binary search tree for a specific key can be programmed recursively or iteratively.
We begin by examining the root node. If the tree is "null", the key we are searching for does not exist in the tree. Otherwise, if the key equals that of the root, the search is successful and we return the node. If the key is less than that of the root, we search the left subtree. Similarly, if the key is greater than that of the root, we search the right subtree. This process is repeated until the key is found or the remaining subtree is "null". If the searched key is not found after a "null" subtree is reached, then the key is not present in the tree. This is easily expressed as a recursive algorithm (implemented in Python):
def search_recursively(key, node):
The same algorithm can be implemented iteratively:
def search_iteratively(key, node):
These two examples do not support duplicates, i. o. w. they rely on the order relation being a total order.
Because in the worst case this algorithm must search from the root of the tree to the leaf farthest from the root, the search operation takes time proportional to the tree's "height" (see tree terminology). On average, binary search trees with keys have height. However, in the worst case, binary search trees can have height, when the unbalanced tree resembles a linked list (degenerate tree).
If the order relation is only a total preorder, a reasonable extension of the functionality is the following: also in case of equality search down to the leaves. Thereby allowing to specify (or hard-wire) a direction, where to insert a duplicate, either to the right or left of all duplicates in the tree so far. If the direction is hard-wired, both choices, right and left, support a stack with insert duplicate as push operation and delete as the pop operation.
def search_duplicatesAllowed(key, node):
A binary tree sort equipped with such a search function becomes stable.
Insertion begins as a search would begin; if the key is not equal to that of the root, we search the left or right subtrees as before. Eventually, we will reach an external node and add the new key-value pair (here encoded as a record 'newNode') as its right or left child, depending on the node's key. In other words, we examine the root and recursively insert the new node to the left subtree if its key is less than that of the root, or the right subtree if its key is greater than or equal to the root.
Here's how a typical binary search tree insertion might be performed in a binary tree in C++:
void insert(Node*& root, int key, int value) {
Alternatively, a non-recursive version might be implemented like this. Using a pointer-to-pointer to keep track of where we came from lets the code avoid explicit checking for and handling of the case where it needs to insert a node at the tree root:
void insert(Node** root, int key, int value) {
The above "destructive" procedural variant modifies the tree in place. It uses only constant heap space (and the iterative version uses constant stack space as well), but the prior version of the tree is lost. Alternatively, as in the following Python example, we can reconstruct all ancestors of the inserted node; any reference to the original tree root remains valid, making the tree a persistent data structure:
def binary_tree_insert(node, key, value):
The part that is rebuilt uses space in the average case and in the worst case.
In either version, this operation requires time proportional to the height of the tree in the worst case, which is time in the average case over all trees, but time in the worst case.
Another way to explain insertion is that in order to insert a new node in the tree, its key is first compared with that of the root. If its key is less than the root's, it is then compared with the key of the root's left child. If its key is greater, it is compared with the root's right child. This process continues, until the new node is compared with a leaf node, and then it is added as this node's right or left child, depending on its key: if the key is less than the leaf's key, then it is inserted as the leaf's left child, otherwise as the leaf's right child.
There are other ways of inserting nodes into a binary tree, but this is the only way of inserting nodes at the leaves and at the same time preserving the BST structure.
When removing a node from a binary "search" tree it is mandatory to maintain the in-order sequence of the nodes.
There are many possibilities to do this. However, the following method which has been proposed by T. Hibbard in 1962 guarantees that the heights of the subject subtrees are changed by at most one.
There are three possible cases to consider:
In all cases, when "D" happens to be the root, make the replacement node root again.
Nodes with two children are harder to delete. A node's in-order successor is its right subtree's left-most child, and a node's in-order predecessor is the left subtree's right-most child. In either case, this node will have only one or no child at all. Delete it according to one of the two simpler cases above.
Consistently using the in-order successor or the in-order predecessor for every instance of the two-child case can lead to an unbalanced tree, so some implementations select one or the other at different times.
Runtime analysis: Although this operation does not always traverse the tree down to a leaf, this is always a possibility; thus in the worst case it requires time proportional to the height of the tree. It does not require more even when the node has two children, since it still follows a single path and does not visit any node twice.
def find_min(self):
def replace_node_in_parent(self, new_value=None) -> None:
def binary_tree_delete(self, key) -> None:
Once the binary search tree has been created, its elements can be retrieved in-order by recursively traversing the left subtree of the root node, accessing the node itself, then recursively traversing the right subtree of the node, continuing this pattern with each node in the tree as it's recursively accessed. As with all binary trees, one may conduct a pre-order traversal or a post-order traversal, but neither are likely to be useful for binary "search" trees. An in-order traversal of a binary search tree will always result in a sorted list of node items (numbers, strings or other comparable items).
The code for in-order traversal in Python is given below. It will call callback (some function the programmer wishes to call on the node's value, such as printing to the screen) for every node in the tree.
def traverse_binary_tree(node, callback):
Traversal requires time, since it must visit every node. This algorithm is also , so it is asymptotically optimal.
Traversal can also be implemented iteratively. For certain applications, e.g. greater equal search, approximative search, an operation for "single step (iterative) traversal" can be very useful. This is, of course, implemented without the callback construct and takes on average and in the worst case.
Sometimes we already have a binary tree, and we need to determine whether it is a BST. This problem has a simple recursive solution.
The BST property—every node on the right subtree has to be larger than the current node and every node on the left subtree has to be smaller than the current node—is the key to figuring out whether a tree is a BST or not. The greedy algorithm—simply traverse the tree, at every node check whether the node contains a value larger than the value at the left child and smaller than the value on the right child—does not work for all cases. Consider the following tree:
In the tree above, each node meets the condition that the node contains a value larger than its left child and smaller than its right child hold, and yet it is not a BST: the value 5 is on the right subtree of the node containing 20, a violation of the BST property.
Instead of making a decision based solely on the values of a node and its children, we also need information flowing down from the parent as well. In the case of the tree above, if we could remember about the node containing the value 20, we would see that the node with value 5 is violating the BST property contract.
So the condition we need to check at each node is:
A recursive solution in C++ can explain this further:
struct TreeNode {
bool isBST(struct TreeNode *node, int minKey, int maxKey) {
codice_1 and codice_2 are done to allow only distinct elements in BST.
If we want the same elements to also be present, then we can use only codice_3 in both places.
The initial call to this function can be something like this:
if (isBST(root, INT_MIN, INT_MAX)) {
} else {
Essentially we keep creating a valid range (starting from [MIN_VALUE, MAX_VALUE]) and keep shrinking it down for each node as we go down recursively.
As pointed out in section #Traversal, an in-order traversal of a binary "search" tree returns the nodes sorted. Thus we only need to keep the last visited node while traversing the tree and check whether its key is smaller (or smaller/equal, if duplicates are to be allowed in the tree) compared to the current key.
A binary search tree can be used to implement a simple sorting algorithm. Similar to heapsort, we insert all the values we wish to sort into a new ordered data structure—in this case a binary search tree—and then traverse it in order.
The worst-case time of codice_4 is —if you feed it a sorted list of values, it chains them into a linked list with no left subtrees. For example, codice_5 yields the tree codice_6.
There are several schemes for overcoming this flaw with simple binary trees; the most common is the self-balancing binary search tree. If this same procedure is done using such a tree, the overall worst-case time is , which is asymptotically optimal for a comparison sort. In practice, the added overhead in time and space for a tree-based sort (particularly for node allocation) make it inferior to other asymptotically optimal sorts such as heapsort for static list sorting. On the other hand, it is one of the most efficient methods of "incremental sorting", adding items to a list over time while keeping the list sorted at all times.
Binary search trees can serve as priority queues: structures that allow insertion of arbitrary key as well as lookup and deletion of the minimum (or maximum) key. Insertion works as previously explained. "Find-min" walks the tree, following left pointers as far as it can without hitting a leaf:
"Find-max" is analogous: follow right pointers as far as possible. "Delete-min" ("max") can simply look up the minimum (maximum), then delete it. This way, insertion and deletion both take logarithmic time, just as they do in a binary heap, but unlike a binary heap and most other priority queue implementations, a single tree can support all of "find-min", "find-max", "delete-min" and "delete-max" at the same time, making binary search trees suitable as double-ended priority queues.
There are many types of binary search trees. AVL trees and red-black trees are both forms of self-balancing binary search trees. A splay tree is a binary search tree that automatically moves frequently accessed elements nearer to the root. In a treap ("tree heap"), each node also holds a (randomly chosen) priority and the parent node has higher priority than its children. Tango trees are trees optimized for fast searches.
T-trees are binary search trees optimized to reduce storage space overhead, widely used for in-memory databases
A degenerate tree is a tree where for each parent node, there is only one associated child node. It is unbalanced and, in the worst case, performance degrades to that of a linked list. If your add node function does not handle re-balancing, then you can easily construct a degenerate tree by feeding it with data that is already sorted. What this means is that in a performance measurement, the tree will essentially behave like a linked list data structure.
D. A. Heger (2004) presented a performance comparison of binary search trees. Treap was found to have the best average performance, while red-black tree was found to have the smallest number of performance variations.
If we do not plan on modifying a search tree, and we know exactly how often each item will be accessed, we can construct an "optimal binary search tree", which is a search tree where the average cost of looking up an item (the "expected search cost") is minimized.
Even if we only have estimates of the search costs, such a system can considerably speed up lookups on average. For example, if we have a BST of English words used in a spell checker, we might balance the tree based on word frequency in text corpora, placing words like "the" near the root and words like "agerasia" near the leaves. Such a tree might be compared with Huffman trees, which similarly seek to place frequently used items near the root in order to produce a dense information encoding; however, Huffman trees store data elements only in leaves, and these elements need not be ordered.
If the sequence in which the elements in the tree will be accessed is unknown in advance, splay trees can be used which are asymptotically as good as any static search tree we can construct for any particular sequence of lookup operations.
"Alphabetic trees" are Huffman trees with the additional constraint on order, or, equivalently, search trees with the modification that all elements are stored in the leaves. Faster algorithms exist for "optimal alphabetic binary trees" (OABTs). | https://en.wikipedia.org/wiki?curid=4320 |
Binary tree
In computer science, a binary tree is a tree data structure in which each node has at most two children, which are referred to as the ' and the '. A recursive definition using just set theory notions is that a (non-empty) binary tree is a tuple ("L", "S", "R"), where "L" and "R" are binary trees or the empty set and "S" is a singleton set containing the root. Some authors allow the binary tree to be the empty set as well.
From a graph theory perspective, binary (and K-ary) trees as defined here are actually arborescences. A binary tree may thus be also called a bifurcating arborescence—a term which appears in some very old programming books, before the modern computer science terminology prevailed. It is also possible to interpret a binary tree as an undirected, rather than a directed graph, in which case a binary tree is an ordered, rooted tree. Some authors use rooted binary tree instead of "binary tree" to emphasize the fact that the tree is rooted, but as defined above, a binary tree is always rooted. A binary tree is a special case of an ordered K-ary tree, where "k" is 2.
In mathematics, what is termed "binary tree" can vary significantly from author to author. Some use the definition commonly used in computer science, but others define it as every non-leaf having exactly two children and don't necessarily order (as left/right) the children either.
In computing, binary trees are used in two very different ways:
To actually define a binary tree in general, we must allow for the possibility that only one of the children may be empty. An artifact, which in some textbooks is called an "extended binary tree" is needed for that purpose. An extended binary tree is thus recursively defined as:
Another way of imagining this construction (and understanding the terminology) is to consider instead of the empty set a different type of node—for instance square nodes if the regular ones are circles.
A binary tree is a rooted tree that is also an ordered tree (a.k.a. plane tree) in which every node has at most two children. A rooted tree naturally imparts a notion of levels (distance from the root), thus for every node a notion of children may be defined as the nodes connected to it a level below. Ordering of these children (e.g., by drawing them on a plane) makes possible to distinguish left child from right child. But this still doesn't distinguish between a node with left but not a right child from a one with right but no left child.
The necessary distinction can be made by first partitioning the edges, i.e., defining the binary tree as triplet (V, E1, E2), where (V, E1 ∪ E2) is a rooted tree (equivalently arborescence) and E1 ∩ E2 is empty, and also requiring that for all "j" ∈ { 1, 2 } every node has at most one E"j" child. A more informal way of making the distinction is to say, quoting the Encyclopedia of Mathematics, that "every node has a left child, a right child, neither, or both" and to specify that these "are all different" binary trees.
Tree terminology is not well-standardized and so varies in the literature.
In combinatorics one considers the problem of counting the number of full binary trees of a given size. Here the trees have no values attached to their nodes (this would just multiply the number of possible trees by an easily determined factor), and trees are distinguished only by their structure; however the left and right child of any node are distinguished (if they are different trees, then interchanging them will produce a tree distinct from the original one). The size of the tree is taken to be the number "n" of internal nodes (those with two children); the other nodes are leaf nodes and there are of them. The number of such binary trees of size "n" is equal to the number of ways of fully parenthesizing a string of symbols (representing leaves) separated by "n" binary operators (representing internal nodes), so as to determine the argument subexpressions of each operator. For instance for one has to parenthesize a string like , which is possible in five ways:
The correspondence to binary trees should be obvious, and the addition of redundant parentheses (around an already parenthesized expression or around the full expression) is disallowed (or at least not counted as producing a new possibility).
There is a unique binary tree of size 0 (consisting of a single leaf), and any other binary tree is characterized by the pair of its left and right children; if these have sizes "i" and "j" respectively, the full tree has size . Therefore, the number formula_15 of binary trees of size "n" has the following recursive description formula_16, and formula_17 for any positive integer "n". It follows that formula_15 is the Catalan number of index "n".
The above parenthesized strings should not be confused with the set of words of length 2"n" in the Dyck language, which consist only of parentheses in such a way that they are properly balanced. The number of such strings satisfies the same recursive description (each Dyck word of length 2"n" is determined by the Dyck subword enclosed by the initial '(' and its matching ')' together with the Dyck subword remaining after that closing parenthesis, whose lengths 2"i" and 2"j" satisfy ); this number is therefore also the Catalan number formula_15. So there are also five Dyck words of length 6:
These Dyck words do not correspond to binary trees in the same way. Instead, they are related by the following recursively defined bijection: the Dyck word equal to the empty string corresponds to the binary tree of size 0 with only one leaf. Any other Dyck word can be written as (formula_21)formula_22, where formula_21,formula_22 are themselves (possibly empty) Dyck words and where the two written parentheses are matched. The bijection is then defined by letting the words formula_21 and formula_22 correspond to the binary trees that are the left and right children of the root.
A bijective correspondence can also be defined as follows: enclose the Dyck word in an extra pair of parentheses, so that the result can be interpreted as a Lisp list expression (with the empty list () as only occurring atom); then the dotted-pair expression for that proper list is a fully parenthesized expression (with NIL as symbol and '.' as operator) describing the corresponding binary tree (which is in fact the internal representation of the proper list).
The ability to represent binary trees as strings of symbols and parentheses implies that binary trees can represent the elements of a free magma on a singleton set.
Binary trees can be constructed from programming language primitives in several ways.
In a language with records and references, binary trees are typically constructed by having a tree node structure which contains some data and references to its left child and its right child. Sometimes it also contains a reference to its unique parent. If a node has fewer than two children, some of the child pointers may be set to a special null value, or to a special sentinel node.
This method of storing binary trees wastes a fair bit of memory, as the pointers will be null (or point to the sentinel) more than half the time; a more conservative representation alternative is threaded binary tree.
In languages with tagged unions such as ML, a tree node is often a tagged union of two types of nodes, one of which is a 3-tuple of data, left child, and right child, and the other of which is a "leaf" node, which contains no data and functions much like the null value in a language with pointers. For example, the following line of code in OCaml (an ML dialect) defines a binary tree that stores a character in each node.
type chr_tree = Empty | Node of char * chr_tree * chr_tree
Binary trees can also be stored in breadth-first order as an implicit data structure in arrays, and if the tree is a complete binary tree, this method wastes no space. In this compact arrangement, if a node has an index "i", its children are found at indices formula_27 (for the left child) and formula_28 (for the right), while its parent (if any) is found at index "formula_29" (assuming the root has index zero). This method benefits from more compact storage and better locality of reference, particularly during a preorder traversal. However, it is expensive to grow and wastes space proportional to 2"h" - "n" for a tree of depth "h" with "n" nodes.
This method of storage is often used for binary heaps.
A succinct data structure is one which occupies close to minimum possible space, as established by information theoretical lower bounds. The number of different binary trees on formula_1 nodes is formula_31, the formula_1th Catalan number (assuming we view trees with identical "structure" as identical). For large formula_1, this is about formula_34; thus we need at least about formula_35 bits to encode it. A succinct binary tree therefore would occupy formula_36 bits.
One simple representation which meets this bound is to visit the nodes of the tree in preorder, outputting "1" for an internal node and "0" for a leaf. If the tree contains data, we can simply simultaneously store it in a consecutive array in preorder. This function accomplishes this:
The string "structure" has only formula_37 bits in the end, where formula_1 is the number of (internal) nodes; we don't even have to store its length. To show that no information is lost, we can convert the output back to the original tree like this:
More sophisticated succinct representations allow not only compact storage of trees but even useful operations on those trees directly while they're still in their succinct form.
There is a one-to-one mapping between general ordered trees and binary trees, which in particular is used by Lisp to represent general ordered trees as binary trees. To convert a general ordered tree to binary tree, we only need to represent the general tree in left-child right-sibling way. The result of this representation will automatically be a binary tree, if viewed from a different perspective. Each node "N" in the ordered tree corresponds to a node "N' " in the binary tree; the "left" child of "N' " is the node corresponding to the first child of "N", and the "right" child of "N' " is the node corresponding to "N" 's next sibling --- that is, the next node in order among the children of the parent of "N". This binary tree representation of a general order tree is sometimes also referred to as a left-child right-sibling binary tree (also known as LCRS tree, doubly chained tree, filial-heir chain).
One way of thinking about this is that each node's children are in a linked list, chained together with their "right" fields, and the node only has a pointer to the beginning or head of this list, through its "left" field.
For example, in the tree on the left, A has the 6 children {B,C,D,E,F,G}. It can be converted into the binary tree on the right.
The binary tree can be thought of as the original tree tilted sideways, with the black left edges representing "first child" and the blue right edges representing "next sibling". The leaves of the tree on the left would be written in Lisp as:
which would be implemented in memory as the binary tree on the right, without any letters on those nodes that have a left child.
There are a variety of different operations that can be performed on binary trees. Some are mutator operations, while others simply return useful information about the tree.
Nodes can be inserted into binary trees in between two other nodes or added after a leaf node. In binary trees, a node that is inserted is specified as to which child it is.
To add a new node after leaf node A, A assigns the new node as one of its children and the new node assigns node A as its parent.
Insertion on internal nodes is slightly more complex than on leaf nodes. Say that the internal node is node A and that node B is the child of A. (If the insertion is to insert a right child, then B is the right child of A, and similarly with a left child insertion.) A assigns its child to the new node and the new node assigns its parent to A. Then the new node assigns its child to B and B assigns its parent as the new node.
Deletion is the process whereby a node is removed from the tree. Only certain nodes in a binary tree can be removed unambiguously.
Suppose that the node to delete is node A. If A has no children, deletion is accomplished by setting the child of A's parent to null. If A has one child, set the parent of A's child to A's parent and set the child of A's parent to A's child.
In a binary tree, a node with two children cannot be deleted unambiguously. However, in certain binary trees (including binary search trees) these nodes "can" be deleted, though with a rearrangement of the tree structure.
Pre-order, in-order, and post-order traversal visit each node in a tree by recursively visiting each node in the left and right subtrees of the root.
In depth-first order, we always attempt to visit the node farthest from the root node that we can, but with the caveat that it must be a child of a node we have already visited. Unlike a depth-first search on graphs, there is no need to remember all the nodes we have visited, because a tree cannot contain cycles. Pre-order is a special case of this. See depth-first search for more information.
Contrasting with depth-first order is breadth-first order, which always attempts to visit the node closest to the root that it has not already visited. See breadth-first search for more information. Also called a "level-order traversal".
In a complete binary tree, a node's breadth-index ("i" − (2"d" − 1)) can be used as traversal instructions from the root. Reading bitwise from left to right, starting at bit "d" − 1, where "d" is the node's distance from the root ("d" = ⌊log2("i"+1)⌋) and the node in question is not the root itself ("d" > 0). When the breadth-index is masked at bit "d" − 1, the bit values 0 and 1 mean to step either left or right, respectively. The process continues by successively checking the next bit to the right until there are no more. The rightmost bit indicates the final traversal from the desired node's parent to the node itself. There is a time-space trade-off between iterating a complete binary tree this way versus each node having pointer/s to its sibling/s. | https://en.wikipedia.org/wiki?curid=4321 |
Borel measure
In mathematics, specifically in measure theory, a Borel measure on a topological space is a measure that is defined on all open sets (and thus on all Borel sets). Some authors require additional restrictions on the measure, as described below.
Let formula_1 be a locally compact Hausdorff space, and let formula_2 be the smallest σ-algebra that contains the open sets of formula_1; this is known as the σ-algebra of Borel sets. A Borel measure is any measure formula_4 defined on the σ-algebra of Borel sets. Some authors require in addition that formula_4 is locally compact, meaning that formula_6 for every compact set formula_7. If a Borel measure formula_4 is both inner regular and outer regular, it is called a regular Borel measure. If formula_4 is both inner regular and locally finite, it is called a Radon measure.
The real line formula_10 with its usual topology is a locally compact Hausdorff space, hence we can define a Borel measure on it. In this case, formula_11 is the smallest σ-algebra that contains the open intervals of formula_10. While there are many Borel measures "μ", the choice of Borel measure that assigns formula_13 for every half-open interval formula_14 is sometimes called "the" Borel measure on formula_10. This measure turns out to be the restriction to the Borel σ-algebra of the Lebesgue measure formula_16, which is a complete measure and is defined on the Lebesgue σ-algebra. The Lebesgue σ-algebra is actually the "completion" of the Borel σ-algebra, which means that it is the smallest σ-algebra that contains all the Borel sets and has a complete measure on it. Also, the Borel measure and the Lebesgue measure coincide on the Borel sets (i.e., formula_17 for every Borel measurable set, where formula_4 is the Borel measure described above).
If "X" and "Y" are second-countable, Hausdorff topological spaces, then the set of Borel subsets formula_19 of their product coincides with the product of the sets formula_20 of Borel subsets of "X" and "Y". That is, the Borel functor
from the category of second-countable Hausdorff spaces to the category of measurable spaces preserves finite products.
The Lebesgue–Stieltjes integral is the ordinary Lebesgue integral with respect to a measure known as the Lebesgue–Stieltjes measure, which may be associated to any function of bounded variation on the real line. The Lebesgue–Stieltjes measure is a regular Borel measure, and conversely every regular Borel measure on the real line is of this kind.
One can define the Laplace transform of a finite Borel measure μ on the real line by the Lebesgue integral
An important special case is where μ is a probability measure or, even more specifically, the Dirac delta function. In operational calculus, the Laplace transform of a measure is often treated as though the measure came from a distribution function "f". In that case, to avoid potential confusion, one often writes
where the lower limit of 0− is shorthand notation for
This limit emphasizes that any point mass located at 0 is entirely captured by the Laplace transform. Although with the Lebesgue integral, it is not necessary to take such a limit, it does appear more naturally in connection with the Laplace–Stieltjes transform.
Given a Borel measure μ on a metric space "X" such that μ("X") > 0 and μ("B"("x", "r")) ≤ "rs" holds for some constant "s" > 0 and for every ball "B"("x", "r") in "X", then the Hausdorff dimension dimHaus("X") ≥ "s". A partial converse is provided by Frostman's lemma:
Lemma: Let "A" be a Borel subset of R"n", and let "s" > 0. Then the following are equivalent:
The Cramér–Wold theorem in measure theory states that a Borel probability measure on formula_26 is uniquely determined by the totality of its one-dimensional projections. It is used as a method for proving joint convergence results. The theorem is named after Harald Cramér and Herman Ole Andreas Wold. | https://en.wikipedia.org/wiki?curid=4322 |
Blackadder
Blackadder is a series of four BBC One pseudohistorical British sitcoms, plus several one-off instalments, which originally aired from 1983 to 1989. All television episodes starred Rowan Atkinson as the antihero Edmund Blackadder and Tony Robinson as Blackadder's dogsbody, Baldrick. Each series was set in a different historical period, with the two protagonists accompanied by different characters, though several reappear in one series or another, e.g., Melchett (Stephen Fry) and Lord Flashheart (Rik Mayall).
The first series, "The Black Adder", was written by Richard Curtis and Rowan Atkinson, while subsequent episodes were written by Curtis and Ben Elton. The shows were produced by John Lloyd. In 2000, the fourth series, "Blackadder Goes Forth", ranked at 16 in the 100 Greatest British Television Programmes, a list created by the British Film Institute. In a 2001 poll by Channel 4 Edmund Blackadder was ranked third on their list of the 100 Greatest TV Characters. In the 2004 TV poll to find Britain's Best Sitcom, "Blackadder" was voted the second-best British sitcom of all time, topped by "Only Fools and Horses". It was also ranked as the 9th-best TV show of all time by "Empire" magazine.
Although each series is set in a different era, all follow the misfortunes of Edmund Blackadder (played by Atkinson), who in each is a member of a British family dynasty present at many significant periods and places in British history. It is implied in each series that the Blackadder character is a descendant of the previous one (the end theme lyrics of series 2 episode "Head" specify that he is the great-grandson of the previous), although it is never specified how or when any of the Blackadders (who are usually bachelors) managed to father children.
In series one, Edmund Blackadder is not particularly bright, and is much the intellectual inferior of his servant, Baldrick (played by Tony Robinson). However, in subsequent series the positions are reversed: Blackadder is clever, shrewd, scheming and manipulative while Baldrick is extremely dim. Each incarnation of Blackadder and Baldrick is also saddled with tolerating the presence of a dimwitted aristocrat. This role was taken in the first two series by Lord Percy Percy, played by Tim McInnerny, with Hugh Laurie playing the role in the third and fourth series, as Prince George, Prince Regent and Lieutenant George, respectively.
Each series is set in a different period of British history, beginning in 1485 and ending in 1917, and comprises six half-hour episodes. The first series, made in 1983, was called "The Black Adder" and was set in the fictional reign of "Richard IV". The second series, "Blackadder II" (1986), was set during the reign of Elizabeth I. "Blackadder the Third" (1987) was set during the late 18th and early 19th centuries in the reign of George III, and "Blackadder Goes Forth" (1989) was set in 1917 in the trenches of the Great War.
"The Black Adder", the first series of "Blackadder", was written by Richard Curtis and Rowan Atkinson and produced by John Lloyd. It originally aired on BBC1 from 15 June 1983 to 20 July 1983, and was a joint production with the Australian Seven Network.
Set in 1485 at the end of the British Middle Ages, the series is written as an alternative history in which King Richard III won the Battle of Bosworth Field only to be mistaken for someone else and murdered, and is succeeded by Richard IV (Brian Blessed), one of the Princes in the Tower. The series follows the exploits of Richard IV's unfavoured second son Edmund, the Duke of Edinburgh (who calls himself "The Black Adder") in his various attempts to increase his standing with his father and his eventual quest to overthrow him.
Conceived while Atkinson and Curtis were working on "Not the Nine O'Clock News", the series dealt comically with a number of aspects of medieval life in Britain: witchcraft, royal succession, European relations, the Crusades, and the conflict between the Church and the Crown. Along with the secret history, many historical events portrayed in the series were anachronistic (for example, Constantinople had already fallen to the Ottoman Empire in 1453, predating the events in the episode by 33 years); this dramatic licence would continue in the subsequent "Blackadders". The filming of the series was highly ambitious, with a large cast and much location shooting. The series also featured Shakespearean dialogue, often adapted for comic effect; the end credits featured the words "Additional Dialogue by William Shakespeare".
"Blackadder II" is set in England during the reign of Queen Elizabeth I (1558–1603), who is portrayed by Miranda Richardson. The principal character is Edmund, Lord Blackadder, the great-grandson of the original Black Adder. During the series, he regularly deals with the Queen, her obsequious Lord Chamberlain Lord Melchett (Stephen Fry) – his rival – and the Queen's demented former nanny Nursie (Patsy Byrne).
Following the BBC's request for improvements (and a severe budget reduction), several changes were made. The second series was the first to establish the familiar Blackadder character: cunning, shrewd and witty, in sharp contrast to the first series' bumbling Prince Edmund. To reduce the cost of production, it was shot with virtually no outdoor scenes (the first series was shot largely on location) and several frequently used indoor sets, such as the Queen's throne room and Blackadder's front room.
A quote from this series ranked number three in a list of the top 25 television "putdowns" of the last 40 years by the "Radio Times" magazine: "The eyes are open, the mouth moves, but Mr. Brain has long since departed, hasn't he, Percy?"
"Blackadder the Third" is set in the late 18th and early 19th centuries, a period known as the Regency. In the series, Edmund Blackadder Esquire is the butler to the Prince Regent, the Prince of Wales (the prince is played by Hugh Laurie as a complete fop and idiot). Despite Edmund's respected intelligence and abilities, he has no personal fortune to speak of, apart from his frequently fluctuating wage packet (as well, it seems, from stealing and selling off the Prince's socks) from the Prince: "If I'm running short of cash, all I have to do is go upstairs and ask Prince Fathead for a rise."
As well as Rowan Atkinson and Tony Robinson in their usual roles, this series starred Hugh Laurie as the Prince Regent, and Helen Atkinson-Wood as Mrs. Miggins. The series features Dr. Samuel Johnson (Robbie Coltrane); William Pitt the Younger (Simon Osborne); the French Revolution (with Chris Barrie, Tim McInnerny as the Scarlet Pimpernel, and Nigel Planer); hammy theatrical actors (Kenneth Connor and Hugh Paddick); a squirrel-hating cross-dressing highwayman (Miranda Richardson); and a duel with the Duke of Wellington (Stephen Fry).
This series is set in 1917, on the Western Front in the trenches of the First World War. Another "big push" is planned, and Captain Blackadder's one goal is to avoid being killed, but his schemes always land him back in the trenches. Blackadder is joined by his batman Private S. Baldrick (Tony Robinson) and idealistic Edwardian twit Lieutenant George (Hugh Laurie). General Melchett (Stephen Fry) rallies his troops from a French château thirty-five miles from the front, where he is aided and abetted by his assistant, Captain Kevin Darling (Tim McInnerny), pencil-pusher supreme and Blackadder's nemesis, whose name is played on for maximum comedic value.
The series' tone is somewhat darker than the other "Blackadder"s; it details the deprivations of trench warfare as well as the incompetence and life-wasting strategies of the top brass. For example, Baldrick is reduced to making coffee from mud and cooking rats, while General Melchett hatches a plan for the troops to walk very slowly toward the German lines, because "it'll be the last thing Fritz will expect."
The final episode, "Goodbyeee", is known for being extraordinarily poignant for a comedy – especially the final scene, which sees the main characters (Blackadder, Baldrick, George, and Darling) finally going "over the top" and charging off into the fog and smoke of no man's land presumably to die. In a list of the 100 Greatest British Television Programmes, drawn up by the British Film Institute in 2000 and voted for by industry professionals, "Blackadder Goes Forth" was placed 16th.
The "Blackadder" pilot was shot but never broadcast on terrestrial TV in the UK (although some scenes were shown in the 25th anniversary special "Blackadder Rides Again"). One notable difference in the pilot, as in many pilots, is the casting. Baldrick is played not by Tony Robinson, but by Philip Fox. Another significant difference is that the character of Prince Edmund presented in the pilot is much closer to the intelligent, conniving Blackadder of the later series than the snivelling, weak buffoon of the original. Set in the year 1582, the script of the pilot is roughly the same as the episode "Born to Be King", albeit with some different jokes, with some lines appearing in other episodes of the series.
This special, set in the English Civil War, was shown as part of Comic Relief's Red Nose Day on Friday 5 February 1988. The 15-minute episode is set in November 1648, during the last days of the Civil War. Sir Edmund Blackadder and his servant, Baldrick, are the last two men loyal to the defeated King Charles I of England (played by Stephen Fry, portrayed as a soft-spoken, ineffective, naive character, with the voice and mannerisms of Charles I's namesake, the current Prince of Wales). However, due to a misunderstanding between Oliver Cromwell (guest-star Warren Clarke) and Baldrick, the king is arrested and sent to the Tower of London. The rest of the episode revolves around Blackadder's attempts to save the king, as well as improve his standing.
The second special was broadcast on Friday 23 December 1988. In a twist on Charles Dickens' "A Christmas Carol", Ebenezer Blackadder is the "kindest and loveliest" man in England. The Spirit of Christmas shows Blackadder the contrary antics of his ancestors and descendants, and reluctantly informs him that if he turns evil his descendants will enjoy power and fortune, while if he remains the same a future Blackadder will live shamefully subjugated to a future incompetent Baldrick. This remarkable encounter causes him to proclaim, "Bad guys have all the fun", and adopt the personality with which viewers are more familiar.
"Blackadder: Back & Forth" was originally shown in the Millennium Dome in 2000, followed by a screening on Sky One in the same year (and later on BBC1). It is set on the turn of the millennium, and features Lord Blackadder placing a bet with his friends – modern versions of Queenie (Miranda Richardson), Melchett (Stephen Fry), George (Hugh Laurie) and Darling (Tim McInnerny) – that he has built a working time machine. While this is intended as a clever con trick, the machine surprisingly works, sending Blackadder and Baldrick back to the Cretaceous period, where they manage to cause the extinction of the dinosaurs through the use of Baldrick's best-worst-and-only pair of underpants as a weapon against a hungry T. Rex. Finding that Baldrick has forgotten to write dates on the machine's dials, the rest of the film follows their attempts to find their way back to 1999, often creating huge historical anomalies in the process that must be corrected before the end. The film includes cameo appearances from Kate Moss and Colin Firth.
Broadcast as part of Children in Need and Comic Relief's joint special The Big Night In during the 2020 Covid-19 pandemic, Fry resumed the role of Lord Melchett, Head of the Royal Household, under lockdown at Melchett Manor to help Prince William deal with educating his children via Zoom and discussing Tiger King, before they both step outside to clap for the NHS. Melchett is said to be isolating with Lord Blackadder, both grandsons to their First World War counterparts
In 1998, as part of HRH Prince Charles' 50th Birthday Gala televised on ITV, Atkinson returned to the Cavalier incarnation of Blackadder reading aloud a letter to the Privy Council of King Charles I. He colourfully refuses their invitation to stage a royal gala, calling such occasions "very, very, very dull" and asserting that there was "more musical talent on display when my servant Baldrick breaks wind."
In 2000, on the BBC's annual Royal Variety Performance, Atkinson portrayed Blackadder as a present-day officer in "Her Majesty's Royal Regiment of Shirkers" and delivered a monologue titled "Blackadder: The Army Years," proposing that Britain regain her former greatness by invading (or at least buying) France.
In 2012, as part of the Prince's Trust charity show "We Are Most Amused", Atkinson and Robinson reprised their roles as Blackadder and Baldrick in a comedy sketch featuring Miranda Hart as leader of a government inquiry into the recent banking crisis. Blackadder, chief executive of a fictional British bank, appearing with Baldrick as his gardener, convinces the panel to publicly blame the entire crisis on Baldrick, to the latter's consternation.
Rowan Atkinson and Richard Curtis developed the idea for the sitcom while working on "Not the Nine O'Clock News". Eager to avoid comparisons to the critically acclaimed "Fawlty Towers", they proposed the idea of a historical sitcom. An unaired pilot episode was made in 1982, and a six-episode series was commissioned. The budget for the series was considerable, with much location shooting particularly at Alnwick Castle in Northumberland and the surrounding countryside in February 1983. The series also used large casts of extras, horses and expensive medieval-style costumes. Atkinson has said about the making of the first series:
The first series was odd, it was very extravagant. It cost a million pounds for the six programmes ... [which] was a lot of money to spend ... It looked great, but it wasn't as consistently funny as we would have liked.
Due to the high cost of the first series, the then-controller of programming of BBC1, Michael Grade, was reluctant to sign off a second series without major improvements to the show and drastic cost-cutting, leaving a gap of three years between the two series. Atkinson did not wish to continue writing for the second series.
A chance meeting between Richard Curtis and comedian Ben Elton led to the decision to collaborate on a new series of Blackadder. Recognising the main faults of the first series, Curtis and Elton agreed that "Blackadder II" would be a studio-only production (along with the inclusion of a live audience during recording, instead of showing the episodes to an audience after taping). Besides adding a greater comedy focus, Elton suggested a major change in character emphasis: Baldrick would become the stupid sidekick, while Edmund Blackadder evolved into a cunning sycophant. This led to the familiar set-up that was maintained in the following series.
Only in the "Back & Forth" millennium special was the shooting once again on location, because this was a production with a budget estimated at £3 million, and was a joint venture between Tiger Aspect, Sky Television, the New Millennium Experience Company and the BBC, rather than the BBC alone.
Each series tended to feature the same set of regular actors in different period settings, although throughout the four series and specials, only Blackadder and Baldrick were constant characters. Several regular cast members recurred as characters with similar names, implying, like Blackadder, that they were descendants.
Various actors have appeared in more than one of the Blackadder series and/or specials. These are:
Ben Elton's arrival after the first series heralded the more frequent recruitment of comic actors from the alternative comedy era for guest appearances, including Robbie Coltrane, Rik Mayall (who had appeared in the final episode of the first series as "Mad Gerald"), Adrian Edmondson, Nigel Planer, Mark Arden, Stephen Frost, Chris Barrie and Jeremy Hardy. Elton himself played an anarchist in "Blackadder the Third".
Gabrielle Glaister played Bob, an attractive girl who poses as a man, in both series 2 and 4. Rik Mayall plays Lord Flashheart, a vulgar friend in his first appearance and then a successful rival of Blackadder in later episodes of series 2 and 4. He also played a decidedly Flashheart-like Robin Hood in "Back & Forth". Lee Cornes also appeared in an episode of all three Curtis-Elton series. He appeared as a guard in the episode "Chains" of "Blackadder II"; as the poet Shelley in the episode "Ink and Incapability' of "Blackadder the Third"; and as firing squad soldier Private Fraser in the episode "Corporal Punishment" of "Blackadder Goes Forth".
More established actors, some at the veteran stage of their careers, were also recruited for roles. These included Peter Cook, John Grillo, Simon Jones, Tom Baker, Jim Broadbent, Hugh Paddick, Frank Finlay, Kenneth Connor, Bill Wallis, Ronald Lacey, Roger Blake, Denis Lill, Warren Clarke and Geoffrey Palmer, who played Field Marshal Sir Douglas Haig in "Goodbyeee", the final episode of "Blackadder Goes Forth". Miriam Margolyes played three different guest roles: The Spanish Infanta in The Queen of Spain's Beard, Lady Whiteadder in Beer, and Queen Victoria in Blackadder's Christmas Carol.
Unusually for a sitcom based loosely on factual events and in the historical past, a man was recruited for one episode essentially to play himself. Political commentator Vincent Hanna played a character billed as "his own great-great-great grandfather" in the episode "Dish and Dishonesty" of "Blackadder the Third". Hanna was asked to take part because the scene was of a by-election in which Baldrick was a candidate and, in the style of modern television, Hanna gave a long-running "live" commentary of events at the count (and interviewed candidates and election agents) to a crowd through the town hall window.
Howard Goodall's theme tune has the same melody throughout all the series, but is played in roughly the style of the period in which it is set. It is performed mostly with trumpets and timpani in "The Black Adder", the fanfares used suggesting typical medieval court fanfares; with a combination of recorder, string quartet and electric guitar in "Blackadder II" (the end theme, with different lyrics each time reflecting on the episode's events, was sung by a countertenor); on oboe, cello and harpsichord (in the style of a minuet) for "Blackadder the Third"; by The Band of the 3rd Battalion, Royal Anglian Regiment in "Blackadder Goes Forth"; sung by carol singers in "Blackadder's Christmas Carol"; and by an orchestra in "Blackadder: The Cavalier Years" and "Blackadder: Back & Forth".
In 2000, the fourth series, "Blackadder Goes Forth", ranked at 16 in the "100 Greatest British Television Programmes", a list created by the British Film Institute. In 2004, a BBC TV poll for "Britain's Best Sitcom", "Blackadder" was voted the second best British sitcom of all time, topped by "Only Fools and Horses". It was also ranked as the 20th Best TV Show of All Time by "Empire" magazine.
Despite regular statements denying any plans for a fifth series, cast members are regularly asked about the possibility of a new series.
In January 2005, Tony Robinson told ITV's "This Morning" that Rowan Atkinson was more keen than he has been in the past to do a fifth series, set in the 1960s (centred on a rock band called the "Black Adder Five", with Baldrick – a.k.a. 'Bald Rick' – as the drummer). In the documentary "Blackadder Rides Again", Robinson stated that the series would present Blackadder as the bastard son of Queen Elizabeth II and running a Beatles-like rock band. Rowan Atkinson, Tony Robinson, Hugh Laurie, Stephen Fry, Tim McInnerny and Miranda Richardson would have reprised their roles, and reportedly, Brian Blessed, Elspet Gray and Robert East would have returned from the first series to play Blackadder's biological family. Robinson in a stage performance 1 June 2007, again mentioned this idea, but in the context of a movie.
One idea mentioned by Curtis was that it was Baldrick who had accidentally assassinated John F. Kennedy. However, aside from a brief mention in June 2005,
there have been no further announcements from the BBC that a new series is being planned. Furthermore, in November 2005, Rowan Atkinson told "BBC Breakfast" that, although he would very much like to do a new series set in Colditz or another prisoner-of-war camp during World War II, something which both he and Stephen Fry reiterated at the end of "Blackadder Rides Again", the chances of it happening are extremely slim.
There were a couple of ideas that had previously floated for the fifth series. "Batadder" was intended to be a parody of "Batman" with Baldrick as the counterpart of Robin (suggested by John Lloyd). This idea eventually came to surface as part of the "Comic Relief" sketch "Spider-Plant Man" in 2005, with Atkinson as the title hero, Robinson as Robin, Jim Broadbent as Batman and Rachel Stevens as Mary Jane. "Star Adder" was to be set in space in the future (suggested by Atkinson), though this too was touched upon in "Blackadder's Christmas Carol".
On 10 April 2007, "Hello!" reported that Atkinson was moving forward with his ideas for a fifth series. He said, "I like the idea of him being a prisoner of war in Colditz. That would have the right level of authority and hierarchy which is apparent in all the Blackadders."
A post on BlackAdderHall.com by Ben Elton in early 2007 said that "Blackadder" would return in some form, whether it be a TV series or film. Elton has since not given any more information on the putative "Blackadder 5".
During an interview in August 2007 about his film "Mr. Bean's Holiday", Atkinson was asked about the possibility of a further "Blackadder" series, to which the simple reply "No, no chance" was given:
Stephen Fry has expressed the view that, since the series went out on such a good "high", a film might not be a good idea.
During his June 2007 stage performance, chronicled on the Tony Robinson's Cunning Night Out DVD, Robinson states that, after filming the "Back & Forth special", the general idea was to reunite for another special in 2010. Robinson jokingly remarked that Hugh Laurie's success on "House" may make that difficult.
At the end of "Blackadder Rides Again", Robinson asked Tim McInnerny if he would do another series and he responded "no", because he thought people would not want to see them as they are now and would rather remember them for how they were. In the same documentary, Rowan Atkinson voiced his similar view; 'Times past; that's what they were!' However, Miranda Richardson and Tony Robinson expressed enthusiasm towards the idea of a series set in the Wild West, whilst John Lloyd favoured an idea for a series with a Neanderthal Blackadder. Lastly, Stephen Fry suggested a series set in a prisoner of war camp during World War II, but later remarked that "perhaps it's best to leave these things as a memory."
On 28 November 2012, Rowan Atkinson reprised the role at the "We are most amused" comedy gala for the Prince's Trust at the Royal Albert Hall. He was joined by Tony Robinson as Baldrick. The sketch involved Blackadder as CEO of Melchett, Melchett and Darling bank facing an enquiry over the banking crisis.
In August 2015, Tony Robinson said in an interview "I do think a new series of Blackadder is on the cards. I have spoken to virtually all the cast about this now. The only problem is Hugh [Laurie]'s fee. He's a huge star now." However, in October 2018, Richard Curtis "dashed hopes" that the show would return for a fifth series.
All series and many of the specials are available on DVD and video. Many are also available on BBC audio cassette. As of 2008, a "Best of BBC" edition box set is available containing all four major series together with "Blackadder's Christmas Carol" and "Back & Forth". All four series and the Christmas special are also available for download on iTunes.
5 February 1990, BBC Enterprises Ltd released the first series on two single videos.
2 October 1989, BBC Enterprises Ltd released the second series on two single videos.
6 March 1989, BBC Enterprises Ltd released the third series on two single videos.
10 September 1990, BBC Enterprises Ltd released the fourth and final series on two single videos.
On 7 September 1992, all eight single Blackadder video releases were re-released as four "complete" double VHS releases. The four entire series videos were re-released as single video releases on 2 October 1995.
On 5 January 1998, five episodes of the first two series were released on a 15 rated compiled video by BBC Worldwide Ltd.
On 4 November 1991, "Blackadder's Christmas Carol" was released on a single video release rated PG (Cat. No. BBCV 4646) | https://en.wikipedia.org/wiki?curid=4326 |
Boii
The Boii (Latin plural, singular "Boius"; ) were a Gallic tribe of the later Iron Age, attested at various times in Cisalpine Gaul (northern Italy), Pannonia (Hungary), parts of Bavaria, in and around Bohemia (after whom the region is named in most languages; comprising the bulk of Czechia), parts of Poland, and Gallia Narbonensis. In addition the archaeological evidence indicates that in the 2nd century BC Celts expanded from Bohemia through the Kłodzko Valley into Silesia, now part of Poland and Czechia.
They first appear in history in connection with the Gallic invasion of north Italy, 390 BC, when they made the Etruscan city of Felsina their new capital, Bononia (Bologna). After a series of wars they were decisively beaten by the Romans in a Battle of Mutina (193 BC) and their territory became part of the Roman province of Cisalpine Gaul. According to Strabo, writing two centuries after the events, rather than being destroyed by the Romans like their Celtic neighbours,
Around 60 BC, a group of Boii joined the Helvetiis' ill-fated attempt to conquer land in western Gaul and were defeated by Julius Caesar, along with their allies, in the Battle of Bibracte. Caesar settled the remnants of that group in Gorgobina, from where they sent two thousand to Vercingetorix's aid at the Battle of Alesia six years later. The eastern Boii on the Danube were incorporated into the Roman Empire in 8 AD.
From all the different names of the same Celtic people in literature and inscriptions it is possible to abstract a Continental Celtic segment, "boio"-. There are two major derivations of this segment, both presupposing that it belongs to the family of Indo-European languages: from 'cow' and from 'warrior.' The Boii would thus be either "the herding people" or "the warrior people".
The "cow" derivation depends most immediately on the Old Irish legal term for "outsider": "ambue", from Proto-Celtic "*ambouios" ( In a reference to the first known historical Boii, Polybius relates that their wealth consisted of cattle and gold, that they depended on agriculture and war, and that a man's status depended on the number of associates and assistants he had. The latter were presumably the "*ambouii", as opposed to the man of status, who was "*bouios", a cattle owner, and the "*bouii" were originally a class, "the cattle owners".
The "warrior" derivation was adopted by the linguist Julius Pokorny, who presented it as being from Indo-European "*bhei(ə)-", "*bhī-", "hit"; however, not finding any Celtic names close to it (except for the Boii), he adduces examples somewhat more widely from originals further back in time: "phohiio-s-", a Venetic personal name; "Boioi", an Illyrian tribe; "Boiōtoi", a Greek tribal name ("the Boeotians"); and a few others. "Boii" would be from the o-grade of "*bhei-", which is "*bhoi-". Such a connection is possible if the original form of "Boii" belonged to a tribe of Proto-Indo-European speakers long before the time of the historic Boii. If that is the case, then the Celtic tribe of central Europe must have been a final daughter population of a linguistically-diversified ancestor tribe.
The same wider connections can be hypothesized for the "cow" derivation: the Boeotians have been known for well over a century as a people of kine, which might have been parallel to the meaning of Italy as a "land of calves". Indo-European reconstructions can be made using "cow" as a basis, such as ; the root may itself be an imitation of the sound a cow makes.
Contemporary derived words include "Boiorix" ("king of the Boii", one of the chieftains of the Cimbri) and "Boiodurum" ("gate/fort of the Boii", modern Passau) in Germany. Their memory also survives in the modern regional names of Bohemia ("Boiohaemum"), a mixed-language form from boio- and Proto-Germanic "*haimaz", "home": "home of the Boii", and , Bavaria, which is derived from the Germanic "Baiovarii" tribe (Germanic "*baja-warjaz": the first component is most plausibly explained as a Germanic version of "Boii"; the second part is a common formational morpheme of Germanic tribal names, meaning 'dwellers', as in Old English "-ware"); this combination "Boii-dwellers" may have meant "those who dwell where the Boii formerly dwelt".
According to the ancient authors, the Boii arrived in northern Italy by crossing the Alps. While of the other tribes who had come to Italy along with the Boii, the Senones, Lingones and Cenomani are also attested in Gaul at the time of the Roman conquest. It remains therefore unclear where exactly the Central European origins of the Boii lay, if somewhere in Gaul, Southern Germany or in Bohemia.
Polybius relates that the Celts were close neighbors of the Etruscan civilization and "cast covetous eyes on their beautiful country". Invading the Po Valley with a large army, they drove out the Etruscans and resettled it, the Boii taking the right bank in the center of the valley. Strabo confirms that the Boii emigrated from their lands across the Alps and were one of the largest tribes of the Celts. The Boii occupied the old Etruscan settlement of Felsina, which they named "Bononia" (modern Bologna). Polybius describes the Celtic way of life in Cisalpine Gaul as follows:
The archaeological evidence from Bologna and its vicinity contradicts the testimony of Polybius and Livy on some points, who say the Boii expelled the Etruscans and perhaps some were forced to leave. It much rather indicates that the Boii neither destroyed nor depopulated Felsinum, but simply moved in and became part of the population by intermarriage. The cemeteries of the period in Bologna contain La Tène weapons and other artifacts, as well as Etruscan items such as bronze mirrors. At Monte Bibele not far away one grave contained La Tène weapons and a pot with an Etruscan female name scratched on it.
In the second half of the 3rd century BC, the Boii allied with the other Cisalpine Gauls and the Etruscans against Rome. They also fought alongside Hannibal, killing the Roman general Lucius Postumius Albinus in 216 BC, whose skull was then turned into a sacrificial bowl. A short time earlier, they had been defeated at the Battle of Telamon in 225 BC, and were again at Placentia in 194 BC (modern Piacenza) and Mutina in 193 BC (modern Modena). Publius Cornelius Scipio Nasica completed the Roman conquest of the Boii in 191 BC, celebrating a triumph for it. After their losses, according to Strabo, a large portion of the Boii left Italy.
Contrary to the interpretation of the classical writers, the Pannonian Boii attested in later sources are not simply the remnants of those who had fled from Italy, but rather another division of the tribe, which had settled there much earlier. The burial rites of the Italian Boii show many similarities with contemporary Bohemia, such as inhumation, which was uncommon with the other Cisalpine Gauls, or the absence of the typically western Celtic torcs. This makes it much more likely that the Cisalpine Boii had actually originated from Bohemia rather than the other way round. Having migrated to Italy from north of the Alps, some of the defeated Celts simply moved back to their kinsfolk.
The Pannonian Boii are mentioned again in the late 2nd century BC when they repelled the Cimbri and Teutones (Strabo VII, 2, 2). Later on, they attacked the city of Noreia (in modern Austria) shortly before a group of Boii (32,000 according to Julius Caesar – the number is probably an exaggeration) joined the Helvetii in their attempt to settle in western Gaul. After the Helvetian defeat at Bibracte, the influential Aedui tribe allowed the Boii survivors to settle on their territory, where they occupied the "oppidum" of Gorgobina. Although attacked by Vercingetorix during one phase of the war, they supported him with two thousand troops at the battle of Alesia (Caesar, "Commentarii de Bello Gallico", VII, 75).
Again, other parts of the Boii had remained closer to their traditional home, and settled in the Slovak and Hungarian lowlands by the Danube and the Mura, with a centre at Bratislava. Around 60 BC they clashed with the rising power of the Dacians under their king Burebista and were defeated. When the Romans finally conquered Pannonia in 8 AD, the Boii seem not to have opposed them. Their former territory was now called "deserta Boiorum" (deserta meaning 'empty or sparsely populated lands'). However, the Boii had not been exterminated: There was a "civitas Boiorum et Azaliorum" (the Azalii being a neighbouring tribe) which was under the jurisdiction of a prefect of the Danube shore ("praefectus ripae Danuvii"). This , a common Roman administrative term designating both a city and the tribal district around it, was later adjoined to the city of Carnuntum.
Plautus refers to the Boii in "Captivi":
There is a play on words: "Boia" means "woman of the Boii", also "convicted criminal's restraint collar".
In volume 21 of his "History of Rome", Livy (59 BC – 17 AD) claims that it was a Boio man that offered to show Hannibal the way across the Alps.
In the first century BC, the Boii living in an oppidum of Bratislava minted Biatecs, high-quality coins with inscriptions (probably the names of kings) in Latin letters. This is the only "written source" provided by the Boii themselves. | https://en.wikipedia.org/wiki?curid=4327 |
Backgammon
Backgammon is one of the oldest known board games. Its history can be traced back nearly 5,000 years to archeological discoveries in Mesopotamia. It is a two-player game where each player has fifteen pieces ("checkers" or "men") that move between twenty-four triangles ("points") according to the roll of two dice. The objective of the game is to be first to "bear off", i.e. move all fifteen checkers off the board. Backgammon is a member of the tables family, one of the oldest classes of board games.
Backgammon involves a combination of strategy and luck (from rolling dice). While the dice may determine the outcome of a single game, the better player will accumulate the better record over series of many games. With each roll of the dice, players must choose from numerous options for moving their checkers and anticipate possible counter-moves by the opponent. The optional use of a doubling cube allows players to raise the stakes during the game.
Like chess, backgammon has been studied with great interest by computer scientists. Owing to this research, backgammon software has been developed that is capable of beating world-class human players (see TD-Gammon for an example).
Backgammon is not controlled by a dominating authority, yet the "rules of play" are agreed by the international tournaments.
Backgammon playing pieces may be termed checkers, draughts, stones, men, counters, pawns, discs, pips, chips, or nips.
The objective is for players to remove ("bear off") all their checkers from the board before their opponent can do the same. As the playing time for each individual game is short, it is often played in matches where victory is awarded to the first player to reach a certain number of points.
Each side of the board has a track of 12 long triangles, called points. The points form a continuous track in the shape of a horseshoe, and are numbered from 1 to 24. In the most commonly used setup, each player begins with fifteen chips, two are placed on their 24-point, three on their 8-point, and five each on their 13-point and their 6-point. The two players move their chips in opposing directions, from the 24-point towards the 1-point.
Points 1 through 6 are called the home board or inner board, and points 7 through 12 are called the outer board. The 7-point is referred to as the bar point, and the 13-point as the midpoint.
To start the game, each player rolls one die, and the player with the higher number moves first using the numbers shown on both dice. If the players roll the same number, they must roll again. Both dice must land completely flat on the right-hand side of the gameboard. The players then take alternate turns, rolling two dice at the beginning of each turn.
After rolling the dice, players must, if possible, move their checkers according to the number shown on each die. For example, if the player rolls a 6 and a 3 (denoted as "6-3"), the player must move one checker six points forward, and another or the same checker three points forward. The same checker may be moved twice, as long as the two moves can be made separately and legally: six and then three, or three and then six. If a player rolls two of the same number, called doubles, that player must play each die twice. For example, a roll of 5-5 allows the player to make four moves of five spaces each. On any roll, a player must move according to the numbers on both dice if it is at all possible to do so. If one or both numbers do not allow a legal move, the player forfeits that portion of the roll and his or her turn ends. If moves can be made according to either one die or the other, but not both, the higher number must be used. If one die is unable to be moved, but such a move is made possible by the moving of the other die, that move is compulsory.
In the course of a move, a checker may land on any point that is unoccupied or is occupied by one or more of the player's own checkers. It may also land on a point occupied by exactly one opposing checker, or "blot". In this case, the blot has been "hit", and is placed in the middle of the board on the bar that divides the two sides of the playing surface. A checker may never land on a point occupied by two or more opposing checkers; thus, no point is ever occupied by checkers from both players simultaneously. There is no limit to the number of checkers that can occupy a point at any given time.
Checkers placed on the bar must re-enter the game through the opponent's home board before any other move can be made. A roll of 1 allows the checker to enter on the 24-point (opponent's 1), a roll of 2 on the 23-point (opponent's 2), and so forth, up to a roll of 6 allowing entry on the 19-point (opponent's 6). Checkers may not enter on a point occupied by two or more opposing checkers. Checkers can enter on unoccupied points, or on points occupied by a single opposing checker; in the latter case, the single checker is hit and placed on the bar. More than one checker can be on the bar at a time. A player may not move any other checkers until all checkers on the bar belonging to that player have re-entered the board. If a player has checkers on the bar, but rolls a combination that does not allow any of those checkers to re-enter, the player does not move. If the opponent's home board is completely "closed" (i.e. all six points are each occupied by two or more checkers), there is no roll that will allow a player to enter a checker from the bar, and that player stops rolling and playing until at least one point becomes open (occupied by one or zero checkers) due to the opponent's moves.
When all of a player's checkers are in that player's home board, that player may start removing them; this is called "bearing off". A roll of 1 may be used to bear off a checker from the 1-point, a 2 from the 2-point, and so on. If all of a player's checkers are on points lower than the number showing on a particular die, the player must use that die to bear off one checker from the highest occupied point. For example, if a player rolls a 6 and a 5, but has no checkers on the 6-point and two on the 5-point, then the 6 and the 5 must be used to bear off the two checkers from the 5-point. When bearing off, a player may also move a lower die roll before the higher even if that means the full value of the higher die is not fully utilized. For example, if a player has exactly one checker remaining on the 6-point, and rolls a 6 and a 1, the player may move the 6-point checker one place to the 5-point with the lower die roll of 1, and then bear that checker off the 5-point using the die roll of 6; this is sometimes useful tactically. As before, if there is a way to use all moves showing on the dice, by moving checkers within the home board or bearing them off, the player must do so. If a player's checker is hit while in the process of bearing off, that player may not bear off any others until it has been re-entered into the game and moved into his/her home board, according to the normal movement rules.
The first player to bear off all fifteen of his/her own checkers wins the game. If the opponent has not yet borne off any checkers when the game ends, the winner scores a "gammon", which counts for double stakes. If the opponent has not yet borne off any checkers and has some on the bar or in the winner's home board, the winner scores a "backgammon", which counts for triple stakes.
To speed up match play and to provide an added dimension for strategy, a doubling cube is usually used. The doubling cube is not a die to be rolled, but rather a marker with the numbers 2, 4, 8, 16, 32, and 64 inscribed on its sides, to denote the current stake. At the start of each game, the doubling cube is placed on the midpoint of the bar with the number 64 showing; the cube is then said to be "centered, on 1". When the cube is centered, either player may start their turn by proposing that the game be played for twice the current stakes. Their opponent must either accept ("take") the doubled stakes or resign ("drop") the game immediately.
Whenever a player accepts doubled stakes, the cube is placed on their side of the board with the corresponding power of two facing upward, to indicate that the right to re-double belongs exclusively to that player. If the opponent drops the doubled stakes, he or she loses the game at the current value of the doubling cube. For instance, if the cube showed the number 2 and a player wanted to redouble the stakes to put it at 4, the opponent choosing to drop the redouble would lose two, or twice the original stake.
There is no limit on the number of redoubles. Although 64 is the highest number depicted on the doubling cube, the stakes may rise to 128, 256, and so on. In money games, a player is often permitted to "beaver" when offered the cube, doubling the value of the game again, while retaining possession of the cube.
A variant of the doubling cube "beaver" is the "raccoon". Players who doubled their opponent, seeing the opponent beaver the cube, may in turn then double the stakes once again ("raccoon") as part of that cube phase before any dice are rolled. The opponent retains the doubling cube. E.g. White doubles Black to 2 points, Black accepts then beavers the cube to 4 points; White, confident of a win, raccoons the cube to 8 points, while Black retains the cube. Such a move adds greatly to the risk of having to face the doubling cube coming back at 8 times its original value when first doubling the opponent (offered at 2 points, counter offered at 16 points) should the luck of the dice change.
Some players may opt to invoke The Murphy rule or the "automatic double rule". If both opponents roll the same opening number, the doubling cube is incremented on each occasion yet remains in the middle of the board, available to either player. The Murphy rule may be invoked with a maximum number of automatic doubles allowed and that limit is agreed to prior to a game or match commencing. When a player decides to double the opponent, the value is then a double of whatever face value is shown (e.g. if two automatic doubles have occurred putting the cube up to 4, the first in-game double will be for 8 points). The Murphy rule is not an official rule in backgammon and is rarely, if ever, seen in use at officially sanctioned tournaments.
The Jacoby rule, named after Oswald Jacoby, allows gammons and backgammons to count for their respective double and triple values only if the cube has already been offered and accepted. This encourages a player with a large lead to double, possibly ending the game, rather than to play it to conclusion hoping for a gammon or backgammon. The Jacoby rule is widely used in money play but is not used in match play.
The Crawford rule, named after John R. Crawford, is designed to make match play more equitable for the player in the lead. If a player is one point away from winning a match, that player's opponent will always want to double as early as possible in order to catch up. Whether the game is worth one point or two, the trailing player must win to continue the match. To balance the situation, the Crawford rule requires that when a player first reaches a score one point short of winning, neither player may use the doubling cube for the following game, called the Crawford game. After the Crawford game, normal use of the doubling cube resumes. The Crawford rule is routinely used in tournament match play. It is possible for a Crawford game never to occur in a match.
If the Crawford rule is in effect, then another option is the Holland rule, named after
Tim Holland, which stipulates that after the Crawford game, a player cannot double until after at least two rolls have been played by each side. It was common in tournament play in the 1980s but is now rarely used.
There are many variants of standard backgammon rules. Some are played primarily throughout one geographic region, and others add new tactical elements to the game. Variants commonly alter the starting position, restrict certain moves, or assign special value to certain dice rolls, but in some geographic regions even the rules and directions of the checkers' movement change, rendering the game fundamentally different.
Acey-deucey is a variant of backgammon in which players start with no checkers on the board, and must bear them on at the beginning of the game. The roll of 1-2 is given special consideration, allowing the player, after moving the 1 and the 2, to select any desired doubles move. A player also receives an extra turn after a roll of 1-2 or of doubles.
Hypergammon is a variant of backgammon in which players have only three checkers on the board, starting with one each on the 24-, 23- and 22-points. The game has been strongly solved, meaning that exact equities are available for all 32 million possible positions.
Nard is a traditional variant from Persia in which basic rules are almost the same except that even a single piece is 'safe'. All 15 pieces start on the 24th wedge.
Nackgammon is a variant of backgammon invented by Nick "Nack" Ballard in which players start with one less checker on the six point and midpoint and two checkers on the 23 point.
Russian backgammon is a variant described in 1895 as: "...much in vogue in Russia, Germany, and other parts of the Continent...". Players start with no checkers on the board, and both players move in the same direction to bear off in a common home board. In this variant, doubles are more powerful: four moves are played as in standard backgammon, followed by four moves according to the difference of the dice value from 7, and then the player has another turn (with the caveat that the turn ends if any portion of it cannot be completed).
Gul Bara and Tapa are also variants of the game popular in southeastern Europe and Turkey. The play will iterate among Backgammon, Gul Bara, and Tapa until one of the players reaches a score of 7 or 5.
Coan ki is an ancient Chinese board game that is very similar.
Plakoto, Fevga and Portes are three versions of backgammon played in Greece. Together, the three are referred to as Tavli.
Misere (Backgammon to Lose) is a variant of backgammon in which the objective is to lose the game.
Other minor variants to the standard game are common among casual players in certain regions. For instance, only allowing a maximum of five checkers on any point (Britain) or disallowing "hit-and-run" in your home board (Middle East).
Tavla, a Turkish variation.
Backgammon has an established opening theory, although it is less detailed than that of chess. The tree of positions expands rapidly because of the number of possible dice rolls and the moves available on each turn. Recent computer analysis has offered more insight on opening plays, but the midgame is reached quickly. After the opening, backgammon players frequently rely on some established general strategies, combining and switching among them to adapt to the changing conditions of a game.
A blot has the highest probability of being hit when it is 6 points away from an opponent's checker (see picture). Strategies can derive from that. The most direct one is simply to avoid being hit, trapped, or held in a stand-off. A "running game" describes a strategy of moving as quickly as possible around the board, and is most successful when a player is already ahead in the race. When this fails, one may opt for a "holding game", maintaining control of a point on one's opponent's side of the board, called an anchor. As the game progresses, this player may gain an advantage by hitting an opponent's blot from the anchor, or by rolling large doubles that allow the checkers to escape into a running game.
The "priming game" involves building a wall of checkers, called a prime, covering a number of consecutive points. This obstructs opposing checkers that are behind the prime. A checker trapped behind a six-point prime cannot escape until the prime is broken. A particularly successful priming effort may lead to a "blitz", which is a strategy of covering the entire home board as quickly as possible while keeping one's opponent on the bar. Because the opponent has difficulty re-entering from the bar or escaping, a player can quickly gain a running advantage and win the game, often with a gammon.
A "backgame" is a strategy that involves holding two or more anchors in an opponent's home board while being substantially behind in the race. The anchors obstruct the opponent's checkers and create opportunities to hit them as they move home. The backgame is generally used only to salvage a game wherein a player is already significantly behind. Using a backgame as an initial strategy is usually unsuccessful.
"Duplication" refers to the placement of checkers such that one's opponent needs the same dice rolls to achieve different goals. For example, players may position all of their blots in such a way that the opponent must roll a 2 in order to hit any of them, reducing the probability of being hit more than once. "Diversification" refers to a complementary tactic of placing one's own checkers in such a way that more numbers are useful.
Many positions require a measurement of a player's standing in the race, for example, in making a doubling cube decision, or in determining whether to run home and begin bearing off. The minimum total of pips needed to move a player's checkers around and off the board is called the "pip count". The difference between the two players' pip counts is frequently used as a measure of the leader's racing advantage. Players often use mental calculation techniques to determine pip counts in live play.
Backgammon is played in two principal variations, "Money" and "Match" play. Money play means that every point counts evenly and every game stands alone, whether money is actually being wagered or not. "Match" play means that the players play until one side scores (or exceeds) a certain number of points. The format has a significant effect on strategy. In a match, the objective is not to win the maximum possible number of points, but rather to simply reach the score needed to win the match. For example, a player leading a 9-point match by a score of 7–5 would be very reluctant to turn the doubling cube, as their opponent could take and make a costless redouble to 4, placing the entire outcome of the match on the current game. Conversely, the trailing player would double very aggressively, particularly if they have chances to win a gammon in the current game. In money play, the theoretically correct checker play and cube action would never vary based on the score.
In 1975, Emmet Keeler and Joel Spencer considered the question of when to double or accept a double using an idealized version of backgammon. In their idealized version, the probability of winning varies randomly over time by Brownian motion, and there are no gammons or backgammons. They showed that the optimal time to offer a double was when the probability of winning reached 80%, and it is wise to accept a double only if the probability of winning is at least 20%. As their assumptions do not correspond perfectly to the real game, actual doubling strategy may vary, but the 80% number still provides a possible rule of thumb.
To reduce the possibility of cheating, most good quality backgammon sets use precision dice and a dice cup. This reduces the likelihood of loaded dice being used, which is the main way of cheating in face-to-face play. A common method of cheating online is the use of a computer program to find the optimal move on each turn; to combat this, many online sites use move-comparison software that identifies when a player's moves resemble those of a backgammon program. Online cheating has therefore become extremely difficult.
In "State of Oregon v Barr", a 1982 court case pivotal to the continued widespread organised playing of backgammon in the US, the State argued backgammon is a game of chance and that it was therefore subject to Oregon's stringent gambling laws. Paul Magriel was a key witness for the defence, contradicting Dr. Roger Nelson, the expert prosecution witness, by saying, "Game theory, however, really applies to games with imperfect knowledge, where something is concealed, such as poker. Backgammon is not such a game. Everything is in front of you. The person who uses that information in the most effective manner will win." After the closing arguments, Judge Stephen S. Walker concluded that backgammon is a game of skill, not a game of chance, and found the defendant, backgammon tournament director Ted Barr, not guilty of promoting gambling.
Early Muslim scholars forbade backgammon. The prohibition was based on sayings attributed to the Prophet Muhammad regarding gambling in general and backgammon in particular, such as: "Whoever plays backgammon it is as if he puts his hand in the pork and pig's blood." Contemporary Ayatollah Ali al-Sistani of Iraq, one of the prominent religious leaders of Shia Muslims, issued a ruling that chess (and by implication backgammon) “is absolutely forbidden even without placing a bet”.
Enthusiasts have formed clubs for social play of backgammon. Local clubs may hold informal gatherings, with members meeting at cafés and bars in the evening to play and converse. A few clubs offer additional services, maintaining their own facilities or offering computer analysis of troublesome plays. Around 2003, some club leaders noticed a growth of interest in backgammon, and attributed it to the game's popularity on the Internet.
A backgammon chouette permits three or more players to participate in a single game, often for money. One player competes against a team of all the other participants, and positions rotate after each game. Chouette play often permits the use of multiple doubling cubes.
Backgammon clubs may also organize tournaments. Large club tournaments sometimes draw competitors from other regions, with final matches viewed by hundreds of spectators. The top players at regional tournaments often compete in major national and international championships. Winners at major tournaments may receive prizes of tens of thousands of dollars.
Starting in January 2018, tournament directors began awarding GammonPoints, a free points registry for tournament directors and players, with GammonPoint awards based on the number of players and strength of field.
The first world championship competition in backgammon was held in Las Vegas, Nevada in 1967. Tim Holland was declared the winner that year and at the tournament the following year. For unknown reasons, there was no championship in 1970, but in 1971, Tim Holland again won the title. The competition remained in Las Vegas until 1975, when it moved to Paradise Island in the Bahamas. The years 1976, 1977 & 1978 saw "dual" World Championships, one in the Bahamas attended by the Americans, and the "European Open Championships" in Monte Carlo with mostly European players. In 1979, Lewis Deyong, who had promoted the Bahamas World Championship for the prior three years, suggested that the two events be combined. Monte Carlo was universally acknowledged as the site of the World Backgammon Championship and has remained as such for thirty years. The Monte Carlo tournament draws hundreds of players and spectators, and is played over the course of a week.
By the 21st century, the largest international tournaments had established the basis of a tour for top professional players. Major tournaments are held yearly worldwide. PartyGaming sponsored the first World Series of Backgammon in 2006 from Cannes and later the 'Backgammon Million' tournament held in the Bahamas in January 2007 with a prize pool of one million dollars, the largest for any tournament to date. In 2008, the World Series of Backgammon ran the world's largest international events in London, the UK Masters, the biggest tournament ever held in the UK with 128 international class players; the Nordic Open, which instantly became the largest in the world with around 500 players in all flights and 153 in the Championship, and Cannes, which hosted the Riviera Cup, the traditional follow-up tournament to the World Championships. Cannes also hosted the WSOB Championship, the WSOB finale, which saw 16 players play three-point shootout matches for €160,000. The event was recorded for television in Europe airing on Eurosport.
The World Backgammon Association (WBA) has been holding the biggest backgammon Tour of the circuit since 2007, the "European Backgammon Tour" (EBGT). In 2011, the WBA collaborated with the online backgammon provider Play65 for the 2011 season of the European Backgammon Tour and with "Betfair" in 2012. The 2013 season of the European Backgammon Tour featured 11 stops and 19 qualified players competing for 19,000 € in a Grand Finale in Lefkosa, Northern Cyprus.
When backgammon is played for money, the most common arrangement is to assign a monetary value to each point, and to play to a certain score, or until either player chooses to stop. The stakes are raised by gammons, backgammons, and use of the doubling cube. Backgammon is sometimes available in casinos. Before the commercialization of artificial neural network programs, proposition bets on specific positions were very common among backgammon players and gamblers. As with most gambling games, successful play requires a combination of luck and skill, as a single dice roll can sometimes significantly change the outcome of the game.
Backgammon software has been developed not only to play and analyze games, but also to facilitate play between humans over the internet. Dice rolls are provided by random or pseudorandom number generators. Real-time online play began with the First Internet Backgammon Server in July 1992, but there are now a range of options; many of which are commercial.
Backgammon has been studied considerably by computer scientists. Neural networks and other approaches have offered significant advances to software for gameplay and analysis.
The first strong computer opponent was BKG 9.8. It was written by Hans Berliner in the late 1970s on a DEC PDP-10 as an experiment in evaluating board game positions. Early versions of BKG played badly even against poor players, but Berliner noticed that its critical mistakes were always at transitional phases in the game. He applied principles of fuzzy logic to improve its play between phases, and by July 1979, BKG 9.8 was strong enough to play against the reigning world champion Luigi Villa. It won the match, 7–1, becoming the first computer program to defeat a world champion in any board game. Berliner stated that the victory was largely a matter of luck, as the computer received more favorable dice rolls.
In the late 1980s, backgammon programmers found more success with an approach based on artificial neural networks. TD-Gammon, developed by Gerald Tesauro of IBM, was the first of these programs to play near the expert level. Its neural network was trained using temporal difference learning applied to data generated from self-play. According to assessments by Bill Robertie and Kit Woolsey, TD-Gammon's play was at or above the level of the top human players in the world. Woolsey said of the program that "There is no question in my mind that its positional judgment is far better than mine."
Tesauro proposed using rollout analysis to compare the performance of computer algorithms against human players. In this method, a Monte-Carlo evaluation of positions is conducted (typically thousands of trials) where different random dice sequences are simulated. The rollout score of the human (or the computer) is the difference of the average game results by following the selected move versus following the best move, then averaged for the entire set of taken moves.
Neural network research has resulted in three modern proprietary programs, JellyFish, Snowie and eXtreme Gammon as well as the shareware BGBlitz and the free software GNU Backgammon. These programs not only play the game, but offer tools for analyzing games and detailed comparisons of individual moves. The strength of these programs lies in their neural networks' weights tables, which are the result of months of training. Without them, these programs play no better than a human novice. For the bearoff phase, backgammon software usually relies on a database containing precomputed equities for all possible bearoff positions.
Computer-versus-computer competitions are also held at Computer Olympiad events.
Remains of a board game from 3000 BC were found at Shahr-e Sukhteh (Persian شهر سوخته) in Iran. The artifacts included two dice and 60 checkers, and the set is believed to be 100 to 200 years older than the Royal Game of Ur. On the board found in Shahr-e Sukhteh the fields are fashioned by the coils of a snake. Board game artifacts found at Jiroft, from about 2000 BC, may represent a game that is more closely related to modern backgammon than those found at Shahr-e Sukhteh and Ur.
Touraj Daryaee (2006)—on the subject of the first written mention of early precursors of backgammon—writes:
In the 11th century "Shahnameh", the Persian poet Ferdowsi credits Burzoe with the invention of the tables game "nard" in the 6th century. He describes an encounter between Burzoe and a Raja visiting from India. The Raja introduces the game of chess, and Burzoe demonstrates "nard", played with dice made from ivory and teak. Today, Nard is the name for the Persian version of backgammon, which has different initial positions and objectives. H. J. R. Murray details many versions of backgammon; modern Nard is noted there as being the same as backgammon and maybe dating back to 300–500 AD in the Babylonian Talmud, although others believe the Talmud references the Greek race game Kubeia.
The history of backgammon can be traced back nearly 5,000 years to its origins in Mesopotamia (modern-day Iraq). The world's oldest set of dice (made from human bone) were recently discovered in that part of the world. Modern Iraqis continue to enjoy playing the game.
The Royal Game of Ur, originating in ancient Mesopotamia before 2600 BC, may also be an ancestor of modern-day table games like backgammon. It used tetrahedral dice. In the modern Middle East, backgammon is a common feature of coffeehouses. Race board games involving dice have a long history in Iraq, including the Royal Game of Ur in Babylon. Today the game is commonly played in Iraq, Lebanon, Egypt, Syria, Jordan, and other Arab countries.
In the (modern) Arab Levant and Iraq it is called "tawle", which means table, and it is also called "shesh besh" (shesh means 6 in Hebrew, Aramaic and Phoenician, but derives from Phoenician. However, shesh also means 6 in Persian and 5 is likely to be closely related to "Penj" meaning 5. Shesh Besh is commonly referred to the situation when a player scores a 5 and 6 at the same time on dice).
Backgammon or nardi () is very popular among Armenians. The word is derived from Persian word "nard" (). There are two games of nardi commonly played:
Short nardi: Set-up and rules the same as backgammon.
Long nardi: A game that starts with all fifteen checkers placed in one line-on 24-point and on 11-point. The two players move their checkers in opposing directions, from the 24-point towards the 1-point, or home board. In long nardi one checker by itself can block a point. There is no hitting in long nardi. The objective of the game is bearing all checkers off the board. There is no doubling cube.
"Tάβλη (tavli)" meaning "table" or "board" in Byzantine Greek, is the oldest game with rules known to be nearly identical to backgammon; it is described in an epigram of Byzantine Emperor Zeno (AD 476–491). The board was the same with 24 points, 12 on each side. As today, each player had 15 checkers and used cubical dice with sides numbered one to six. The object of the game, to be the first to bear off all of one's checkers, was also the same. Hitting a blot, reentering a piece from the bar, and bearing off, all followed the modern rules. The only differences with modern backgammon were the use of an extra die (three rather than two) and the starting of all pieces off the board (with them entering in the same way that pieces on the bar enter in modern backgammon). The name τάβλη is still used for backgammon in Greece, where it is frequently played in town plateias and cafes. The epigram of Zeno describes a particularly bad dice roll the emperor had for his given position. Zeno, who was white, had a stack of seven checkers, three stacks of two checkers and two "blots", checkers that stand alone on a point and are therefore in danger of being put outside the board by an incoming opponent checker. Zeno threw the three dice with which the game was played and obtained 2, 5 and 6. As in backgammon, Zeno could not move to a space occupied by two opponent (black) pieces. The white and black checkers were so distributed on the points that the only way to use all of the three results, as required by the game rules, was to break the three stacks of two checkers into blots, exposing them and ruining the game for Zeno.
The τάβλη of Zeno's time is believed to be a direct descendant of the earlier Roman "Ludus duodecim scriptorum" ("Game of twelve lines") with that board's middle row of points removed, and only the two outer rows remaining. Ludus duodecim scriptorum used a board with three rows of 12 points each with the 15 checkers being moved in opposing directions by the two players across three rows according to the roll of the three cubical dice. Little specific text about the gameplay of Ludus duodecim scriptorum has survived; it may have been related to the older Ancient Greek dice game Kubeia. The earliest known mention of the game is in Ovid's Ars Amatoria (The Art of Love) (written between 1 BC and 8 AD). In Roman times this game was also known as alea, and a likely apocryphal Latin story linked this name, and the game, to a Trojan soldier named Alea.
Race board games involving dice have existed for millennia in the Near East and eastern Mediterranean, including the game senet of Ancient Egypt. The ancient Egyptian game senet was excavated, along with illustrations, from Egyptian royal tombs dating to 3500 BC. Though using a board that is quite different from backgammon, it may be a predecessor.
Backgammon, which is known as "tavla", from Byzantine Greek τάβλη, is a very popular game in Turkey, and it is customary to name the dice rolls with their Persian number names, with local spellings: yek (1), dü (2), se (3), cehar (4), penç (5), şeş (6).
The usual Tavla rules are same as in the neighboring Arab countries and Greece, as established over a millennium ago, but there are also many quite different variants. The usual tavla is also known as erkek tavlası meaning boys' or men's tavla. The other variant kız tavlası (i.e. girls' tavla) is a game that depends only on the dice and involves no strategy. Another variant, asker tavlası (meaning soldiers' tavla), has the pieces thrown to the board randomly and the opponents try to flip their pieces over the opponents' pieces to beat them.
Backgammon is popular among Greeks. It is a game in which Greeks usually tease their opponent and they create a lively atmosphere. The game is called "Tavli", derived in Byzantine times from the Latin word "tabula". A game, almost identical to backgammon, called Tavli (Byzantine Greek: τάβλη) is described in an epigram of the Byzantine Emperor Zeno (AD 476–481). There are four games of Tavli commonly played:
Portes: Set-up and rules the same as backgammon, except that backgammons count as gammons (2 points) and there is no doubling cube.
Plakoto: A game where one checker can trap another checker on the same point.
Fevga: A game where one checker by itself can block a point.
Asodio: Also known as Acey-deucey where all checkers are off the board, and you enter by rolling either doubles or acey-deucey.
These games are played one after another, in matches of three, five, or seven points. Before starting a match, each player rolls 1 die, and the player with the highest roll picks up both dice and re-rolls (i.e. it is possible to roll doubles for the opening move). Players use the same pair of dice in turns. After the first game, the winner of the previous game starts first. Each game counts as 1 point, if the opponent has borne off at least 1 stone, otherwise 2 points (gammon/backgammon). There is no doubling cube.
Backgammon was popular in China for a time and was known as "shuanglu" (雙陸/双陆, shuānglù), with the book written during the Southern Song (1127–1279) period recording over ten variants - but over time it was replaced by other games such as xiangqi (Chinese chess).
In Japan "ban-sugoroku" is thought to have been introduced from China in the 6th century and is mentioned in Genji monogatari. As a gambling game it was made illegal several times. In the early Edo-era, a new and quick gambling game called "Chō-han" appeared and sugoroku quickly dwindled. By the 13th century, the board game Go, originally played only by the aristocracy, had become popular among the general public.
In Korea, it is called Ssang-ryuk or Jeopo.
The "jeux de tables" (Games of Tables), predecessors of modern backgammon, first appeared in France during the 11th century and became a favorite pastime of gamblers. In 1254, Louis IX issued a decree prohibiting his court officials and subjects from playing. Tables games were played in Germany in the 12th century, and had reached Iceland by the 13th century. In Spain, the Alfonso X manuscript "Libro de los juegos", completed in 1283, describes rules for a number of dice and table games in addition to its extensive discussion of chess. By the 17th century, table games had spread to Sweden. A wooden board and checkers were recovered from the wreck of the Vasa among the belongings of the ship's officers. Backgammon appears widely in paintings of this period, mainly those of Dutch and German painters (Van Ostade, Jan Steen, Hieronymus Bosch, Bruegel and others). Some surviving artworks are "Cardsharps" by Caravaggio (the backgammon board is in the lower left) and "The Triumph of Death" by Pieter Bruegel the Elder (the backgammon board is in the lower right). Others are the Hell of Bosch and interior of an Inn by Jan Steen.
In the 16th century, Elizabethan laws and church regulations prohibited playing tables, but by the 18th century, backgammon was popular among the English clergy. Edmund Hoyle published "A Short Treatise on the Game of Back-Gammon" in 1743; this described rules and strategy for the game and was bound together with a similar text on whist.
In English, the word "backgammon" is most likely derived from "back" and Middle English ""gamen"", meaning "game" or "play". The earliest use documented by the Oxford English Dictionary was in 1650.
The most recent major development in backgammon was the addition of the doubling cube. It was first introduced in the 1920s in New York City among members of gaming clubs in the Lower East Side. The cube required players not only to select the best move in a given position, but also to estimate the probability of winning from that position, transforming backgammon into the expected value-driven game played in the 20th and 21st centuries.
The popularity of backgammon surged in the mid-1960s, in part due to the charisma of Prince Alexis Obolensky who became known as "The Father of Modern Backgammon". "Obe", as he was called by friends, co-founded the International Backgammon Association, which published a set of official rules. He also established the World Backgammon Club of Manhattan, devised a backgammon tournament system in 1963, then organized the first major international backgammon tournament in March 1964, which attracted royalty, celebrities and the press. The game became a huge fad and was played on college campuses, in discothèques and at country clubs; stockbrokers and bankers began playing at conservative men's clubs. People young and old all across the country dusted off their boards and checkers. Cigarette, liquor and car companies began to sponsor tournaments and Hugh Hefner held backgammon parties at the Playboy Mansion. Backgammon clubs were formed and tournaments were held, resulting in a World Championship promoted in Las Vegas in 1967.
Most recently, the United States Backgammon Federation (USBGF) was organized in 2009 to repopularize the game in the United States. Board and committee members include many of the top players, tournament directors and writers in the worldwide backgammon community. The USBGF has recently created a Standards of Ethical Practice to address issues on which tournament rules fail to touch. | https://en.wikipedia.org/wiki?curid=4329 |
Book of Joshua
The Book of Joshua ( "") is the sixth book in the Hebrew Bible and the Christian Old Testament, and is the first book of the Deuteronomistic history, the story of Israel from the conquest of Canaan to the Babylonian exile. It tells of the campaigns of the Israelites in central, southern and northern Canaan, the destruction of their enemies, and the division of the land among the Twelve Tribes, framed by two set-piece speeches, the first by God commanding the conquest of the land, and, at the end, the last by Joshua warning of the need for faithful observance of the Law ("torah") revealed to Moses.
Almost all scholars agree that the Book of Joshua holds little historical value for early Israel and most likely reflects a much later period. The earliest parts of the book are possibly chapters 2–11, the story of the conquest; these chapters were later incorporated into an early form of Joshua written late in the reign of king Josiah (reigned 640–609 BCE), but the book was not completed until after the fall of Jerusalem to the Neo-Babylonian Empire in 586 BCE, and possibly not until after the return from the Babylonian exile in 539 BCE.
I. Transfer of leadership to Joshua (1:1–18)
II. Entrance into and conquest of Canaan (2:1–12:24)
III. Division of the land among the tribes (13:1–22:34)
IV. Conclusion (23:1–24:33)
Chapter 1 commences "after the death of Moses" (Joshua 1:1) and presents the first of three important moments in Joshua marked with major speeches and reflections by the main characters; here first God, and then Joshua, make speeches about the goal of conquest of the Promised Land; in chapter 12, the narrator looks back on the conquest; and in chapter 23 Joshua gives a speech about what must be done if Israel is to live in peace in the land.
God commissions Joshua to take possession of the land and warns him to keep faith with the Mosaic covenant. God's speech foreshadows the major themes of the book: the crossing of the Jordan River and conquest of the land, its distribution, and the imperative need for obedience to the Law. Joshua's own immediate obedience is seen in his speeches to the Israelite commanders and to the Transjordanian tribes, and the Transjordanians' affirmation of Joshua's leadership echoes Yahweh's assurances of victory.
Rahab, a Canaanite woman of the Bible, sets in motion the entrance into Canaan by the Israelites. To avoid repeating failed attempts by Moses to have notable men of Israel predict the success rate of entry into Canaan mentioned in the book of Numbers, Joshua tasks two regular men with entering Jericho as spies. They arrive at Rahab's house and spend the night. The king of Jericho, having heard of possible Israelite spies, demands that Rahab reveal the men. She tells him that she is unaware of their whereabouts, when in reality, she hid them on her roof under flax. The next morning, Rahab professes her faith in God to the men and acknowledges her belief that Canaan was meant for the Israelites to inhabit. Because of Rahab's actions, the Israelites are able to enter Canaan.
The Israelites cross the Jordan River through a miraculous intervention of God and the Ark of the Covenant and are circumcised at Gibeath-Haaraloth (translated as "hill of foreskins"), renamed Gilgal in memory. "Gilgal" sounds like "Gallothi", ""I have removed"", but is more likely to translate as ""circle of standing stones"". The conquest begins with the battle of Jericho, followed by Ai (central Canaan), after which Joshua builds an altar to Yahweh at Mount Ebal in northern Canaan and renews the Covenant in a ceremony with elements of a divine land-grant ceremony, similar to ceremonies known from Mesopotamia.
The narrative then switches to the south. The Gibeonites trick the Israelites into entering an alliance with them by saying that they are not Canaanites. This prevents the Israelites from exterminating them, but they are enslaved instead. An alliance of Amorite kingdoms headed by the Canaanite king of Jerusalem is defeated with Yahweh's miraculous help of stopping the Sun and the Moon, and hurling down large hailstones (Joshua 10:10–14). The enemy kings were eventually hanged on trees. The Deuteronomist author may have used the then-recent 701 BCE campaign of the Assyrian king Sennacherib in the Kingdom of Judah as his model; the hanging of the captured kings is in accordance with Assyrian practice of the 8th century BCE.
With the south conquered the narrative moves to the northern campaign. A powerful multi-national (or more accurately, multi-ethnic) coalition headed by the king of Hazor, the most important northern city, is defeated with Yahweh's help. Hazor itself is then captured and destroyed. Chapter 11:16–23 summarises the extent of the conquest: Joshua has taken the entire land, almost entirely through military victories, with only the Gibeonites agreeing to peaceful terms with Israel. The land then "had rest from war" (Joshua 11:23, repeated at 14:15). Chapter 12 lists the vanquished kings on both sides of the Jordan River: the two kings who ruled east of the Jordan who were defeated under Moses' leadership (Joshua 12:1-6; cf. Numbers 21), and the 31 kings on the west of the Jordan who were defeated under Joshua's leadership (Joshua 12:7–24). The list of the 31 kings is quasi-tabular:
Having described how the Israelites and Joshua have carried out the first of their God's commands, the narrative now turns to the second: to "put the people in possession of the land." Joshua is "old, advanced (or stricken) in years" by this time (Joshua 13:1).
This land distribution is a "covenantal land grant": Yahweh, as king, is issuing each tribe its territory. The "Cities of Refuge" and Levitical cities are attached to the end, since it is necessary for the tribes to receive their grants before they allocate parts of it to others. The Transjordanian tribes are dismissed, affirming their loyalty to Yahweh.
The book reaffirms Moses' allocation of land east of the Jordan to the tribes of Reuben and Gad and the half-tribe of Manasseh (Joshua 13:8–32; cf. ), and then describes how Joshua divided the newly conquered land of Canaan into parcels, and assigned them to the tribes by lot. Joshua 14:1 also makes reference to the role of Eleazar the priest (ahead of Joshua) in the distribution process. The description serves a theological function to show how the promise of the land was realized in the biblical narrative; its origins are unclear, but the descriptions may reflect geographical relations among the places named.
The wording of suggests that the tribes of Reuben, Gad, Judah, Ephraim and Manasseh received their land allocation some time before the "remaining seven tribes", and a 21-member expedition set out to survey the remainder of the land with a view to organising the allocation to the tribes of Simeon, Benjamin, Asher, Naphtali, Zebulun, Issachar and Dan. Subsequently, 48 cities with their surrounding lands were allocated to the Tribe of Levi (, cf. ).
Omitted in the Masoretic Text, but present in the Septuagint, is a statement that:
By the end of chapter 21, the narrative records that the fulfilment of God's promise of land, rest and supremacy over the enemies of the Israelites was complete (Joshua 21:43–45). The tribes to whom Moses had granted land east of the Jordan are authorized to return home to Gilead (here used in the widest sense for the whole Transjordan district), having faithfully 'kept the charge' (Joshua 22:3, English Revised Version) of supporting the tribes occupying Canaan. They are granted "riches… with very much livestock, with silver, with gold, with bronze, with iron, and with very much clothing" as a reward (Joshua 22:1–9).
Joshua, in his old age and conscious that he is "going the way of all the earth" (Joshua 23:14), gathers the leaders of the Israelites together and reminds them of Yahweh's great works for them, and of the need to love Yahweh (Joshua 23:11). The Israelites are told – just as Joshua himself had been told (Joshua 1:7) – that they must comply with "all that is written in the Book of the Law of Moses" (Joshua 23:6), neither "turn[ing] aside from it to the right hand or to the left" (i.e. by adding to the law, or diminishing from it).
Joshua meets again with all the people at Shechem in chapter 24 and addresses them a second time. He recounts the history of God's formation of the Israelite nation, beginning with "Terah, the father of Abraham and Nahor, [who] lived beyond the Euphrates River and worshiped other gods" (). He invited the Israelites to choose between serving the Lord who had delivered them from Egypt, or the gods which their ancestors had served on the other side of the Euphrates, or the gods of the Amorites in whose land they now lived. The people chose to serve the Lord, a decision which Joshua recorded in the Book of the Law of God. He then erected a memorial stone "under the oak that was by the sanctuary of the Lord" in Shechem (Joshua 24:1-27). The oak is associated with the Oak of Moreh where Abram had set up camp during his travels in this area (Genesis 12:6). Thus "Joshua made a covenant with the people", literally "cut a covenant", a phrase common to the Hebrew, Greek, and Latin languages. It derives from the custom of sacrifice, in which the victims were cut in pieces and offered to the deity invoked in ratification of the engagement.
The people then returned to their inheritance i.e. their allocated lands (Joshua 24:28).
The Book of Joshua closes with three concluding items (referred to in the Jerusalem Bible as "Two Additions"):
There were no Levitical cities given to the descendants of Aaron in Ephraim, so theologians Carl Friedrich Keil and Franz Delitzsch supposed the land may have been at Geba in the territory of the Tribe of Benjamin: "the situation, 'upon the mountains of Ephraim', is not at variance with this view, as these mountains extended, according to Judges 4:5, etc., far into the territory of Benjamin".
In some manuscripts and editions of the Septuagint, there is an additional verse relating to the apostasy of the Israelites after Joshua's death.
The Book of Joshua is an anonymous work. The Babylonian Talmud, written in the 3rd to 5th centuries CE, attributed it to Joshua himself, but this idea was rejected as untenable by John Calvin (1509–64), and by the time of Thomas Hobbes (1588–1679) it was recognised that the book must have been written much later than the period it depicted. There is now general agreement that it was composed as part of a larger work, the Deuteronomistic history, stretching from the Book of Deuteronomy to the Books of Kings. composed first at the court of king Josiah in the late 7th century BCE, and extensively revised in the 6th century BCE.
The prevailing scholarly view is that Joshua is not a factual account of historical events. The apparent setting of Joshua is the 13th century BCE which was a time of widespread city-destruction, but with a few exceptions (Hazor, Lachish) the destroyed cities are not the ones the Bible associates with Joshua, and the ones it does associate with him show little or no sign of even being occupied at the time. Given its lack of historicity, Carolyn Pressler in her commentary for the "Westminster Bible Companion" series suggests that readers of Joshua should give priority to its theological message ("what passages teach about God") and be aware of what these would have meant to audiences in the 7th and 6th centuries BCE. Richard Nelson explained that the needs of the centralised monarchy favoured a single story of origins, combining old traditions of an exodus from Egypt, belief in a national god as "divine warrior," and explanations for ruined cities, social stratification and ethnic groups, and contemporary tribes.
In the 1930s Martin Noth made a sweeping criticism of the usefulness of the Book of Joshua for history. Noth was a student of Albrecht Alt, who emphasized form criticism and the importance of etiology. Alt and Noth posited a peaceful movement of the Israelites into various areas of Canaan, "contra" the Biblical account.
Albright questioned the "tenacity" of etiologies, which were key to Noth's analysis of the campaigns in Joshua. Archaeological evidence in the 1930s showed that the city of Ai, an early target for conquest in the putative Joshua account, had existed and been destroyed, but in the 22nd century BCE. Some alternate sites for Ai have been proposed which would partially resolve the discrepancy in dates, but these sites have not been widely accepted. In 1951 Kathleen Kenyon showed that Jericho was from the Middle Bronze Age (c. 2100–1550 BCE), not the Late Bronze Age (c. 1550–1200 BCE). Kenyon argued that the early Israelite campaign could not be historically corroborated, but rather explained as an etiology of the location and a representation of the Israelite settlement.
In 1955, G. Ernest Wright discussed the correlation of archaeological data to the early Israelite campaigns, which he divided into three phases per the Book of Joshua. He pointed to two sets of archaeological findings that "seem to suggest that the biblical account is in general correct regarding the nature of the late thirteenth and twelfth-eleventh centuries in the country" (i.e., "a period of tremendous violence"). He gives particular weight to what were then recent digs at Hazor by Yigael Yadin.
As an alternative to both the military conquest and uncontested infiltration hypotheses, George Mendenhall and Norman Gottwald suggested that the Israelites emerged through a kind of peasant revolt against their Canaanite lords. Arguing against this view Gary A. Rendsburg says that archaeological findings (presented by Israel Finkelstein in 1986) undermined this idea because the Israelites' first settled areas were not held by Canaanites, whose cities were sustained alongside Israelite areas.
The Book of Joshua holds little historical value. The archaeological evidence shows that Jericho and Ai were not occupied in the Near Eastern Late Bronze Age. The story of the conquest most likely represents the nationalist propaganda of the 8th century BCE kings of Judah and their claims to the territory of the Kingdom of Israel; incorporated into an early form of Joshua written late in the reign of king Josiah (reigned 640–609 BCE). The book was probably revised and completed after the fall of Jerusalem to the Neo-Babylonian Empire in 586 BCE, and possibly after the return from the Babylonian exile in 538 BCE.
The overarching theological theme of the Deuteronomistic history is faithfulness and God's mercy, and their opposites, faithlessness and God's wrath. In the Book of Judges, the Books of Samuel, and the Books of Kings, the Israelites become faithless and God ultimately shows his anger by sending his people into exile. But in Joshua Israel is obedient, Joshua is faithful, and God fulfills his promise and gives them the land. Yahweh's war campaign in Canaan validates Israel's entitlement to the land and provides a paradigm of how Israel was to live there: twelve tribes, with a designated leader, united by covenant in warfare and in worship of Yahweh alone at a single sanctuary, all in obedience to the commands of Moses as found in the Book of Deuteronomy.
The Book of Joshua takes forward Deuteronomy's theme of Israel as a single people worshipping Yahweh in the land God has given them. Yahweh, as the main character in the book, takes the initiative in conquering the land, and Yahweh's power wins the battles. For example, the walls of Jericho fall because Yahweh fights for Israel, not because the Israelites show superior fighting ability. The potential disunity of Israel is a constant theme, the greatest threat of disunity coming from the tribes east of the Jordan. Chapter 22:19 even hints that the land across the Jordan is unclean and that the tribes who live there have secondary status.
Land is the central topic of Joshua. The introduction to Deuteronomy recalled how Yahweh had given the land to the Israelites but then withdrew the gift when Israel showed fear and only Joshua and Caleb had trusted in God. The land is Yahweh's to give or withhold, and the fact that he has promised it to Israel gives Israel an inalienable right to take it. For exilic and post-exilic readers, the land was both the sign of Yahweh's faithfulness and Israel's unfaithfulness, as well as the centre of their ethnic identity. In Deuteronomistic theology, "rest" meant Israel's unthreatened possession of the land, the achievement of which began with the conquests of Joshua.
Joshua "carries out a systematic campaign against the civilians of Canaan — men, women and children — that amounts to genocide." In doing this he is carrying out "herem" as commanded by Yahweh in Deuteronomy 20:17: "You shall not leave alive anything that breathes". The purpose is to drive out and dispossess the Canaanites, with the implication that there are to be no treaties with the enemy, no mercy, and no intermarriage. "The extermination of the nations glorifies Yahweh as a warrior and promotes Israel's claim to the land," while their continued survival "explores the themes of disobedience and penalty and looks forward to the story told in Judges and Kings." The divine call for massacre at Jericho and elsewhere can be explained in terms of cultural norms (Israel was not the only Iron Age state to practice "herem") and theology (a measure to ensure Israel's purity as well as the fulfillment of God's promise), but Patrick D. Miller in his commentary on Deuteronomy remarks, "there is no real way to make such reports palatable to the hearts and minds of contemporary readers and believers."
Obedience versus disobedience is a constant theme of the work. Obedience ties in the Jordan crossing, the defeat of Jericho and Ai, circumcision and Passover, and the public display and reading of the Law. Disobedience appears in the story of Achan (stoned for violating the "herem" command), the Gibeonites, and the altar built by the Transjordan tribes. Joshua's two final addresses challenge the Israel of the future (the readers of the story) to obey the most important command of all, to worship Yahweh and no other gods. Joshua thus illustrates the central Deuteronomistic message, that obedience leads to success and disobedience to ruin.
The Deuteronomistic history draws parallels in proper leadership between Moses, Joshua and Josiah. God's commission to Joshua in chapter 1 is framed as a royal installation. The people's pledge of loyalty to Joshua as the successor of Moses recalls royal practices. The covenant-renewal ceremony led by Joshua was the prerogative of the kings of Judah. God's command to Joshua to meditate on the "book of the law" day and night parallels the description of Josiah in 2 Kings 23:25 as a king uniquely concerned with the study of the law. The two figures had identical territorial goals; Josiah died in 609 BCE while attempting to annex the former Israel to his own kingdom of Judah.
Some of the parallels with Moses can be seen in the following, and not exhaustive, list:
Joshua has become an iconic figure for the Jewish Zionist movement, and many Israeli settlements sit on land taken by force from Palestinians. David Ben-Gurion saw in the war narrative of Joshua an ideal basis for a unifying national myth for his new State of Israel, framed against a common enemy, the Arabs. He met with politicians and scholars such as Bible scholar Shemaryahu Talmon to discuss Joshua's supposed conquests and later published a book of the meeting transcripts; in a lecture at Ben-Gurion's home, archaeologist Yigael Yadin argued for the historicity of the Israelite military campaign pointing to the conquests of Hazor, Bethel, and Lachish. Zionism thus presented the 1948 war (the war which saw the creation of the State of Israel) as a "miraculous" clearing of the land based on Joshua, and the Bible as a mandate for the expulsion of the Palestinians.
By the same token, the Biblical narrative of conquest has been used as an apparatus of critique against Zionism. For example, Michael Prior criticizes the use of the campaign in Joshua to favor "colonial enterprises" (in general, not only Zionism) and have been interpreted as validating ethnic cleansing. He asserts that the Bible was used to make the treatment of Palestinians more palatable morally. A related moral condemnation can be seen in "The political sacralization of imperial genocide: contextualizing Timothy Dwight's "The Conquest of Canaan"" by Bill Templer. This kind of critique is not new; Jonathan Boyarin notes how Frederick W. Turner blamed Israel's monotheism for the very idea of genocide, which Boyarin found "simplistic" yet with precedents. | https://en.wikipedia.org/wiki?curid=4331 |
Book of Daniel
The Book of Daniel is a 2nd-century BC biblical apocalypse combining a prophecy of history with an eschatology (a portrayal of end times) cosmic in scope and political in focus. In more mundane language, it is "an account of the activities and visions of Daniel, a noble Jew exiled at Babylon," its message being that just as the God of Israel saved Daniel and his friends from their enemies, so he would save all of Israel in their present oppression.
In the Hebrew Bible, it is found in the "Ketuvim" (writings), while in Christian Bibles it is grouped with the Major Prophets. The book divides into two parts, a set of six court tales in chapters 1–6 written mostly in Aramaic, followed by four apocalyptic visions in chapters 7–12, written mostly in Hebrew. The deuterocanon contains three additional stories: the Song of the Three Holy Children, Susanna, and Bel and the Dragon.
The book's influence has resonated through later ages, from the Dead Sea Scrolls community and the authors of the gospels and Revelation, to various movements from the 2nd century to the Protestant Reformation and modern millennialist movements—on which it continues to have a profound influence.
The Book of Daniel is divided between the court tales of chapters 1–6 and the apocalyptic visions of 7–12, and between the Hebrew of chapters 1 and 8–12 and the Aramaic of chapters 2–7. The division is reinforced by the chiastic arrangement of the Aramaic chapters (see below), and by a chronological progression in chapters 1–6 from Babylonian to Median rule, and from Babylonian to Persian rule in chapters 7–12. Various suggestions have been made by scholars to explain the fact that the genre division does not coincide with the other two, but it appears that the language division and concentric structure of chapters 2–6 are artificial literary devices designed to bind the two halves of the book together. The following outline is provided by Collins in his commentary on Daniel:
PART I: Tales (chapters 1:1–6:29)
PART II: Visions (chapters 7:1–12:13)
There is a recognised chiasm (a concentric literary structure in which the main point of a passage is placed in the centre and framed by parallel elements on either side in "ABBA" fashion) in the chapter arrangement of the Aramaic section. The following is taken from Paul Redditt's "Introduction to the Prophets":
In the third year of King Jehoiakim, God allows Jerusalem to fall into the power of Nebuchadnezzar II, king of Babylon. Young Israelites of noble and royal family, "without physical defect, and handsome," versed in wisdom and competent to serve in the palace of the king, are taken to Babylon to be taught the literature and language of that nation. Among them are Daniel and his three companions, who refuse to touch the royal food and wine. Their overseer fears for his life in case the health of his charges deteriorates, but Daniel suggests a trial and the four emerge healthier than their counterparts from ten days of nothing but vegetables and water. They are allowed to continue to refrain from eating the king's food, and to Daniel God gives insight into visions and dreams. When their training is done Nebuchadnezzar finds them 'ten times better' than all the wise men in his service and therefore keeps them at his court, where Daniel continues until the first year of King Cyrus.
In the second year of his reign Nebuchadnezzar has a dream. When he wakes up, he realizes that the dream has some important message, so he consults his wise men. Wary of their potential to fabricate an explanation, the king refuses to tell the wise men what he saw in his dream. Rather, he demands that his wise men tell him what the content of the dream was, and then interpret it. When the wise men protest that this is beyond the power of any man, he sentences all, including Daniel and his friends, to death. Daniel receives an explanatory vision from God: Nebuchadnezzar had seen an enormous statue with a head of gold, breast and arms of silver, belly and thighs of bronze, legs of iron, and feet of mixed iron and clay, then saw the statue destroyed by a rock that turned into a mountain filling the whole earth. Daniel explains the dream to the king: the statue symbolized four successive kingdoms, starting with Nebuchadnezzar, all of which would be crushed by God's kingdom, which would endure forever. Nebuchadnezzar acknowledges the supremacy of Daniel's god, raises Daniel over all his wise men, and places Daniel and his companions over the province of Babylon.
Daniel's companions Shadrach, Meshach, and Abednego refuse to bow to King Nebuchadnezzar's golden statue and are thrown into a fiery furnace. Nebuchadnezzar is astonished to see a fourth figure in the furnace with the three, one "with the appearance like a son of the gods." So the king called the three to come out of the fire, and blessed the God of Israel, and decreed that any who blasphemed against him should be torn limb from limb.
Nebuchadnezzar recounts a dream of a huge tree that is suddenly cut down at the command of a heavenly messenger. Daniel is summoned and interprets the dream. The tree is Nebuchadnezzar himself, who for seven years will lose his mind and live like a wild beast. All of this comes to pass until, at the end of the specified time, Nebuchadnezzar acknowledges that "heaven rules" and his kingdom and sanity are restored.
Belshazzar and his nobles blasphemously drink from sacred Jewish temple vessels, offering praise to inanimate gods, until a hand mysteriously appears and writes upon the wall. The horrified king summons Daniel, who upbraids him for his lack of humility before God and interprets the message: Belshazzar's kingdom will be given to the Medes and Persians. Belshazzar rewards Daniel and raises him to be third in the kingdom, and that very night Belshazzar is slain and Darius the Mede takes the kingdom.
Darius elevates Daniel to high office, exciting the jealousy of other officials. Knowing of Daniel's devotion to his God, his enemies trick the king into issuing an edict forbidding worship of any other god or man for a 30-day period. Daniel continues to pray three times a day to God towards Jerusalem; he is accused and King Darius, forced by his own decree, throws Daniel into the lions' den. But God shuts up the mouths of the lions, and the next morning Darius rejoices to find him unharmed. The king casts Daniel's accusers into the lions' pit together with their wives and children to be instantly devoured, while he himself acknowledges Daniel's God as he whose kingdom shall never be destroyed.
In the first year of Belshazzar Daniel has a dream of four monstrous beasts arising from the sea. The fourth, a beast with ten horns, devours the whole earth, treading it down and crushing it, and a further small horn appears and uproots three of the earlier horns. The Ancient of Days judges and destroys the beast, and "one like a son of man" is given everlasting kingship over the entire world. A divine being explains that the four beasts represent four kings, but that "the holy ones of the Most High" would receive the everlasting kingdom. The fourth beast would be a fourth kingdom with ten kings, and another king who would pull down three kings and make war on the "holy ones" for "a time, two times and a half," after which the heavenly judgement will be made against him and the "holy ones" will receive the everlasting kingdom.
In the third year of Belshazzar Daniel has vision of a ram and goat. The ram has two mighty horns, one longer than the other, and it charges west, north and south, overpowering all other beasts. A goat with a single horn appears from the west and destroys the ram. The goat becomes very powerful until the horn breaks off and is replaced by four lesser horns. A small horn that grows very large, it stops the daily temple sacrifices and desecrates the sanctuary for two thousand three hundred "evening and mornings" (which could be either 1,150 or 2,300 days) until the temple is cleansed. The angel Gabriel informs him that the ram represents the Medes and Persians, the goat is Greece, and the "little horn" is a wicked king.
In the first year of Darius the Mede, Daniel meditates on the word of Jeremiah that the desolation of Jerusalem would last seventy years; he confesses the sin of Israel and pleads for God to restore Israel and the "desolated sanctuary" of the Temple. The angel Gabriel explains that the seventy years stand for seventy "weeks" of years (490 years), during which the Temple will first be restored, then later defiled by a "prince who is to come," "until the decreed end is poured out."
Daniel 10: In the third year of Cyrus Daniel sees in his vision an angel (called "a man", but clearly a supernatural being) who explains that he is in the midst of a war with the "prince of Persia", assisted only by Michael, "your prince." The "prince of Greece" will shortly come, but first he will reveal what will happen to Daniel's people.
Daniel 11: A future king of Persia will make war on the king of Greece, a "mighty king" will arise and wield power until his empire is broken up and given to others, and finally the king of the south (identified in verse 8 as Egypt) will go to war with the "king of the north." After many battles (described in great detail) a "contemptible person" will become king of the north; this king will invade the south two times, the first time with success, but on his second he will be stopped by "ships of Kittim." He will turn back to his own country, and on the way his soldiers will desecrate the Temple, abolish the daily sacrifice, and set up the abomination of desolation. He will defeat and subjugate Libya and Egypt, but "reports from the east and north will alarm him," and he will meet his end "between the sea and the holy mountain."
Daniel 12: At this time Michael will come. It will be a time of great distress, but all those whose names are written will be delivered. "Multitudes who sleep in the dust of the earth will awake, some to everlasting life, others to shame and everlasting contempt; those who are wise will shine like the brightness of the heavens, and those who lead many to righteousness, like the stars for ever and ever." In the final verses the remaining time to the end is revealed: "a time, times and half a time" (three years and a half). Daniel fails to understand and asks again what will happen, and is told: "From the time that the daily sacrifice is abolished and the abomination that causes desolation is set up, there will be 1,290 days. Blessed is the one who waits for and reaches the end of the 1,335 days."
The Greek text of Daniel is considerably longer than the Hebrew, due to three additional stories: they remain in Catholic and Orthodox Bibles but were rejected by the Protestant movement in the 16th century on the basis that they were absent from Jewish Bibles.
The visions of chapters 7–12 reflect the crisis which took place in Judea in 167–164 BC when Antiochus IV Epiphanes, the Greek king of the Seleucid Empire, threatened to destroy traditional Jewish worship in Jerusalem. When Antiochus came to the throne in 175 BC the Jews were largely pro-Seleucid. The High Priestly family was split by rivalry, and one member, Jason, offered the king a large sum to be made High Priest. Jason also asked—or more accurately, paid—to be allowed to make Jerusalem a "polis", or Greek city. This meant, among other things, that city government would be in the hands of the citizens, which meant in turn that citizenship would be a valuable commodity, to be purchased from Jason. None of this threatened the Jewish religion, and the reforms were widely welcomed, especially among the Jerusalem aristocracy and the leading priests. Three years later Jason was deposed when another priest, Menelaus, offered Antiochus an even larger sum for the post of High Priest.
Antiochus invaded Egypt twice, in 169 BC with success, but on the second incursion, in late 168 BC, he was forced to withdraw by the Romans. Jason, hearing a rumour that Antiochus was dead, attacked Menelaus to take back the High Priesthood. Antiochus drove Jason out of Jerusalem, plundered the Temple, and introduced measures to pacify his Egyptian border by imposing complete Hellenisation: the Jewish Book of the Law was prohibited and on 15 December 167 BC an "abomination of desolation", probably a Greek altar, was introduced into the Temple. With the Jewish religion now clearly under threat a resistance movement sprang up, led by the Maccabee brothers, and over the next three years it won sufficient victories over Antiochus to take back and purify the Temple.
The crisis which the author of Daniel addresses is the defilement of the altar in Jerusalem in 167 BC (first introduced in chapter 8:11): the daily offering which used to take place twice a day, at morning and evening, stopped, and the phrase "evenings and mornings" recurs through the following chapters as a reminder of the missed sacrifices. But whereas the events leading up to the sacking of the Temple in 167 BC and the immediate aftermath are remarkably accurate, the predicted war between the Syrians and the Egyptians (11:40–43) never took place, and the prophecy that Antiochus would die in Palestine (11:44–45) was inaccurate (he died in Persia). The obvious conclusion is that the account must have been completed near the end of the reign of Antiochus but before his death in December 164 BC, or at least before news of it reached Jerusalem, and the consensus of modern scholarship is accordingly that the book dates to the period 167–163 BCE.
It is generally accepted that Daniel originated as a collection of Aramaic court tales later expanded by the Hebrew revelations. The court tales may have originally circulated independently, but the edited collection was probably composed in the third or early second century BC. Chapter 1 was composed (in Aramaic) at this time as a brief introduction of to provide historical context, introduce the characters of the tales, and explain how Daniel and his friends came to Babylon. The visions of chapters 7–12 were added and chapter 1 translated into Hebrew at the third stage when the final book was being drawn together.
Daniel is one of a large number of Jewish apocalypses, all of them pseudonymous. The stories of the first half are considered legendary in origin, and the visions of the second the product of anonymous authors in the Maccabean period (2nd century BC).
Although the entire book is traditionally ascribed to Daniel the seer, chapters 1–6 are in the voice of an anonymous narrator, except for chapter 4 which is in the form of a letter from king Nebuchadnezzar; only the second half (chapters 7–12) is presented by Daniel himself, introduced by the anonymous narrator in chapters 7 and 10. The real author/editor of Daniel was probably an educated Jew, knowledgeable in Greek learning, and of high standing in his own community. The book is a product of "Wisdom" circles, but the type of wisdom is (the discovery of heavenly secrets from earthly signs) rather than the wisdom of learning—the main source of wisdom in Daniel is God's revelation.
It is possible that the name of Daniel was chosen for the hero of the book because of his reputation as a wise seer in Hebrew tradition. Ezekiel, who lived during the Babylonian exile, mentioned him in association with Noah and Job (Ezekiel 14:14) as a figure of legendary wisdom (28:3), and a hero named Daniel (more accurately Dan'el, but the spelling is close enough for the two to be regarded as identical) features in a late 2nd millennium myth from Ugarit. "The legendary Daniel, known from long ago but still remembered as an exemplary character ... serves as the principal human 'hero' in the biblical book that now bears his name"; Daniel is the wise and righteous intermediary who is able to interpret dreams and thus convey the will of God to humans, the recipient of visions from on high that are interpreted to him by heavenly intermediaries.
The prophecies of Daniel are accurate down to the career of Antiochus IV Epiphanes, king of Syria and oppressor of the Jews, but not in its prediction of his death: the author seems to know about Antiochus' two campaigns in Egypt (169 and 167 BC), the desecration of the Temple (the "abomination of desolation"), and the fortification of the Akra (a fortress built inside Jerusalem), but he seems to know nothing about the reconstruction of the Temple or about the actual circumstances of Antiochus' death in late 164 BC. Chapters 10–12 must therefore have been written between 167 and 164 BC. There is no evidence of a significant time lapse between those chapters and chapters 8 and 9, and chapter 7 may have been written just a few months earlier again.
Further evidence of the book's date is in the fact that Daniel is excluded from the Hebrew Bible's canon of the prophets, which was closed around 200 BC, and the Wisdom of Sirach, a work dating from around 180 BC, draws on almost every book of the Old Testament except Daniel, leading scholars to suppose that its author was unaware of it. Daniel is, however, quoted in a section of the Sibylline Oracles commonly dated to the middle of the 2nd century BC, and was popular at Qumran at much the same time, suggesting that it was known from the middle of that century.
The Book of Daniel is preserved in the 12-chapter Masoretic Text and in two longer Greek versions, the original Septuagint version, c. 100 BC, and the later Theodotion version from c. 2nd century AD. Both Greek texts contain three additions to Daniel: The Prayer of Azariah and Song of the Three Holy Children; the story of Susannah and the Elders; and the story of Bel and the Dragon. Theodotion is much closer to the Masoretic Text and became so popular that it replaced the original Septuagint version in all but two manuscripts of the Septuagint itself. The Greek additions were apparently never part of the Hebrew text.
Eight copies of the Book of Daniel, all incomplete, have been found at Qumran, two in Cave 1, five in Cave 4, and one in Cave 6. Between them, they preserve text from eleven of Daniel's twelve chapters, and the twelfth is quoted in the Florilegium (a compilation scroll) 4Q174, showing that the book at Qumran did not lack this conclusion. All eight manuscripts were copied between 125 BC (4QDanc) and about 50 AD (4QDanb), showing that Daniel was being read at Qumran only about 40 years after its composition. All appear to preserve the 12-chapter Masoretic version rather than the longer Greek text. None reveal any major disagreements against the Masoretic, and the four scrolls that preserve the relevant sections (1QDana, 4QDana, 4QDanb, and 4QDand) all follow the bilingual nature of Daniel where the book opens in Hebrew, switches to Aramaic at 2:4b, then reverts to Hebrew at 8:1.
"(This section deals with modern scholarly reconstructions of the meaning of Daniel to its original authors and audience)"
The Book of Daniel is an apocalypse, a literary genre in which a heavenly reality is revealed to a human recipient; such works are characterized by visions, symbolism, an other-worldly mediator, an emphasis on cosmic events, angels and demons, and pseudonymity (false authorship). The production of apocalypses occurred commonly from 300 BC to 100 AD, not only among Jews and Christians, but also among Greeks, Romans, Persians and Egyptians, and Daniel is a representative apocalyptic seer, the recipient of divine revelation: he has learned the wisdom of the Babylonian magicians and surpassed them, because his God is the true source of knowledge; he is one of the "maskilim" (משכלים), the wise ones, who have the task of teaching righteousness and whose number may be considered to include the authors of the book itself. The book is also an eschatology, as the divine revelation concerns the end of the present age, a predicted moment in which God will intervene in history to usher in the final kingdom. It gives no real details of the end-time, but it seems that God's kingdom will be on this earth, that it will be governed by justice and righteousness, and that the tables will be turned on the Seleucids and those Jews who have cooperated with them.
The message of the Book of Daniel is that, just as the God of Israel saved Daniel and his friends from their enemies, so he would save all Israel in their present oppression. The book is filled with monsters, angels, and numerology, drawn from a wide range of sources, both biblical and non-biblical, that would have had meaning in the context of 2nd-century Jewish culture, and while Christian interpreters have always viewed these as predicting events in the New Testament—"the Son of God", "the Son of Man", Christ and the Antichrist—the book's intended audience is the Jews of the 2nd century BC. The following explains a few of these predictions, as understood by modern biblical scholars.
The concepts of immortality and resurrection, with rewards for the righteous and punishment for the wicked, have roots much deeper than Daniel, but the first clear statement is found in the final chapter of that book: "Many of those who sleep in the dust of the earth shall awake, some to everlasting life, and some to everlasting shame and contempt." Without this belief, Christianity, in which the resurrection of Jesus plays a central role, would have disappeared, like the movements following other charismatic Jewish figures of the 1st century.
Daniel was quoted and referenced by both Jews and Christians in the 1st century AD as predicting the imminent end-time. Moments of national and cultural crisis continually reawakened the apocalyptic spirit, through the Montanists of the 2nd/3rd centuries, persecuted for their millennialism, to the more extreme elements of the 16th-century Reformation such as the Zwickau prophets and the Münster Rebellion. During the English Civil War, the Fifth Monarchy Men took their name and political program from Daniel 7, demanding that Oliver Cromwell allow them to form a "government of saints" in preparation for the coming of the Messiah; when Cromwell refused, they identified him instead as the Beast usurping the rightful place of King Jesus. Daniel remains one of the most influential apocalypses in modern America, foretelling the history of Jesus and the Second Coming.
The influence of Daniel has not been confined to Judaism and Christianity: In the Middle Ages Muslims created horoscopes whose authority was attributed to Daniel. More recently the Bahá'í movement, which originated in Persian Shi'ite Islam, justified its existence on the 1260-day prophecy of Daniel, holding that it foretold the coming of the Twelfth Imam and an age of peace and justice in the year 1844, which is the year 1260 of the Muslim era.
Daniel belongs not only to the religious tradition but also to the wider Western intellectual and artistic heritage. It was easily the most popular of the prophetic books for the Anglo-Saxons, who nevertheless treated it not as prophecy but as a historical book, "a repository of dramatic stories about confrontations between God and a series of emperor-figures who represent the highest reach of man". In the early modern period the physicist Isaac Newton paid special attention to it, and Francis Bacon borrowed a motto from it for his work "Novum Organum". Philosophers, such as Baruch Spinoza drew on it. In the 20th century its apocalyptic second half attracted the attention of Carl Jung. The book has also inspired musicians, from medieval liturgical drama to the 20th century compositions of Darius Milhaud. Artists including Michelangelo, Rembrandt and Eugène Delacroix have all drawn on its imagery. | https://en.wikipedia.org/wiki?curid=4333 |
Batman
Batman is a fictional superhero appearing in American comic books published by DC Comics. The character was created by artist Bob Kane and writer Bill Finger, and first appeared in "Detective Comics" #27 in 1939. Originally named the "Bat-Man," the character is also referred to by such epithets as the Caped Crusader, the Dark Knight, and the World's Greatest Detective.
Batman's secret identity is Bruce Wayne, a wealthy American playboy, philanthropist, and owner of Wayne Enterprises. Batman originated from an incident in Bruce's childhood; after witnessing the murder of his parents Dr. Thomas Wayne and Martha Wayne, he swore vengeance against criminals, an oath tempered by a sense of justice. Bruce trains himself physically and intellectually and crafts a bat-inspired persona to fight crime.
Batman operates in the fictional Gotham City with assistance from various supporting characters, including his butler Alfred, police commissioner Jim Gordon, and vigilante allies such as Robin. Unlike most superheroes, Batman does not possess any inhuman superpowers. He does, however, possess a genius-level intellect and is a peerless martial artist, and his vast wealth affords him an extraordinary arsenal of weaponry and equipment. A large assortment of villains make up Batman's rogues gallery, including his nemesis, the Joker.
The character became popular soon after his introduction in 1939 and gained his own comic book title, "Batman", the following year. As the decades went on, different interpretations of the character emerged. The late 1960s "Batman" television series used a camp aesthetic, which continued to be associated with the character for years after the show ended. Various creators worked to return the character to his dark roots, culminating in 1986 with "The Dark Knight Returns" by Frank Miller. The success of Warner Bros. Pictures' live-action "Batman" feature films have helped maintain the character's prominence in mainstream culture.
Batman has been licensed and featured in various adaptations, from radio to television and film, and appears in merchandise sold around the world, such as apparel, toys, and video games. Kevin Conroy, Rino Romano, Anthony Ruivivar, Peter Weller, Bruce Greenwood, Jason O'Mara, and Will Arnett, among others, have provided the character's voice for animated adaptations. Batman has been depicted in both film and television by Lewis Wilson, Robert Lowery, Adam West, Michael Keaton, Val Kilmer, George Clooney, Bruce Thomas, Christian Bale, Ben Affleck, David Mazouz, Iain Glen, Kevin Conroy, and Warren Christie. Robert Pattinson will portray the character in the upcoming 2021 film, "The Batman".
In early 1939, the success of Superman in "Action Comics" prompted editors at National Comics Publications (the future DC Comics) to request more superheroes for its titles. In response, Bob Kane created "the Bat-Man". Collaborator Bill Finger recalled that "Kane had an idea for a character called 'Batman,' and he'd like me to see the drawings. I went over to Kane's, and he had drawn a character who looked very much like Superman with kind of ... reddish tights, I believe, with boots ... no gloves, no gauntlets ... with a small domino mask, swinging on a rope. He had two stiff wings that were sticking out, looking like bat wings. And under it was a big sign ... BATMAN". The bat-wing-like cape was suggested by Bob Kane, inspired as a child by Leonardo Da Vinci's sketch of an ornithopter flying device.
Finger suggested giving the character a cowl instead of a simple domino mask, a cape instead of wings, and gloves; he also recommended removing the red sections from the original costume. Finger said he devised the name Bruce Wayne for the character's secret identity: "Bruce Wayne's first name came from Robert Bruce, the Scottish patriot. Wayne, being a playboy, was a man of gentry. I searched for a name that would suggest colonialism. I tried Adams, Hancock ... then I thought of Mad Anthony Wayne." He later said his suggestions were influenced by Lee Falk's popular "The Phantom", a syndicated newspaper comic-strip character with which Kane was also familiar.
Kane and Finger drew upon contemporary 1930s popular culture for inspiration regarding much of the Bat-Man's look, personality, methods, and weaponry. Details find predecessors in pulp fiction, comic strips, newspaper headlines, and autobiographical details referring to Kane himself. As an aristocratic hero with a double identity, Batman had predecessors in the Scarlet Pimpernel (created by Baroness Emmuska Orczy, 1903) and Zorro (created by Johnston McCulley, 1919). Like them, Batman performed his heroic deeds in secret, averted suspicion by playing aloof in public, and marked his work with a signature symbol. Kane noted the influence of the films "The Mark of Zorro" (1920) and "The Bat Whispers" (1930) in the creation of the character's iconography. Finger, drawing inspiration from pulp heroes like Doc Savage, The Shadow, Dick Tracy, and Sherlock Holmes, made the character a master sleuth.
In his 1989 autobiography, Kane detailed Finger's contributions to Batman's creation:
Kane signed away ownership in the character in exchange for, among other compensation, a mandatory byline on all Batman comics. This byline did not originally say "Batman created by Bob Kane"; his name was simply written on the title page of each story. The name disappeared from the comic book in the mid-1960s, replaced by credits for each story's actual writer and artists. In the late 1970s, when Jerry Siegel and Joe Shuster began receiving a "created by" credit on the Superman titles, along with William Moulton Marston being given the byline for creating Wonder Woman, Batman stories began saying "Created by Bob Kane" in addition to the other credits.
Finger did not receive the same recognition. While he had received credit for other DC work since the 1940s, he began, in the 1960s, to receive limited acknowledgment for his Batman writing; in the letters page of "Batman" #169 (February 1965) for example, editor Julius Schwartz names him as the creator of the Riddler, one of Batman's recurring villains. However, Finger's contract left him only with his writing page rate and no byline. Kane wrote, "Bill was disheartened by the lack of major accomplishments in his career. He felt that he had not used his creative potential to its fullest and that success had passed him by." At the time of Finger's death in 1974, DC had not officially credited Finger as Batman co-creator.
Jerry Robinson, who also worked with Finger and Kane on the strip at this time, has criticized Kane for failing to share the credit. He recalled Finger resenting his position, stating in a 2005 interview with "The Comics Journal":
Although Kane initially rebutted Finger's claims at having created the character, writing in a 1965 open letter to fans that "it seemed to me that Bill Finger has given out the impression that he and not myself created the "Batman, t' as well as Robin and all the other leading villains and characters. This statement is fraudulent and entirely untrue." Kane himself also commented on Finger's lack of credit. "The trouble with being a 'ghost' writer or artist is that you must remain rather anonymously without 'credit'. However, if one wants the 'credit', then one has to cease being a 'ghost' or follower and become a leader or innovator."
In 1989, Kane revisited Finger's situation, recalling in an interview:
In September 2015, DC Entertainment revealed that Finger would be receiving credit for his role in Batman's creation on the 2016 superhero film "" and the second season of "Gotham" after a deal was worked out between the Finger family and DC. Finger received credit as a creator of Batman for the first time in a comic in October 2015 with "Batman and Robin Eternal" #3 and "Batman: Arkham Knight Genesis" #3. The updated acknowledgment for the character appeared as "Batman created by Bob Kane with Bill Finger".
The first Batman story, "The Case of the Chemical Syndicate", was published in "Detective Comics" #27 (May 1939). Finger said, "Batman was originally written in the style of the pulps", and this influence was evident with Batman showing little remorse over killing or maiming criminals. Batman proved a hit character, and he received his own solo title in 1940 while continuing to star in "Detective Comics". By that time, Detective Comics was the top-selling and most influential publisher in the industry; Batman and the company's other major hero, Superman, were the cornerstones of the company's success. The two characters were featured side-by-side as the stars of "World's Finest Comics", which was originally titled "World's Best Comics" when it debuted in fall 1940. Creators including Jerry Robinson and Dick Sprang also worked on the strips during this period.
Over the course of the first few Batman strips elements were added to the character and the artistic depiction of Batman evolved. Kane noted that within six issues he drew the character's jawline more pronounced, and lengthened the ears on the costume. "About a year later he was almost the full figure, my mature Batman", Kane said. Batman's characteristic utility belt was introduced in "Detective Comics" #29 (July 1939), followed by the boomerang-like batarang and the first bat-themed vehicle, the Batplane, in #31 (Sept. 1939). The character's origin was revealed in #33 (Nov. 1939), unfolding in a two-page story that establishes the brooding persona of Batman, a character driven by the death of his parents. Written by Finger, it depicts a young Bruce Wayne witnessing his parents' murder at the hands of a mugger. Days later, at their grave, the child vows that "by the spirits of my parents [I will] avenge their deaths by spending the rest of my life warring on all criminals".
The early, pulp-inflected portrayal of Batman started to soften in "Detective Comics" #38 (April 1940) with the introduction of Robin, Batman's junior counterpart. Robin was introduced, based on Finger's suggestion, because Batman needed a "Watson" with whom Batman could talk. Sales nearly doubled, despite Kane's preference for a solo Batman, and it sparked a proliferation of "kid sidekicks". The first issue of the solo spin-off series "Batman" was notable not only for introducing two of his most persistent enemies, the Joker and Catwoman, but for a pre-Robin inventory story, originally meant for "Detective Comics #38", in which Batman shoots some monstrous giants to death. That story prompted editor Whitney Ellsworth to decree that the character could no longer kill or use a gun.
By 1942, the writers and artists behind the Batman comics had established most of the basic elements of the Batman mythos. In the years following World War II, DC Comics "adopted a postwar editorial direction that increasingly de-emphasized social commentary in favor of lighthearted juvenile fantasy". The impact of this editorial approach was evident in "Batman" comics of the postwar period; removed from the "bleak and menacing world" of the strips of the early 1940s, Batman was instead portrayed as a respectable citizen and paternal figure that inhabited a "bright and colorful" environment.
Batman was one of the few superhero characters to be continuously published as interest in the genre waned during the 1950s. In the story "The Mightiest Team in the World" in "Superman" #76 (June 1952), Batman teams up with Superman for the first time and the pair discover each other's secret identity. Following the success of this story, "World's Finest Comics" was revamped so it featured stories starring both heroes together, instead of the separate Batman and Superman features that had been running before. The team-up of the characters was "a financial success in an era when those were few and far between"; this series of stories ran until the book's cancellation in 1986.
Batman comics were among those criticized when the comic book industry came under scrutiny with the publication of psychologist Fredric Wertham's book "Seduction of the Innocent" in 1954. Wertham's thesis was that children imitated crimes committed in comic books, and that these works corrupted the morals of the youth. Wertham criticized Batman comics for their supposed homosexual overtones and argued that Batman and Robin were portrayed as lovers. Wertham's criticisms raised a public outcry during the 1950s, eventually leading to the establishment of the Comics Code Authority, a code that is no longer in use by the comic book industry. The tendency towards a "sunnier Batman" in the postwar years intensified after the introduction of the Comics Code. Scholars have suggested that the characters of Batwoman (in 1956) and the pre-Barbara Gordon Bat-Girl (in 1961) were introduced in part to refute the allegation that Batman and Robin were gay, and the stories took on a campier, lighter feel.
In the late 1950s, Batman stories gradually became more science fiction-oriented, an attempt at mimicking the success of other DC characters that had dabbled in the genre. New characters such as Batwoman, the original Bat-Girl, Ace the Bat-Hound, and Bat-Mite were introduced. Batman's adventures often involved odd transformations or bizarre space aliens. In 1960, Batman debuted as a member of the Justice League of America in "The Brave and the Bold" #28 (Feb. 1960), and went on to appear in several Justice League comic book series starting later that same year.
By 1964, sales of Batman titles had fallen drastically. Bob Kane noted that, as a result, DC was "planning to kill Batman off altogether". In response to this, editor Julius Schwartz was assigned to the Batman titles. He presided over drastic changes, beginning with 1964's "Detective Comics" #327 (May 1964), which was cover-billed as the "New Look". Schwartz introduced changes designed to make Batman more contemporary, and to return him to more detective-oriented stories. He brought in artist Carmine Infantino to help overhaul the character. The Batmobile was redesigned, and Batman's costume was modified to incorporate a yellow ellipse behind the bat-insignia. The space aliens, time travel, and characters of the 1950s such as Batwoman, Ace, and Bat-Mite were retired. Bruce Wayne's butler Alfred was killed off (though his death was quickly reversed) while a new female relative for the Wayne family, Aunt Harriet, came to live with Bruce Wayne and Dick Grayson.
The debut of the "Batman" television series in 1966 had a profound influence on the character. The success of the series increased sales throughout the comic book industry, and "Batman" reached a circulation of close to 900,000 copies. Elements such as the character of Batgirl and the show's campy nature were introduced into the comics; the series also initiated the return of Alfred. Although both the comics and TV show were successful for a time, the camp approach eventually wore thin and the show was canceled in 1968. In the aftermath, the Batman comics themselves lost popularity once again. As Julius Schwartz noted, "When the television show was a success, I was asked to be campy, and of course when the show faded, so did the comic books."
Starting in 1969, writer Dennis O'Neil and artist Neal Adams made a deliberate effort to distance Batman from the campy portrayal of the 1960s TV series and to return the character to his roots as a "grim avenger of the night". O'Neil said his idea was "simply to take it back to where it started. I went to the DC library and read some of the early stories. I tried to get a sense of what Kane and Finger were after."
O'Neil and Adams first collaborated on the story "The Secret of the Waiting Graves" in "Detective Comics" #395 (Jan. 1970). Few stories were true collaborations between O'Neil, Adams, Schwartz, and inker Dick Giordano, and in actuality these men were mixed and matched with various other creators during the 1970s; nevertheless the influence of their work was "tremendous". Giordano said: "We went back to a grimmer, darker Batman, and I think that's why these stories did so well ..." While the work of O'Neil and Adams was popular with fans, the acclaim did little to improve declining sales; the same held true with a similarly acclaimed run by writer Steve Englehart and penciler Marshall Rogers in "Detective Comics" #471–476 (Aug. 1977 – April 1978), which went on to influence the 1989 movie "Batman" and be adapted for "", which debuted in 1992. Regardless, circulation continued to drop through the 1970s and 1980s, hitting an all-time low in 1985.
Frank Miller's limited series "The Dark Knight Returns" (February–June 1986) returned the character to his darker roots, both in atmosphere and tone. The comic book, which tells the story of a 55-year-old Batman coming out of retirement in a possible future, reinvigorated interest in the character. "The Dark Knight Returns" was a financial success and has since become one of the medium's most noted touchstones. The series also sparked a major resurgence in the character's popularity.
That year Dennis O'Neil took over as editor of the Batman titles and set the template for the portrayal of Batman following DC's status quo-altering miniseries "Crisis on Infinite Earths". O'Neil operated under the assumption that he was hired to revamp the character and as a result tried to instill a different tone in the books than had gone before. One outcome of this new approach was the "" storyline in "Batman" #404–407 (Feb.–May 1987), in which Frank Miller and artist David Mazzucchelli redefined the character's origins. Writer Alan Moore and artist Brian Bolland continued this dark trend with 1988's 48-page one-shot "", in which the Joker, attempting to drive Commissioner Gordon insane, cripples Gordon's daughter Barbara, and then kidnaps and tortures the commissioner, physically and psychologically.
The Batman comics garnered major attention in 1988 when DC Comics created a 900 number for readers to call to vote on whether Jason Todd, the second Robin, lived or died. Voters decided in favor of Jason's death by a narrow margin of 28 votes (see "").
The 1993 "" story arc introduced a new villain, Bane, who critically injures Batman after pushing him to the limits of his endurance. Jean-Paul Valley, known as Azrael, is called upon to wear the Batsuit during Bruce Wayne's convalescence. Writers Doug Moench, Chuck Dixon, and Alan Grant worked on the Batman titles during "Knightfall", and would also contribute to other Batman crossovers throughout the 1990s. 1998's "" storyline served as the precursor to 1999's "No Man's Land", a year-long storyline that ran through all the Batman-related titles dealing with the effects of an earthquake-ravaged Gotham City. At the conclusion of "No Man's Land", O'Neil stepped down as editor and was replaced by Bob Schreck.
Another writer who rose to prominence on the Batman comic series, was Jeph Loeb. Along with longtime collaborator Tim Sale, they wrote two miniseries ("" and "") that pit an early in his career version of Batman against his entire rogues gallery (including Two-Face, whose origin was re-envisioned by Loeb) while dealing with various mysteries involving serial killers Holiday and the Hangman. In 2003, Loeb teamed with artist Jim Lee to work on another mystery arc: "" for the main Batman book. The 12–issue storyline has Batman and Catwoman teaming up against Batman's entire rogues gallery, including an apparently resurrected Jason Todd, while seeking to find the identity of the mysterious supervillain Hush. While the character of Hush failed to catch on with readers, the arc was a sales success for DC. The series became #1 on the Diamond Comic Distributors sales chart for the first time since "Batman" #500 (Oct. 1993) and Todd's appearance laid the groundwork for writer Judd Winick's subsequent run as writer on "Batman", with another multi-issue arc, "", which ran from "Batman" #637–650 (April 2005 – April 2006).
In 2005, DC launched "All-Star Batman and Robin", a stand-alone comic series set outside the main DC Universe continuity. Written by Frank Miller and drawn by Jim Lee, the series was a commercial success for DC Comics, although it was widely panned by critics for its writing and strong depictions of violence.
Starting in 2006, Grant Morrison and Paul Dini were the regular writers of "Batman" and "Detective Comics", with Morrison reincorporating controversial elements of Batman lore. Most notably of these elements were the science fiction themed storylines of the 1950s Batman comics, which Morrison revised as hallucinations Batman suffered under the influence of various mind-bending gases and extensive sensory deprivation training. Morrison's run climaxed with "Batman R.I.P.", which brought Batman up against the villainous "Black Glove" organization, which sought to drive Batman into madness. "Batman R.I.P." segued into "Final Crisis" (also written by Morrison), which saw the apparent death of Batman at the hands of Darkseid. In the 2009 miniseries "", Wayne's former protégé Dick Grayson becomes the new Batman, and Wayne's son Damian becomes the new Robin. In June 2009, Judd Winick returned to writing "Batman", while Grant Morrison was given his own series, titled "Batman and Robin".
In 2010, the storyline "" saw Bruce travel through history, eventually returning to the present day. Although he reclaimed the mantle of Batman, he also allowed Grayson to continue being Batman as well. Bruce decided to take his crime-fighting cause globally, which is the central focus of "Batman Incorporated". DC Comics would later announce that Grayson would be the main character in "Batman", "Detective Comics", and "Batman and Robin", while Wayne would be the main character in "Batman Incorporated". Also, Bruce appeared in another ongoing series, "".
In September 2011, DC Comics' entire line of superhero comic books, including its "Batman" franchise, were canceled and relaunched with new #1 issues as part of "The New 52" reboot. Bruce Wayne is the only character to be identified as Batman and is featured in "Batman", "Detective Comics", "Batman and Robin", and "". Dick Grayson returns to the mantle of Nightwing and appears in his own ongoing series. While many characters have their histories significantly altered to attract new readers, Batman's history remains mostly intact. "Batman Incorporated" was relaunched in 2012–2013 to complete the "Leviathan" storyline.
With the beginning of "The New 52", Scott Snyder was the writer of the "Batman" title. His first major story arc was "Night of the Owls", where Batman confronts the Court of Owls, a secret society that has controlled Gotham for centuries. The second story arc was "Death of the Family", where the Joker returns to Gotham and simultaneously attacks each member of the Batman family. The third story arc was "", which redefined Batman's origin in The New 52. It followed "Batman" #0, published in June 2012, which explored the character's early years. The final storyline before the "Convergence" (2015) storyline was "", depicting the supposed final battle between Batman and the Joker when he unleashes the deadly Endgame virus onto Gotham City. The storyline ends with Batman and the Joker's supposed deaths.
Starting with "Batman" (vol. 2) #41, Commissioner James Gordon takes over Bruce's mantle as a new, state-sanctioned, robotic-Batman, debuting in the Free Comic Book Day special comic "Divergence". However, Bruce Wayne is soon revealed to be alive, albeit now suffering almost total amnesia of his life as Batman and only remembering his life as Bruce Wayne through what he has learned from Alfred. Bruce Wayne finds happiness and proposes to his girlfriend, Julie Madison, but Mr. Bloom heavily injures Jim Gordon and takes control of Gotham City and threatens to destroy the city by energizing a particle reactor to create a "strange star" to swallow the city. Bruce Wayne discovers the truth that he was Batman and after talking to a stranger who smiles a lot (it is heavily implied that this is the amnesic Joker) he forces Alfred to implant his memories as Batman, but at the cost of his memories as the reborn Bruce Wayne. He returns and helps Jim Gordon defeat Mr. Bloom and shut down the reactor. Gordon gets his job back as the commissioner, and the government Batman project is shut down.
In 2015, DC Comics released "", the sequel to Frank Miller's "The Dark Knight Returns" and "The Dark Knight Strikes Again".
In June 2016, the "DC Rebirth" event relaunched DC Comics' entire line of comic book titles. "Batman" was rebooted as starting with a one-shot issue entitled "Batman: Rebirth" #1 (Aug. 2016). The series then began shipping twice-monthly as a third volume, starting with "Batman" (vol. 3) #1 (Aug. 2016). The third volume of "Batman" was written by Tom King, and artwork was provided by David Finch and Mikel Janín. The "Batman" series introduced two vigilantes, Gotham and Gotham Girl. "Detective Comics" resumed its original numbering system starting with June 2016's #934, and the "New 52" series was labeled as volume 2, with issues numbering from #1-52. Similarly with the "Batman" title, the "New 52" issues were labeled as volume 2 and encompassed issues #1-52. Writer James Tynion IV and artists Eddy Barrows and Alvaro Martinez worked on "Detective Comics" #934, and the series initially featured a team consisting of Tim Drake, Stephanie Brown, Cassandra Cain, and Clayface, led by Batman and Batwoman.
DC Comics ended the "Rebirth" branding in December 2017, opting to include everything under a larger "DC Universe" banner and naming. The continuity established by "Rebirth" continues across DC's comic book titles, including volume one of "Detective Comics" and volume three of "Batman".
After the conclusion of "Batman" (vol. 3) #85 a new creative team consisting of James Tynion IV with art by Tony S. Daniel and Danny Miki replaced Tom King, David Finch and Mikel Janín.
Batman's secret identity is Bruce Wayne, a wealthy American industrialist. As a child, Bruce witnessed the murder of his parents, Dr. Thomas Wayne and Martha Wayne, which ultimately led him to craft the Batman persona and seek justice against criminals. He resides on the outskirts of Gotham City in his personal residence, Wayne Manor. Wayne averts suspicion by acting the part of a superficial playboy idly living off his family's fortune and the profits of Wayne Enterprises, his inherited conglomerate. He supports philanthropic causes through his nonprofit Wayne Foundation, but is more widely known as a celebrity socialite. In public, he frequently appears in the company of high-status women, which encourages tabloid gossip. Although Bruce Wayne leads an active romantic life, his vigilante activities as Batman account for most of his time.
Various modern stories have portrayed the extravagant, playboy image of Bruce Wayne as a facade. This is in contrast to the Post-"Crisis" Superman, whose Clark Kent persona is the true identity, while the Superman persona is the facade. In "Batman Unmasked", a television documentary about the psychology of the character, behavioral scientist Benjamin Karney notes that Batman's personality is driven by Bruce Wayne's inherent humanity; that "Batman, for all its benefits and for all of the time Bruce Wayne devotes to it, is ultimately a tool for Bruce Wayne's efforts to make the world better". Bruce Wayne's principles include the desire to prevent future harm and a vow not to kill. Bruce Wayne believes that our actions define us, we fail for a reason and anything is possible.
Writers of Batman and Superman stories have often compared and contrasted the two. Interpretations vary depending on the writer, the story, and the timing. Grant Morrison notes that both heroes "believe in the same kind of things" despite the day/night contrast their heroic roles display. He notes an equally stark contrast in their real identities. Bruce Wayne and Clark Kent belong to different social classes: "Bruce has a butler, Clark has a boss." T. James Musler's book "Unleashing the Superhero in Us All" explores the extent to which Bruce Wayne's vast personal wealth is important in his life story, and the crucial role it plays in his efforts as Batman.
Will Brooker notes in his book "Batman Unmasked" that "the confirmation of the Batman's identity lies with the young audience ... he doesn't have to be Bruce Wayne; he just needs the suit and gadgets, the abilities, and most importantly the morality, the humanity. There's just a sense about him: 'they trust him ... and they're never wrong."
Batman's primary character traits can be summarized as ""wealth; physical prowess; deductive abilities and obsession"". The details and tone of Batman comic books have varied over the years due to different creative teams. Dennis O'Neil noted that character consistency was not a major concern during early editorial regimes: "Julie Schwartz did a Batman in "Batman" and "Detective" and Murray Boltinoff did a Batman in the "Brave and the Bold" and apart from the costume they bore very little resemblance to each other. Julie and Murray did not want to coordinate their efforts, nor were they asked to do so. Continuity was not important in those days."
The driving force behind Bruce Wayne's character is his parents' murder and their absence. Bob Kane and Bill Finger discussed Batman's background and decided that "there's nothing more traumatic than having your parents murdered before your eyes". Despite his trauma, he sets his mind on studying to become a scientist and to train his body into physical perfection to fight crime in Gotham City as Batman, an inspired idea from Wayne's insight into the criminal mind.
Another of Batman's characterizations is that of a vigilante; in order to stop evil that started with the death of his parents, he must sometimes break the law himself. Although manifested differently by being re-told by different artists, it is nevertheless that the details and the prime components of Batman's origin have never varied at all in the comic books, the "reiteration of the basic origin events holds together otherwise divergent expressions". The origin is the source of the character's traits and attributes, which play out in many of the character's adventures. He also speaks over 40 different languages.
Batman is often treated as a vigilante by other characters in his stories. Frank Miller views the character as "a dionysian figure, a force for anarchy that imposes an individual order". Dressed as a bat, Batman deliberately cultivates a frightening persona in order to aid him in crime-fighting, a fear that originates from the criminals' own guilty conscience. Miller is often credited with reintroducing anti-heroic traits into Batman's characterization, such as his brooding personality, willingness to use violence and torture, and increasingly alienated behavior. Batman, shortly a year after his debut and the introduction of Robin, was changed in 1940 after DC editor Whitney Ellsworth felt the character would be tainted by his lethal methods and DC established their own ethical code, subsequently he was retconned as having a stringent moral code. Miller's Batman was closer to the original pre-Robin version, who was willing to kill criminals if necessary.
On several occasions former Robin Dick Grayson has served as Batman; most notably in 2009 while Wayne was believed dead, and served as a second Batman even after Wayne returned in 2010. As part of DC's 2011 continuity relaunch, Grayson returned to being Nightwing following the "Flashpoint" crossover event.
In an interview with IGN, Morrison detailed that having Dick Grayson as Batman and Damian Wayne as Robin represented a "reverse" of the normal dynamic between Batman and Robin, with, "a more light-hearted and spontaneous Batman and a scowling, badass Robin". Morrison explained his intentions for the new characterization of Batman: "Dick Grayson is kind of this consummate superhero. The guy has been Batman's partner since he was a kid, he's led the Teen Titans, and he's trained with everybody in the DC Universe. So he's a very different kind of Batman. He's a lot easier; a lot looser and more relaxed."
Over the years, there have been numerous others to assume the name of Batman, or to officially take over for Bruce during his leaves of absence. Jean Paul Valley, also known as Azrael, assumed the cowl after the events of the "" saga. James Gordon donned a mech-suit after the events of "", and served as Batman in 2015 and 2016.
Additionally, members of the group Batman, Incorporated, Bruce Wayne's experiment at franchising his brand of vigilantism, have at times stood in as the official Batman in cities around the world. Various others have also taken up the role of Batman in stories set in alternative universes and possible futures, including, among them, various former proteges of Bruce Wayne.
Batman's interactions with both villains and cohorts have, over time, developed a strong supporting cast of characters.
Batman faces a variety of foes ranging from common criminals to outlandish supervillains. Many of them mirror aspects of the Batman's character and development, often having tragic origin stories that lead them to a life of crime. These foes are commonly referred to as Batman's "rogues gallery". Batman's "most implacable foe" is the Joker, a homicidal maniac with a clown-like appearance. The Joker is considered by critics to be his perfect adversary, since he is the antithesis of Batman in personality and appearance; the Joker has a maniacal demeanor with a colorful appearance, while Batman has a serious and resolute demeanor with a dark appearance. As a "personification of the irrational", the Joker represents "everything Batman [opposes]". Other long-time recurring foes that are part of Batman's rogues gallery include Catwoman (a cat burglar anti-heroine who is an occasional ally and romantic interest), the Penguin, Ra's al Ghul, Two-Face, the Riddler, the Scarecrow, Mr. Freeze, Poison Ivy, Harley Quinn, Bane, Clayface, and Killer Croc, among others. Many of Batman's adversaries are often psychiatric patients at Arkham Asylum.
Batman's butler, Alfred Pennyworth, first appeared in "Batman" #16 (1943). He serves as Bruce Wayne's loyal father figure and is one of the few persons to know his secret identity. Alfred raised Bruce after his parents' death and knows him on a very personal level. He is sometimes portrayed as a sidekick to Batman and the only other resident of Wayne Manor aside from Bruce. The character "[lends] a homely touch to Batman's environs and [is] ever ready to provide a steadying and reassuring hand" to the hero and his sidekick.
The informal name "Batman family" is used for a group of characters closely allied with Batman, generally masked vigilantes who either have been trained by Batman or operate in Gotham City with his tacit approval. They include: Barbara Gordon, Commissioner Gordon's daughter, who has fought crime under the vigilante identity of Batgirl and, during a period in which she was confined to a wheelchair due to a gunshot wound inflicted by the Joker, the computer hacker Oracle; Helena Bertinelli, the sole surviving member of a mob family turned vigilante, who has worked with Batman on occasion, primarily as the Huntress and as Batgirl for a brief stint; Cassandra Cain, the daughter of professional assassins David Cain, and Lady Shiva, who succeeded Bertinelli as Batgirl.
Lucius Fox, a technology specialist and Bruce Wayne's business manager who is well aware of his employer's clandestine vigilante activities; Dr. Leslie Thompkins, a family friend who like Alfred became a surrogate parental figure to Bruce Wayne after the deaths of his parents, and is also aware of his secret identity; Vicki Vale, an investigative journalist who often reports on Batman's activities for the Gotham Gazette; Ace the Bat-Hound, Batman's canine partner who was mainly active in the 1950s and 1960s; and Bat-Mite, an extra-dimensional imp mostly active in the 1960s who idolizes Batman.
As Batman's ally in the Gotham City police, Commissioner James "Jim" Gordon debuted along with Batman in "Detective Comics" #27 and has been a consistent presence ever since. As a crime-fighting everyman, he shares Batman's goals while offering, much as the character of Watson does in Sherlock Holmes stories, a normal person's perspective on the work of Batman's extraordinary genius.
Batman is at times a member of superhero teams such as the Justice League of America and the Outsiders. Batman has often been paired in adventures with his Justice League teammate Superman, notably as the co-stars of "World's Finest" and "Superman/Batman" series. In pre-Crisis continuity, the two are depicted as close friends; however, in current continuity, they are still close friends but an uneasy relationship, with an emphasis on their differing views on crime-fighting and justice. In "Superman/Batman" #3 (Dec. 2003), Superman observes, "Sometimes, I admit, I think of Bruce as a man in a costume. Then, with some gadget from his utility belt, he reminds me that he has an extraordinarily inventive mind. And how lucky I am to be able to call on him."
Robin, Batman's vigilante partner, has been a widely recognized supporting character for many years. Bill Finger stated that he wanted to include Robin because "Batman didn't have anyone to talk to, and it got a little tiresome always having him thinking." The first Robin, Dick Grayson, was introduced in 1940. In the 1970s he finally grew up, went off to college and became the hero Nightwing. A second Robin, Jason Todd, appeared in the 1980s. In the stories he was eventually badly beaten and then killed in an explosion set by the Joker, but was later revived. He used the Joker's old persona, the Red Hood, and became an antihero vigilante with no qualms about using firearms or deadly force. Carrie Kelley, the first female Robin to appear in Batman stories, was the final Robin in the continuity of Frank Miller's graphic novels "The Dark Knight Returns" and "The Dark Knight Strikes Again", fighting alongside an aging Batman in stories set out of the mainstream continuity.
The third Robin in the mainstream comics is Tim Drake, who first appeared in 1989. He went on to star in his own comic series, and currently goes by Red Robin, a variation on the traditional Robin persona. In the first decade of the new millennium, Stephanie Brown served as the fourth in-universe Robin between stints as her self-made vigilante identity the Spoiler, and later as Batgirl. After Stephanie Brown's apparent death, Drake resumed the role of Robin for a time. The role eventually passed to Damian Wayne, the 10-year-old son of Bruce Wayne and Talia al Ghul, in the late 2000s. Damian's tenure as "du jour" Robin ended when the character was killed off in the pages of "Batman Incorporated" in 2013. Batman's next young sidekick is Harper Row, a streetwise young woman who avoids the name Robin but followed the ornithological theme nonetheless; she debuted the codename and identity of Bluebird in 2014. Unlike the Robins, Bluebird is willing and permitted to use a gun, albeit non-lethal; her weapon of choice is a modified rifle that fires taser rounds. In 2015, a new series began titled "We Are Robin," focused on a group of teenagers using the Robin persona to fight crime in Gotham City.
Helena Wayne is the biological daughter of Bruce Wayne and Selina Kyle of an alternate universe established in the early 1960s (Multiverse) where the Golden Age stories took place. Damian Wayne is the biological son of Bruce Wayne and Talia al Ghul, and thus the grandson of Ra's al Ghul. Terry McGinnis is the biological son of Bruce Wayne in the DC animated universe, and has taken over the role as Batman when Bruce has become too old to do so.
Writers have varied in the approach over the years to the "playboy" aspect of Bruce Wayne's persona. Some writers show his playboy reputation as a manufactured illusion to support his mission as Batman, while others have depicted Bruce Wayne as genuinely enjoying the benefits of being "Gotham's most eligible bachelor". Bruce Wayne has been portrayed as being romantically linked with many women throughout his various incarnations. The most significant relationships occurred with Selina Kyle, who is also Catwoman and Talia al Ghul, as both women gave birth to his biological offsprings, Helena Wayne and Damian Wayne, respectively.
Batman's first romantic interest was Julie Madison in "Detective Comics" #31 (Sept. 1939); however, their romance was short-lived. Some of Batman's romantic interests have been women with a respected status in society, such as Julie Madison, Vicki Vale, and Silver St. Cloud. Batman has also been romantically involved with allies, such as Kathy Kane (Batwoman), Sasha Bordeaux, and Wonder Woman, and with villains, such as Selina Kyle (Catwoman), Jezebel Jet, Pamela Isley (Poison Ivy), and Talia al Ghul.
While most of Batman's romantic relationships tend to be short in duration, Catwoman has been his most enduring romance throughout the years. The attraction between Batman and Catwoman, whose real name is Selina Kyle, is present in nearly every version and medium in which the characters appear, including a love story between their two secret identities as early as in the 1966 film "Batman". Although Catwoman is typically portrayed as a villain, Batman and Catwoman have worked together in achieving common goals and are usually depicted as having a romantic connection.
In an early 1980s storyline, Selina Kyle and Bruce Wayne develop a relationship, in which the closing panel of the final story shows her referring to Batman as "Bruce". However, a change in the editorial team brought a swift end to that storyline and, apparently, all that transpired during the story arc. Out of costume, Bruce and Selina develop a romantic relationship during "". The story shows Selina saving Bruce from Poison Ivy. However, the relationship ends when Bruce rejects her advances twice; once as Bruce and once as Batman. In "", he stands her up on two holidays, causing her to leave him for good and to leave Gotham City for a while. When the two meet at an opera many years later, during the events of the 12-issue story arc called "", Bruce comments that the two no longer have a relationship as Bruce and Selina. However, "Hush" sees Batman and Catwoman allied against the entire rogues gallery and rekindling their romantic relationship. In "'Hush", Batman reveals his true identity to Catwoman.
The Earth-Two Batman, a character from a parallel world, partners with and marries the reformed Earth-Two Selina Kyle, as shown in "Superman Family" #211. They have a daughter named Helena Wayne, who becomes the Huntress. Along with Dick Grayson, the Earth-Two Robin, the Huntress takes the role as Gotham's protector once Bruce Wayne retires to become police commissioner, a position he occupies until he is killed during one final adventure as Batman.
Batman and Catwoman are shown having a sexual encounter on the roof of a building in "Catwoman" (vol. 4) #1 (2011); the same issue implies that the two have an ongoing sexual relationship. Following the 2016 "DC Rebirth" continuity reboot, the two once again have a sexual encounter on top of a building in "Batman" (vol. 3) #14 (2017).
Following the 2016 "DC Rebirth" continuity reboot, Batman and Catwoman work together in the third volume of "Batman". The two also have a romantic relationship, in which they are shown having a sexual encounter on a rooftop and sleeping together. Bruce proposes to Selina in "Batman" (vol. 3) #24 (2017), and in issue #32, Selina asks Bruce to propose to her again. When he does so, she says, "Yes."
"Batman Annual" (vol. 3) #2 (Jan. 2018) centers on a romantic storyline between Batman and Catwoman. Towards the end, the story is flash-forwarded to the future, in which Bruce Wayne and Selina Kyle are a married couple in their golden years. Bruce receives a terminal medical diagnosis, and Selina cares for him until his death.
Batman has no inherent superhuman powers; he relies on "his own scientific knowledge, detective skills, and athletic prowess". Batman's inexhaustible wealth gives him access to advanced technologies, and as a proficient scientist, he is able to use and modify these technologies to his advantage. In the stories, Batman is regarded as one of the world's greatest detectives, if not the world's greatest crime solver. Batman has been repeatedly described as having a genius-level intellect, being one of the greatest martial artists in the DC Universe, and having peak human physical conditioning. As a polymath, his knowledge and expertise in countless disciplines is nearly unparalleled by any other character in the DC Universe. He has traveled the world acquiring the skills needed to aid him in his endeavors as Batman. In the "" story arc, Superman considers Batman to be one of the most brilliant minds on the planet.
Batman has trained extensively in various different fighting styles, making him one of the best hand-to-hand fighters in the DC Universe. He has fully utilized his photographic memory to master a total of 127 different forms of martial arts. Superman describes Batman as "the most dangerous man on Earth", able to defeat an entire team of superpowered extraterrestrials by himself in order to rescue his imprisoned teammates in Grant Morrison's first storyline in "JLA."
Batman is strongly disciplined, and he has the ability to function under great physical pain and resist most forms of telepathy and mind control. He is a master of disguise, multilingual, and an expert in espionage, often gathering information under the identity of a notorious gangster named Matches Malone. Batman is highly skilled in stealth movement and escapology, which allows him to appear and disappear at will and to break free of nearly inescapable deathtraps with little to no harm.
Batman is an expert in interrogation techniques and his intimidating and frightening appearance alone is often all that is needed in getting information from suspects. Despite having the potential to harm his enemies, Batman's most defining characteristic is his strong commitment to justice and his reluctance to take a life. This unyielding moral rectitude has earned him the respect of several heroes in the DC Universe, most notably that of Superman and Wonder Woman.
Among physical and other crime fighting related training, he is also proficient at other types of skills. Some of these include being a licensed pilot (in order to operate the Batplane), as well as being able to operate other types of machinery. In some publications, he underwent some magician training.
Batman is also frequently described as a "peak human," and it has been revealed that he has (mostly unintentionally) bathed in Lazarus Pits in the past, which may or may not attribute some of his more meta-human feats.
Batman utilizes a vast arsenal of specialized, high-tech vehicles and gadgets in his war against crime, the designs of which usually share a bat motif. Batman historian Les Daniels credits Gardner Fox with creating the concept of Batman's arsenal with the introduction of the utility belt in "Detective Comics" #29 (July 1939) and the first bat-themed weapons the batarang and the "Batgyro" in "Detective Comics" #31 and #32 (Sept., Oct. 1939).
Batman's body armored costume incorporates the imagery of a bat in order to frighten criminals. The details of the Batman costume change repeatedly through various decades, stories, media and artists' interpretations, but the most distinctive elements remain consistent: a scallop-hem cape; a cowl covering most of the face; a pair of bat-like ears; a stylized bat emblem on the chest; and the ever-present utility belt. Finger and Kane originally conceptualized Batman as having a black cape and cowl and grey suit, but conventions in coloring called for black to be highlighted with blue. Hence, the costume's colors have appeared in the comics as dark blue and grey; as well as black and grey. In the Tim Burton's "Batman" and "Batman Returns" films, Batman has been depicted as completely black with a bat in the middle surrounded by a yellow background. Christopher Nolan's "The Dark Knight Trilogy" depicted Batman wearing high-tech gear painted completely black with a black bat in the middle. Ben Affleck's Batman in the DC Extended Universe films wears a suit grey in color with a black cowl, cape, and bat symbol.
Batman's batsuit aids in his combat against enemies, having the properties of both Kevlar and Nomex. It protects him from gunfire and other significant impacts. His gloves typically feature three scallops that protrude from long, gauntlet-like cuffs, although in his earliest appearances he wore short, plain gloves without the scallops. The overall look of the character, particularly the length of the cowl's ears and of the cape, varies greatly depending on the artist. Dennis O'Neil said, "We now say that Batman has two hundred suits hanging in the Batcave so they don't have to look the same ... Everybody loves to draw Batman, and everybody wants to put their own spin on it."
Batman's primary vehicle is the Batmobile, which is usually depicted as an imposing black car, often with tailfins that suggest a bat's wings. Batman also has an aircraft called the Batplane (later called the "Batwing"), along with various other means of transportation. In proper practice, the "bat" prefix (as in Batmobile or batarang) is rarely used by Batman himself when referring to his equipment, particularly after some portrayals (primarily the 1960s "Batman" live-action television show and the "Super Friends" animated series) stretched the practice to campy proportions. For example, the 1960s television show depicted a Batboat, Bat-Sub, and Batcycle, among other bat-themed vehicles. The 1960s television series Batman has an arsenal that includes such "bat-" names as the Bat-computer, Bat-scanner, bat-radar, bat-cuffs, bat-pontoons, bat-drinking water dispenser, bat-camera with polarized bat-filter, bat-shark repellent bat-spray, and Bat-rope. The storyline "A Death in the Family" suggests that given Batman's grim nature, he is unlikely to have adopted the "bat" prefix on his own. In "The Dark Knight Returns", Batman tells Carrie Kelley that the original Robin came up with the name "Batmobile" when he was young, since that is what a kid would call Batman's vehicle. The Batmobile was redesigned in 2011 when DC Comics relaunched its entire line of comic books, with the Batmobile being given heavier armor and new aesthetics.
Batman keeps most of his field equipment in his utility belt. Over the years it has shown to contain an assortment of crime-fighting tools, weapons, and investigative and technological instruments. Different versions of the belt have these items stored in compartments, often as pouches or hard cylinders attached evenly around it. Batman is often depicted as carrying a projectile which shoots a retractable grappling hook attached to a cable. This allows him to attach to distant objects, be propelled into the air, and thus swing from the rooftops of Gotham City. An exception to the range of Batman's equipment are guns, which he refuses to use on principle, since a gun was used in his parents' murder.
When Batman is needed, the Gotham City police activate a searchlight with a bat-shaped insignia over the lens called the Bat-Signal, which shines into the night sky, creating a bat-symbol on a passing cloud which can be seen from any point in Gotham. The origin of the signal varies, depending on the continuity and medium.
In various incarnations, most notably the 1960s "Batman" TV series, Commissioner Gordon also has a dedicated phone line, dubbed the Bat-Phone, connected to a bright red telephone (in the TV series) which sits on a wooden base and has a transparent top. The line connects directly to Batman's residence, Wayne Manor, specifically both to a similar phone sitting on the desk in Bruce Wayne's study and the extension phone in the Batcave.
The Batcave is Batman's secret headquarters, consisting of a series of caves beneath his mansion, Wayne Manor. As his command center, the Batcave serves multiple purposes; supercomputer, surveillance, redundant power-generators, forensics lab, medical infirmary, private study, training dojo, fabrication workshop, arsenal, hangar and garage. It houses the vehicles and equipment Batman uses in his campaign to fight crime. It is also a trophy room and storage facility for Batman's unique memorabilia collected over the years from various cases he has worked on. In both the comic "" #45 and the 2005 film "Batman Begins", the cave is said to have been part of the Underground Railroad.
Batman's history has undergone many retroactive continuity revisions, both minor and major. Elements of the character's history have varied greatly. Scholars William Uricchio and Roberta E. Pearson noted in the early 1990s, "Unlike some fictional characters, the Batman has no primary urtext set in a specific period, but has rather existed in a plethora of equally valid texts constantly appearing over more than five decades."
The central fixed event in the Batman stories is the character's origin story. As a young boy, Bruce Wayne was horrified and traumatized when he watched his parents, the physician Dr. Thomas Wayne and his wife Martha, murdered with a gun by a mugger named Joe Chill. Batman refuses to utilize any sort of gun on the principle that a gun was used to murder his parents. This event drove him to train his body to its peak condition and fight crime in Gotham City as Batman. Pearson and Uricchio also noted beyond the origin story and such events as the introduction of Robin, "Until recently, the fixed and accruing and hence, canonized, events have been few in number", a situation altered by an increased effort by later Batman editors such as Dennis O'Neil to ensure consistency and continuity between stories.
In Batman's first appearance in "Detective Comics" #27, he is already operating as a crime-fighter. Batman's origin is first presented in "Detective Comics" #33 (Nov. 1939) and is later expanded upon in "Batman" #47. As these comics state, Bruce Wayne is born to Dr. Thomas Wayne and his wife Martha, two very wealthy and charitable Gotham City socialites. Bruce is brought up in Wayne Manor, and leads a happy and privileged existence until the age of eight, when his parents are killed by a small-time criminal named Joe Chill while on their way home from a movie theater. That night, Bruce Wayne swears an oath to spend his life fighting crime. He engages in intense intellectual and physical training; however, he realizes that these skills alone would not be enough. "Criminals are a superstitious cowardly lot", Wayne remarks, "so my disguise must be able to strike terror into their hearts. I must be a creature of the night, black, terrible ..." As if responding to his desires, a bat suddenly flies through the window, inspiring Bruce to craft the Batman persona.
In early strips, Batman's career as a vigilante earns him the ire of the police. During this period, Bruce Wayne has a fiancée named Julie Madison. In "Detective Comics" #38, Wayne takes in an orphaned circus acrobat, Dick Grayson, who becomes his vigilante partner, Robin. Batman also becomes a founding member of the Justice Society of America, although he, like Superman, is an honorary member, and thus only participates occasionally. Batman's relationship with the law thaws quickly, and he is made an honorary member of Gotham City's police department. During this time, Alfred Pennyworth arrives at Wayne Manor, and after deducing the Dynamic Duo's secret identities, joins their service as their butler.
The Silver Age of Comic Books in DC Comics is sometimes held to have begun in 1956 when the publisher introduced Barry Allen as a new, updated version of THE Flash. Batman is not significantly changed by the late 1950s for the continuity which would be later referred to as Earth-One. The lighter tone Batman had taken in the period between the Golden and Silver Ages led to the stories of the late 1950s and early 1960s that often feature many science-fiction elements, and Batman is not significantly updated in the manner of other characters until "Detective Comics" #327 (May 1964), in which Batman reverts to his detective roots, with most science-fiction elements jettisoned from the series.
After the introduction of DC Comics' multiverse in the 1960s, DC established that stories from the Golden Age star the Earth-Two Batman, a character from a parallel world. This version of Batman partners with and marries the reformed Earth-Two Catwoman (Selina Kyle). The two have a daughter, Helena Wayne, who becomes the Huntress. She assumes the position as Gotham's protector along with Dick Grayson, the Earth-Two Robin, once Bruce Wayne retires to become police commissioner. Wayne holds the position of police commissioner until he is killed during one final adventure as Batman. Batman titles, however, often ignored that a distinction had been made between the pre-revamp and post-revamp Batmen (since unlike the Flash or Green Lantern, Batman comics had been published without interruption through the 1950s) and would occasionally make reference to stories from the Golden Age. Nevertheless, details of Batman's history were altered or expanded upon through the decades. Additions include meetings with a future Superman during his youth, his upbringing by his uncle Philip Wayne (introduced in "Batman" #208 (Feb. 1969)) after his parents' death, and appearances of his father and himself as prototypical versions of Batman and Robin, respectively. In 1980, then-editor Paul Levitz commissioned the "Untold Legend of the Batman" miniseries to thoroughly chronicle Batman's origin and history.
Batman meets and regularly works with other heroes during the Silver Age, most notably Superman, whom he began regularly working alongside in a series of team-ups in "World's Finest Comics", starting in 1954 and continuing through the series' cancellation in 1986. Batman and Superman are usually depicted as close friends. As a founding member of the Justice League of America, Batman appears in its first story, in 1960's "The Brave and the Bold" #28. In the 1970s and 1980s, "The Brave and the Bold" became a Batman title, in which Batman teams up with a different DC Universe superhero each month.
In 1969, Dick Grayson attends college as part of DC Comics' effort to revise the Batman comics. Additionally, Batman also moves from his mansion, Wayne Manor into a penthouse apartment atop the Wayne Foundation building in downtown Gotham City, in order to be closer to Gotham City's crime. Batman spends the 1970s and early 1980s mainly working solo, with occasional team-ups with Robin and/or Batgirl. Batman's adventures also become somewhat darker and more grim during this period, depicting increasingly violent crimes, including the first appearance (since the early golden age) of the Joker as a homicidal psychopath, and the arrival of Ra's al Ghul, a centuries-old terrorist who knows Batman's secret identity. In the 1980s, Dick Grayson becomes Nightwing.
In the final issue of "The Brave and the Bold" in 1983, Batman quits the Justice League and forms a new group called the Outsiders. He serves as the team's leader until "Batman and the Outsiders" #32 (1986) and the comic subsequently changed its title.
After the 12-issue miniseries "Crisis on Infinite Earths", DC Comics retconned the histories of some major characters in an attempt at updating them for contemporary audiences. Frank Miller retold Batman's origin in the storyline "" from "Batman" #404–407, which emphasizes a grittier tone in the character. Though the Earth-Two Batman is erased from history, many stories of Batman's silver-age/Earth-One career (along with an amount of golden-age ones) remain canonical in the post-Crisis universe, with his origins remaining the same in essence, despite alteration. For example, Gotham's police are mostly corrupt, setting up further need for Batman's existence. The guardian Phillip Wayne is removed, leaving young Bruce to be raised by Alfred Pennyworth. Additionally, Batman is no longer a founding member of the Justice League of America, although he becomes leader for a short time of a new incarnation of the team launched in 1987. To help fill in the revised backstory for Batman following "Crisis", DC launched a new Batman title called "Legends of the Dark Knight" in 1989 and has published various miniseries and one-shot stories since then that largely take place during the "Year One" period.
Subsequently, Batman begins exhibiting an excessive, reckless approach to his crimefighting, a result of the pain of losing Jason Todd. Batman works solo until the decade's close, when Tim Drake becomes the new Robin.
Many of the major Batman storylines since the 1990s have been intertitle crossovers that run for a number of issues. In 1993, DC published "". During the storyline's first phase, the new villain Bane paralyzes Batman, leading Wayne to ask Azrael to take on the role. After the end of "Knightfall", the storylines split in two directions, following both the Azrael-Batman's adventures, and Bruce Wayne's quest to become Batman once more. The story arcs realign in "KnightsEnd", as Azrael becomes increasingly violent and is defeated by a healed Bruce Wayne. Wayne hands the Batman mantle to Dick Grayson (then Nightwing) for an interim period, while Wayne trains for a return to the role.
The 1994 company-wide crossover storyline "Zero Hour: Crisis in Time" changes aspects of DC continuity again, including those of Batman. Noteworthy among these changes is that the general populace and the criminal element now considers Batman an urban legend rather than a known force.
Batman once again becomes a member of the Justice League during Grant Morrison's 1996 relaunch of the series, titled "JLA". During this time, Gotham City faces catastrophe in the decade's closing crossover arc. In 1998's "" storyline, Gotham City is devastated by an earthquake and ultimately cut off from the United States. Deprived of many of his technological resources, Batman fights to reclaim the city from legions of gangs during 1999's "No Man's Land".
Meanwhile, Batman's relationship with the Gotham City Police Department changed for the worse with the events of "Batman: Officer Down" and "Batman: War Games/War Crimes"; Batman's long-time law enforcement allies Commissioner Gordon and Harvey Bullock are forced out of the police department in "Officer Down", while "War Games" and "War Crimes" saw Batman become a wanted fugitive after a contingency plan of his to neutralize Gotham City's criminal underworld is accidentally triggered, resulting in a massive gang war that ends with the sadistic Black Mask the undisputed ruler of the city's criminal gangs. Lex Luthor arranges for the murder of Batman's on-again, off-again love interest Vesper Lynd (introduced in the mid-1990s) during the "Bruce Wayne: Murderer?" and "" story arcs. Though Batman is able to clear his name, he loses another ally in the form of his new bodyguard Sasha, who is recruited into the organization known as "Checkmate" while stuck in prison due to her refusal to turn state's evidence against her employer. While he was unable to prove that Luthor was behind the murder of Vesper, Batman does get his revenge with help from Talia al Ghul in "Superman/Batman" #1–6.
DC Comics' 2005 limited series "Identity Crisis" reveals that JLA member Zatanna had edited Batman's memories to prevent him from stopping the Justice League from lobotomizing Dr. Light after he raped Sue Dibny. Batman later creates the Brother I satellite surveillance system to watch over and, if necessary, kill the other heroes after he remembered. The revelation of Batman's creation and his tacit responsibility for the Blue Beetle's death becomes a driving force in the lead-up to the "Infinite Crisis" miniseries, which again restructures DC continuity. Batman and a team of superheroes destroy Brother Eye and the OMACs, though, at the very end, Batman reaches his apparent breaking point when Alexander Luthor Jr. seriously wounds Nightwing. Picking up a gun, Batman nearly shoots Luthor in order to avenge his former sidekick, until Wonder Woman convinces him to not pull the trigger.
Following "Infinite Crisis", Bruce Wayne, Dick Grayson (having recovered from his wounds), and Tim Drake retrace the steps Bruce had taken when he originally left Gotham City, to "rebuild Batman". In the "" storyline, Batman and Robin return to Gotham City after their year-long absence. Part of this absence is captured during Week 30 of the "52" series, which shows Batman fighting his inner demons. Later on in "52", Batman is shown undergoing an intense meditation ritual in Nanda Parbat. This becomes an important part of the regular "Batman" title, which reveals that Batman is reborn as a more effective crime fighter while undergoing this ritual, having "hunted down and ate" the last traces of fear in his mind. At the end of the "Face the Face" story arc, Bruce officially adopts Tim (who had lost both of his parents at various points in the character's history) as his son. The follow-up story arc in "Batman", "Batman and Son", introduces Damian Wayne, who is Batman's son with Talia al Ghul. Although originally, in "Son of the Demon", Bruce's coupling with Talia was implied to be consensual, this arc retconned it into Talia forcing herself on Bruce.
Batman, along with Superman and Wonder Woman, reforms the Justice League in the new "Justice League of America" series, and is leading the newest incarnation of the Outsiders.
Grant Morrison's 2008 storyline, "Batman R.I.P." featured Batman being physically and mentally broken by the enigmatic villain Doctor Hurt and attracted news coverage in advance of its highly promoted conclusion, which would speculated to feature the death of Bruce Wayne. However, though Batman is shown to possibly perish at the end of the arc, the two-issue arc "Last Rites", which leads into the crossover storyline "Final Crisis", shows that Batman survives his helicopter crash into the Gotham City River and returns to the Batcave, only to be summoned to the Hall of Justice by the JLA to help investigate the New God Orion's death. The story ends with Batman retrieving the god-killing bullet used to kill Orion, setting up its use in "Final Crisis". In the pages of "Final Crisis" Batman is reduced to a charred skeleton. In "Final Crisis" #7, Wayne is shown witnessing the passing of the first man, Anthro. Wayne's "death" sets up the three-issue "" miniseries in which Wayne's ex-proteges compete for the "right" to assume the role of Batman, which concludes with Grayson becoming Batman, while Tim Drake takes on the identity of the Red Robin. Dick and Damian continue as Batman and Robin, and in the crossover storyline "Blackest Night", what appears to be Bruce's corpse is reanimated as a Black Lantern zombie, but is later shown that Bruce's corpse is one of Darkseid's failed Batman clones. Dick and Batman's other friends conclude that Bruce is alive.
Bruce subsequently returned in Morrison's miniseries "", which depicted his travels through time from prehistory to present-day Gotham. Bruce's return set up "Batman Incorporated", an ongoing series which focused on Wayne franchising the Batman identity across the globe, allowing Dick and Damian to continue as Gotham's Dynamic Duo. Bruce publicly announced that Wayne Enterprises will aid Batman on his mission, known as "Batman, Incorporated". However, due to rebooted continuity that occurred as part of DC Comics' 2011 relaunch of all of its comic books, "The New 52", Dick Grayson was restored as Nightwing with Wayne serving as the sole Batman once again. The relaunch also interrupted the publication of "Batman, Incorporated", which resumed its story in 2012–2013 with changes to suit the new status quo.
Batman has become a pop culture icon, recognized around the world. The character's presence has extended beyond his comic book origins; events such as the release of the 1989 "Batman" film and its accompanying merchandising "brought the Batman to the forefront of public consciousness". In an article commemorating the sixtieth anniversary of the character, "The Guardian" wrote, "Batman is a figure blurred by the endless reinvention that is modern mass culture. He is at once an icon and a commodity: the perfect cultural artefact for the 21st century."
The character of Batman has appeared in various media aside from comic books, such as newspaper syndicated comic strips, books, radio dramas, television, a stage show, and several theatrical feature films. The first adaptation of Batman was as a daily newspaper comic strip which premiered on October 25, 1943. That same year the character was adapted in the 15-part serial "Batman", with Lewis Wilson becoming the first actor to portray Batman on screen. While Batman never had a radio series of his own, the character made occasional guest appearances in "The Adventures of Superman" starting in 1945 on occasions when Superman voice actor Bud Collyer needed time off. A second movie serial, "Batman and Robin", followed in 1949, with Robert Lowery taking over the role of Batman. The exposure provided by these adaptations during the 1940s "helped make [Batman] a household name for millions who never bought a comic book".
In the 1964 publication of Donald Barthelme's collection of short stories "Come Back, Dr. Caligari", Barthelme wrote "The Joker's Greatest Triumph". Batman is portrayed for purposes of spoof as a pretentious French-speaking rich man.
The "Batman" television series, starring Adam West, premiered in January 1966 on the ABC television network. Inflected with a camp sense of humor, the show became a pop culture phenomenon. In his memoir, "Back to the Batcave", West notes his dislike for the term 'camp' as it was applied to the 1960s series, opining that the show was instead a farce or lampoon, and a deliberate one, at that. The series ran for 120 episodes; ending in 1968. In between the first and second season of the "Batman" television series, the cast and crew made the theatrical film "Batman" (1966). The Who recorded the theme song from the "Batman" show for their 1966 EP Ready Steady Who, and The Kinks performed the theme song on their 1967 album Live at Kelvin Hall.
The popularity of the "Batman" TV series also resulted in the first animated adaptation of Batman in "The Batman/Superman Hour"; the Batman segments of the series were repackaged as "The Adventures of Batman" and "Batman with Robin the Boy Wonder" which produced thirty-three episodes between 1968 and 1977. From 1973 until 1986, Batman had a starring role in ABC's "Super Friends" series; which was animated by Hanna-Barbera. Olan Soule was the voice of Batman in all these shows, but was eventually replaced during "Super Friends" by Adam West, who also voiced the character in Filmation's 1977 series "The New Adventures of Batman".
In 1992, "" premiered on the Fox television network; produced by Warner Bros. Animation and featuring Kevin Conroy as the voice of Batman. The series received considerable acclaim for its darker tone, mature writing, stylistic design, and thematic complexity compared to previous superhero cartoons, in addition to multiple Emmy Awards. The series' success led to the theatrical film "" (1993), as well as various spin-off TV series; including "", "The New Batman Adventures", "Justice League" and "Justice League Unlimited" (each of which also featured Conroy as Batman's voice). The futuristic series "Batman Beyond" also took place in this same animated continuity and featured a newer, younger Batman voiced by Will Friedle, with the elderly Bruce Wayne (again voiced by Conroy) as a mentor.
In 2004, an unrelated animated series titled "The Batman" made its debut with Rino Romano voicing Batman. In 2008, this show was replaced by another animated series, "", featuring Diedrich Bader's voice as Batman. In 2013, a new CGI-animated series titled "Beware the Batman" made its debut, with Anthony Ruivivar voicing Batman.
In 2014, the live-action TV series "Gotham" premiered on the Fox network, featuring David Mazouz as a 12-year-old Bruce Wayne. In 2018, when the series was renewed for its fifth and final season it was announced that Batman would make an appearance in the show's series finale's flash-forward.
Iain Glen portrays Bruce Wayne in the live-action series "Titans", appearing in the show's second season in 2019. Prior to Glen, Batman was played by stunt doubles Alain Moussi and Maxim Savarias in the first season.
To commemorate the 75th anniversary of the character, Warner Bros aired the television short film, "", that was also posted on DC's YouTube channel.
In August 2019, it was announced that Kevin Conroy would make his live-action television debut as an older Bruce Wayne in the upcoming Arrowverse crossover, "Crisis on Infinite Earths". In the crossover, he portrayed a parallel universe iteration of Batman from Earth-99. In "Batwoman", the Earth-Prime version of Bruce Wayne / Batman is portrayed by Warren Christie.
In 1989, twenty-three years after the aforementioned film based on the TV series starring Adam West, Warner Bros. released the live-action feature film "Batman"; directed by Tim Burton and starring Michael Keaton as the title character. The film was a huge success; not only was it the top-grossing film of the year, but at the time was the fifth highest-grossing film in history. The film also won the Academy Award for Best Art Direction. The film's success spawned three sequels: "Batman Returns" (1992), "Batman Forever" (1995) and "Batman & Robin" (1997); the latter two of which were directed by Joel Schumacher instead of Burton, and replaced Keaton as Batman with Val Kilmer and George Clooney, respectively. The second Schumacher film failed to outgross any of its predecessors and was critically panned; causing Warner Bros. to cancel the planned fourth sequel, "Batman Unchained", and end the initial film series.
In 2005, "Batman Begins" was released by Warner Bros. as a reboot of the film series; directed by Christopher Nolan and starring Christian Bale as Batman. Its sequel, "The Dark Knight" (2008), set the record for the highest grossing opening weekend of all time in the U.S., earning approximately $158 million, and became the fastest film to reach the $400 million mark in the history of American cinema (eighteenth day of release). These record-breaking attendances saw "The Dark Knight" end its run as the second-highest domestic grossing film (at the time) with $533 million, bested then only by "Titanic". The film also won two Academy Awards, including Best Supporting Actor for the late Heath Ledger. It was eventually followed by "The Dark Knight Rises" (2012), which served as a conclusion to Nolan's film series.
Since 2008, Batman has also starred in various direct-to-video animated films under the DC Universe Animated Original Movies banner. Kevin Conroy has reprised his voice role of Batman for several of these films, while others have featured celebrity voice actors in the role; including Jeremy Sisto, William Baldwin, Bruce Greenwood, Ben McKenzie, and Peter Weller. A Lego-themed version of Batman was also featured as one of the protagonists in the animated film "The Lego Movie" (2014), with Will Arnett providing the voice. Arnett reprised the voice role for the spin-off film "The Lego Batman Movie" (2017).
In 2016, Ben Affleck began portraying Batman in the DC Extended Universe with the release of the film "", directed by Zack Snyder, a younger child version of the character was played by Brandon Spink in the same film. Affleck also made a cameo appearance as Batman in David Ayer's film "Suicide Squad" (2016). Affleck reprised the role in the 2017 film "Justice League", also set in the DC Extended Universe.
Dante Pereira-Olson portrays a young Bruce Wayne in the 2019 film "Joker".
In 2019 Robert Pattinson was cast to play Batman in an upcoming reboot of the franchise, "The Batman", releasing in 2021.
Since 1986, Batman has starred in multiple video games, most of which were adaptations of the various cinematic or animated incarnations of the character. Among the most successful of these games is the "" series. The first installment, "" (2009), was released by Rocksteady Studios to critical acclaim; review aggregator Metacritic reports it as having received 92% positive reviews. It was followed by the sequel "" (2011), which also received widespread acclaim and holds a Metacritic ranking of 94%. A prequel game titled "" (2013) was later released by WB Games Montréal. A fourth game titled "" (2015) has also been released by Rocksteady. As with most animated "Batman" productions, Kevin Conroy has provided the voice of the character for these games; excluding "Arkham Origins", in which the younger Batman is voiced by Roger Craig Smith. In 2016, Telltale Games released "" adventure game, which changed the Wayne Family's history as it is depicted in the Batman mythos. A sequel, titled "", was released in 2017.
Mayfair Games published the "DC Heroes" role-playing game in 1985, then published the 80-page supplement "Batman" the following year, written by Mike Stackpole, with cover art by Ed Hannigan. In 1989, Mayfair Games published an updated 96-page softcover "Batman Sourcebook", again written by Mike Stackpole, with additional material by J. Santana, Louis Prosperi, Jack Barker and Ray Winninger, with graphic design by Gregory Scott, and cover and interior art by DC Comics staff.
Mayfair released a simplified version of "DC Heroes" called "The Batman Role-Playing Game" in 1989 to coincide with the "Batman" film.
Gay interpretations of the character have been part of the academic study of Batman since psychologist Fredric Wertham asserted in "Seduction of the Innocent" in 1954 that "Batman stories are psychologically homosexual ... The Batman type of story may stimulate children to homosexual fantasies, of the nature of which they may be unconscious." Andy Medhurst wrote in his 1991 essay "Batman, Deviance, and Camp" that Batman is interesting to gay audiences because "he was one of the first fictional characters to be attacked on the grounds of his presumed homosexuality". Professor of film and cultural studies Will Brooker argues the validity of a queer reading of Batman, and that gay readers would naturally find themselves drawn to the lifestyle depicted within, whether the character of Bruce Wayne himself were explicitly gay or not. He also identifies a homophobic element to the vigor with which mainstream fandom rejects the possibility of a gay reading of the character. In 2005, painter Mark Chamberlain displayed a number of watercolors depicting both Batman and Robin in suggestive and sexually explicit poses, prompting DC to threaten legal action.
Creators associated with the character have expressed their own opinions. Writer Alan Grant has stated, "The Batman I wrote for 13 years isn't gay ... everybody's Batman all the way back to Bob Kane ... none of them wrote him as a gay character. Only Joel Schumacher might have had an opposing view." Frank Miller views the character as sublimating his sexual urges into crimefighting so much so that he's "borderline pathological", concluding "He'd be much healthier if he "were" gay." Grant Morrison said that "Gayness is built into Batman ... Obviously as a fictional character he's intended to be heterosexual, but the basis of the whole concept is utterly gay."
Batman has been the subject of psychological study for some time, and there have been a number of interpretations into the character's psyche.
In "Batman and Psychology: A Dark and Stormy Knight", Travis Langley argues that the concept of archetypes as described by psychologists Carl Jung and Joseph Campbell is present in the Batman mythos, such that the character represents the "shadow archetype". This archetype, according to Langley, represents a person's own dark side; it is not necessarily an evil one, but rather one that is hidden from the outside and concealed from both the world and oneself. Langley argues that Bruce Wayne confronts his own darkness early in life; he chooses to use it to instill fear in wrongdoers, with his bright and dark sides working together to fight evil. Langley uses the Jungian perspective to assert that Batman appeals to our own need to face our "shadow selves". Dr. Travis Langley also taught a class called Batman, a title he was adamant about. "I could have called it something like the Psychology of Nocturnal Vigilantism, but no. I called it Batman," Langley says.
Several psychologists have explored Bruce Wayne/Batman's mental health. Robin. S. Rosenberg evaluated his actions and problems to determine if they reach the level of mental disorders. She examined the possibility of several mental health issues, including dissociative identity disorder, obsessive–compulsive disorder, and several others. She concluded that Bruce Wayne/Batman may have a disorder or a combination of disorders but due to his fictional nature, a definitive diagnosis will remain unknown. | https://en.wikipedia.org/wiki?curid=4335 |
Euclidean space
Euclidean space is the fundamental space of classical geometry. Originally it was the three-dimensional space of Euclidean geometry, but in modern mathematics there are Euclidean spaces of any nonnegative integer dimension, including the three-dimensional space and the "Euclidean plane" (dimension two). It was introduced by the Ancient Greek mathematician Euclid of Alexandria, and the qualifier "Euclidean" is used to distinguish it from other spaces that were later discovered in physics and modern mathematics.
Ancient Greek geometers introduced Euclidean space for modeling the physical universe. Their great innovation was to "prove" all properties of the space as theorems by starting from a few fundamental properties, called "postulates", which either were considered as evident (for example, there is exactly one straight line passing through two points), or seemed impossible to prove (parallel postulate).
After the introduction at the end of 19th century of non-Euclidean geometries, the old postulates were re-formalized to define Euclidean spaces through axiomatic theory. Another definition of Euclidean spaces by means of vector spaces and linear algebra has been shown to be equivalent to the axiomatic definition. It is this definition that is more commonly used in modern mathematics, and detailed in this article.
In all definitions, Euclidean spaces consist of points, which are defined only by the properties that they must have for forming a Euclidean space.
There is essentially only one Euclidean space of each dimension; that is, all Euclidean spaces of a given dimension are isomorphic. Therefore, in many cases, it is possible to work with a specific Euclidean space, which is generally the real -space formula_1 equipped with the dot product. An isomorphism from a Euclidean space to formula_2 associates with each point an -tuple of real numbers which locate that point in the Euclidean space and are called the "Cartesian coordinates" of that point.
Euclidean space was introduced by ancient Greeks as an abstraction of our physical space. Their great innovation, appearing in Euclid's "Elements" was to build and "prove" all geometry by starting from a few very basic properties, which are abstracted from the physical world, and cannot be mathematically proved because of the lack of more basic tools. These properties are called postulates, or axioms in modern language. This way of defining Euclidean space is still in use under the name of synthetic geometry.
In 1637, René Descartes introduced Cartesian coordinates and showed that this allows reducing geometric problems to algebraic computations with numbers. This reduction of geometry to algebra was a major change of point of view, as, until then, the real numbers—that is, rational numbers and non-rational numbers together–were defined in terms of geometry, as lengths and distance.
Euclidean geometry was not applied in spaces of more than three dimensions until the 19th century. Ludwig Schläfli generalized Euclidean geometry to spaces of "n" dimensions using both synthetic and algebraic methods, and discovered all of the regular polytopes (higher-dimensional analogues of the Platonic solids) that exist in Euclidean spaces of any number of dimensions.
Despite the wide use of Descartes' approach, which was called analytic geometry, the definition of Euclidean space remained unchanged until the end of 19th century. The introduction of abstract vector spaces allowed their use in defining Euclidean spaces with a purely algebraic definition. This new definition has been shown to be equivalent to the classical definition in terms of geometric axioms. It is this algebraic definition that is now most often used for introducing Euclidean spaces.
One way to think of the Euclidean plane is as a set of points satisfying certain relationships, expressible in terms of distance and angles. For example, there are two fundamental operations (referred to as motions) on the plane. One is translation, which means a shifting of the plane so that every point is shifted in the same direction and by the same distance. The other is rotation around a fixed point in the plane, in which all points in the plane turn around that fixed point through the same angle. One of the basic tenets of Euclidean geometry is that two figures (usually considered as subsets) of the plane should be considered equivalent (congruent) if one can be transformed into the other by some sequence of translations, rotations and reflections (see below).
In order to make all of this mathematically precise, the theory must clearly define what is a Euclidean space, and the related notions of distance, angle, translation, and rotation. Even when used in physical theories, Euclidean space is an abstraction detached from actual physical locations, specific reference frames, measurement instruments, and so on. A purely mathematical definition of Euclidean space also ignores questions of units of length and other physical dimensions: the distance in a "mathematical" space is a number, not something expressed in inches or metres.
The standard way to mathematically define a Euclidean space, as carried out in the remainder of this article, is to define a Euclidean space as a set of points on which acts a real vector space, the "space of translations" which is equipped with an inner product. The action of translations makes the space an affine space, and this allows defining lines, planes, subspaces, dimension, and parallelism. The inner product allows defining distance and angles.
The set formula_2 of -tuples of real numbers equipped with the dot product is a Euclidean space of dimension . Conversely, the choice of a point called the "origin" and an orthonormal basis of the space of translations is equivalent with defining an isomorphism between a Euclidean space of dimension and formula_2 viewed as a Euclidean space.
It follows that everything that can be said about a Euclidean space can also be said about formula_5 Therefore, many authors, specially at elementary level, call formula_2 the "standard Euclidean space" of dimension , or simply "the" Euclidean space of dimension .
A reason for introducing such an abstract definition of Euclidean spaces, and for working with it instead of formula_2 is that it is often preferable to work in a "coordinate-free" and "origin-free" manner (that is, without choosing a preferred basis and a preferred origin). Another reason is that there is no origin nor any basis in the physical world.
A is a finite-dimensional inner product space over the real numbers.
A Euclidean space is an affine space over the reals such that the associated vector space is a Euclidean vector space. Euclidean spaces are sometimes called "Euclidean affine spaces" for distinguishing them from Euclidean vector spaces.
The "dimension" of a Euclidean space is the dimension of its associated vector space.
If is a Euclidean space, its associated vector space is also called its space of translations, and often denoted formula_8
The elements of are called "points" and are commonly denoted by capital letters. The elements of formula_9 are called "translations", "Euclidean vectors" or "free vectors".
The action of a translation on a point provides a point that is denoted . This action satisfies
The fact that the action is free and transitive means that for every pair of points there is exactly one vector such that . This vector is denoted or formula_11
As previously explained, some of the basic properties of Euclidean spaces result of the structure of affine space. They are described in and its subsections. The properties resulting from the inner product are explained in and its subsections.
For any vector space, the addition acts freely and transitively on the vector space itself. Thus a Euclidean vector space can be viewed as a Euclidean space that has itself as associated vector space.
A typical case of Euclidean vector space is formula_2 viewed as a vector space equipped with the dot product as an inner product. The importance of this particular example of Euclidean space lies in the fact that every Euclidean space is isomorphic to it. More precisely, given a Euclidean space of dimension , the choice of a point, called a "origin" and an orthonormal basis of formula_9 defines an isomorphism of Euclidean spaces from to formula_5
As every Euclidean space of dimension is isomorphic to it, the Euclidean space formula_2 is sometimes called the "standard Euclidean space" of dimension .
Some basic properties of Euclidean spaces depend only of the fact that a Euclidean space is an affine space. They are called affine properties and include the concepts of lines, subspaces, and parallelism. which are detailed in next subsections.
Let be a Euclidean space and formula_9 its associated vector space.
A "flat", "Euclidean subspace" or "affine subspace" of is a subset of such that
is a linear subspace of formula_8 A Euclidean subspace is a Euclidean space with formula_19 as associated vector space. This linear subspace formula_19 is called the "direction" of .
If is a point of then
Conversely, if is a point of and is a linear subspace of formula_22 then
is a Euclidean subspace of direction .
A Euclidean vector space (that is, a Euclidean space such that formula_24) has two sorts of subspaces: its Euclidean subspaces and its linear subspaces. Linear subspaces are Euclidean subspaces and a Euclidean subspace is a linear subspace if and only if it contains the zero vector.
In a Euclidean space, a "line" is a Euclidean subspace of dimension one. Since a vector space of dimension one is spanned by any nonzero vector a line is a set of the form
where and are two distinct points.
It follows that "there is exactly one line that passes through (contains) two distinct points." This implies that two distinct lines intersect in at most one point.
A more symmetric representation of the line passing through and is
where is an arbitrary point (not necessary on the line).
In a Euclidean vector space, the zero vector is usually chosen for ; this allows simplifying the preceding formula into
A standard convention allows using this formula in every Euclidean space, see .
The "line segment", or simply "segment", joining the points and is the subset of the points such that in the preceding formulas. It is denoted or ; that is
Two subspaces and of the same dimension in a Euclidean space are "parallel" if they have the same direction. Equivalently, they are parallel, if there is a translation vector that maps one to the other:
Given a point and a subspace , there exists exactly one subspace that contains and is parallel to , which is formula_30 In the case where is a line (subspace of dimension one), this property is Playfair's axiom.
It follows that in a Euclidean plane, two lines either meet in one point or are parallel.
The concept of parallel subspaces has been extended to subspaces of different dimensions: two subspaces are parallel if the direction of one of them is contained in the direction to the other.
The vector space formula_9 associated to a Euclidean space is an inner product space. This implies a symmetric bilinear form
that is positive definite (that is formula_33 is always positive for ).
The inner product of a Euclidean space is often called "dot product" and denoted . This is specially the case when a Cartesian coordinate system has been chosen, as, in this case, the inner product of two vectors is the dot product of their coordinate vectors. For this reason, and for historical reasons, the dot notation is more commonly used than the bracket notation for the inner product of Euclidean spaces. This article will follow this usage; that is formula_34 will be denoted in the remainder of this article.
The Euclidean norm of a vector is
The inner product and the norm allows expressing and proving all metric and topological properties of Euclidean geometry. The next subsection describe the most fundamental ones. "In these subsections," "denotes an arbitrary Euclidean space, and formula_9 denotes its vector space of translations."
The "distance" (more precisely the "Euclidean distance") between two points of a Euclidean space is the norm of the translation vector that maps one point to the other; that is
The "length" of a segment is the distance between its endpoints. It is often denoted .
The distance is a metric, as it satisfies the triangular inequality
Moreover, the equality is true if and only if belongs to the segment .
This inequality means that the length of any edge of a triangle is smaller than the sum of the lengths of the other edges. This is the origin of the term "triangular inequality".
With the Euclidean distance, every Euclidean space is a complete metric space.
Two nonzero vectors and of formula_9 are "perpendicular" or "orthogonal" if their inner product is zero:
Two linear subspaces of formula_9 are orthogonal if every nonzero vector of the first one is perpendicular to every nonzero vector of the second one. This implies that the intersection of the linear subspace is reduced to the zero vector.
Two lines, and more generally two Euclidean subspaces are orthogonal if their direction are orthogonal. Two orthogonal lines that intersect are said "perpendicular".
Two segments and that share a common endpoint are "perpendicular" or "form a right angle" if the vectors formula_42 and formula_43 are orthogonal.
If and form a right angle, one has
This is the Pythagorean theorem. Its proof is easy in this context, as, expressing this in terms of the inner product, one has, using bilinearity and symmetry of the inner product:
The (non-oriented) "angle" between two nonzero vectors and in formula_9 is
where is the principal value of the arccosine function. By Cauchy–Schwarz inequality, the argument of the arcsine is in the interval . Therefore is real, and (or } if angles are measured in degrees).
Angles are not useful in a Euclidean line, as they can be only 0 or "".
In an oriented Euclidean plane, one can define the "oriented angle" of two vectors. The oriented angle of two vectors and is then the opposite of the oriented angle of and . In this case, the angle of two vectors can have any value modulo an integer multiple of . In particular, a reflex angle equals the negative angle .
The angle of two vectors does not change if they are multiplied by positive numbers. More precisely, if and are two vectors, and and are real numbers, then
If and are three points in a Euclidean space, the angle of the segments and is the angle of the vectors formula_42 and formula_50 As the multiplication of vectors by positive numbers do not change the angle, the angle of two half-lines with initial point can be defined: it is the angle of the segments and , where and are arbitrary points, one on each half-line. Although this is less used, one can define similarly the angle of segments or half-lines that do not share an initial points.
The angle of two lines is defined as follows. If is the angle of two segments, one on each line, the angle of any two other segments, one on each line, is either or . One of these angles is in the interval , and the other being in . The "non-oriented angle" of the two lines is the one in the interval . In an oriented Euclidean plane, the "oriented angle" of two lines belongs to the interval .
Every Euclidean vector space has an orthonormal basis (in fact, infinitely many in dimension higher than one, and two in dimension one), that is a basis formula_51 of unit vectors (formula_52) that are pairwise orthogonal (formula_53 for ). More precisely, given any basis formula_54 the Gram–Schmidt process computes an orthonormal basis such that, for every , the linear spans of formula_55 and formula_56 are equal.
Given a Euclidean space , a "Cartesian frame" is a set of data consisting of an orthonormal basis of formula_22 and a point of , called the "origin" and often denoted . A Cartesian frame formula_58 allows defining Cartesian coordinates for both and formula_9 in the following way.
The Cartesian coordinates of a vector are the coefficients of on the basis formula_60 As the basis is orthonormal, the th coefficient is the dot product formula_61
The Cartesian coordinates of a point of are the Cartesian coordinates of the vector formula_62
As a Euclidean space is an affine space, one can consider an affine frame on it, which is the same as a Euclidean frame, except that the basis is not required to be orthonormal. This define affine coordinates, sometimes called "skew coordinates" for emphasizing that the basis vectors are not pairwise orthogonal.
An affine basis of a Euclidean space of dimension is a set of points that are not contained in a hyperplan. An affine basis define barycentric coordinates for every point.
Many other coordinates systems can be defined on a Euclidean space of dimension , in the following way. Let be a homeomorphism (or, more often, a diffeomorphism) from a dense open subset of to an open subset of formula_5 The "coordinates" of a point of are the components of . The polar coordinate system (dimension 2) and the spherical and cylindrical coordinate systems (dimension 3) are defined this way.
For points that are outside the domain of , coordinates may sometimes be defined as the limit of coordinates of neighbour points, but these coordinates may be not uniquely defined, and may be not continuous in the neighborhood of the point. For example, for the spherical coordinate system, the longitude is not defined at the pole, and on the antimeridian, the longitude passes discontinuousely from –180° to +180°.
This way of defining coordinates extends easily to other mathematical structures, and in particular to manifolds.
An isometry between two metric spaces is a bijection preserving the distance, that is
In the case of a Euclidean vector space, an isometry preserves the norm
since the norm of a vector is its distance from the zero vector. It preserves also the inner product
since
An isometry Euclidean vector spaces is a linear isomorphism.
An isometry formula_68 of Euclidean spaces defines an isometry
formula_69 of the associated Euclidean vector spaces. This implies that two isometric Euclidean spaces have the same dimension. Conversely, if and are Euclidean spaces, , , and formula_69 is an isometry, then the map formula_68 defined by
is an isometry of Euclidean spaces.
It follows from the preceding results that an isometry of Euclidean spaces maps lines to lines, and, more generally Euclidean subspaces to Euclidean subspaces of the same dimension, and that the restriction of the isometry on these subspaces are isometries of these subspaces.
If is a Euclidean space, its associated vector space formula_9 can be considered as a Euclidean space. Every point defines an isometry of Euclidean spaces
which maps to the zero vector and has the identity as associated linear map. The inverse isometry is the map
A Euclidean frame allows defining the map
which is an isometry of Euclidean spaces. The inverse isometry is
"This means that, up to an isomorphism, there is exactly one Euclidean space of a given dimension."
This justifies that many authors talk of formula_2 as "the" Euclidean space of dimension .
An isometry from a Euclidean space onto itself is called "Euclidean isometry", "Euclidean transformation" or "rigid transformation". The rigid transformations of a Euclidean space form a group (under composition), called the "Euclidean group" and often denoted of .
The simplest Euclidean transformations are translations
They are in bijective correspondence with vectors. This is a reason for calling "space of translations" the vector space associated to a Euclidean space. The translations form a normal subgroup of the Euclidean group.
A Euclidean isometry of a Euclidean space defines a linear isometry formula_80 of the associated vector space (by "linear isometry", it is meant an isometry that is also a linear map) in the following way: denoting by the vector formula_81, if is an arbitrary point of , one has
It is straightforward to prove that this is a linear map that does not depend from the choice of
The map formula_83 is a group homomorphism from the Euclidean group onto the group of linear isometries, called the orthogonal group. The kernel of this homomorphism is the translation group, showing that it is a normal subgroup of the Euclidean group.
The isometries that fix a given point form the stabilizer subgroup of the Euclidean group with respect to . The restriction to this stabilizer of above group homomorphism is an isomorphism. So the isometries that fix a given point form a group isomorphic to the orthogonal group.
Let be a point, an isometry, and the translation that maps to . The isometry formula_84 fixes . So formula_85 and "the Euclidean group is the semidirect product of the translation group and the orthogonal group."
The special orthogonal group is the normal subgroup of the orthogonal group that preserves handeness. It is a subgroup of index two of the orthogonal group. Its inverse image by the group homomorphism formula_83 is a normal subgroup of index two of the Euclidean group, which is called the "special Euclidean group" or the "displacement group". Its elements are called "rigid motions" or "displacements".
Rigid motions include the identity, translations, rotations (the rigid motions that fix at least a point), and also screw motions.
Typical examples of rigid transformations that are not rigid motions are reflections, which are rigid transformations that fix a hyperplane and are not the identity. They are also the transformations consisting in changing the sign of one coordinate over some Euclidean frame.
As the special Euclidean group is a subgroup of index two of the Euclidean group, given a reflection , every rigid transformation that is not a rigid motion is the product of and a rigid motion. A glide reflection is an example of a rigid transformation that is not a rigid motion or a reflection.
All groups that have been considered in this section are Lie groups and algebraic groups.
The Euclidean distance makes a Euclidean space a metric space, and thus a topological space. This topology is called the Euclidean topology. In the case of formula_1 this topology is also the product topology.
The open sets are the subsets that contains an open ball around each of their points. In other words, open balls form a base of the topology.
The topological dimension of a Euclidean space equals its dimension. This implies that Euclidean spaces of different dimensions are not homeomorphic. Moreover, the theorem of invariance of domain asserts that a subset of a Euclidean space is open (for the subspace topology) if and only if it is homeomorphic to an open subset of a Euclidean space of the same dimension.
Euclidean spaces are complete and locally compact. That is, a closed subset of a Euclidean space is compact if it is bounded (that is, contained in a ball). In particular, closed balls are compact.
The definition of Euclidean spaces that has been described in this article differs fundamentally of Euclid's one. In reality, Euclid did not define formally the space, because it was thought as a description of the physical world that exists independently of human mind. The need of a formal definition appeared only at the end of 19th century, with the introduction of non-Euclidean geometries.
Two different approaches have been used. Felix Klein suggested to define geometries through their symmetries. The presentation of Euclidean spaces given in this article, is essentially issued from his Erlangen program, with the emphasis given on the groups of translations and isometries.
On the other hand, David Hilbert proposed a set of axioms, inspired by Euclid's postulates. They belong to synthetic geometry, as they do not involve any definition of real numbers. Later G. D. Birkhoff and Alfred Tarski proposed simpler sets of axioms, which use real numbers (see Birkhoff's axioms and Tarski's axioms).
In "Geometric Algebra", Emil Artin has proved that all these definitions of a Euclidean space are equivalent. It is rather easy to prove that all definitions of Euclidean spaces satisfy Hilbert's axioms, and that those involving real numbers (including the above given definition) are equivalent. The difficult part of Artin's proof is the following. In Hilbert's axioms, congruence is an equivalence relation on segments. One can thus define the "length" of a segment as its equivalence class. One must thus prove that this length satisfies properties that characterize nonnegative real numbers. It is what did Artin, with axioms that are not Hilbert's ones, but are equivalent.
Since ancient Greeks, Euclidean space is used for modeling shapes in the physical world. It is thus used in many sciences such as physics, mechanics, and astronomy. It is also widely used in all technical areas that are concerned with shapes, figure, location and position, such as architecture, geodesy, topography, navigation, industrial design, or technical drawing.
Space of dimensions higher than three occurs in several modern theories of physics; see Higher dimension. They occur also in configuration spaces of physical systems.
Beside Euclidean geometry, Euclidean spaces are also widely used in other areas of mathematics. Tangent spaces of differentiable manifolds are Euclidean vector spaces. More generally, a manifold is a space that is locally approximated by Euclidean spaces. Most non-Euclidean geometries can be modeled by a manifold, and embedded in a Euclidean space of higher dimension. For example, an elliptic space can be modeled by an ellipsoid. It is common to represent in a Euclidean space mathematics objects that are "a priori" not of a geometrical nature. An example among many is the usual representation of graphs.
Since the introduction, at the end of 19th century, of Non-Euclidean geometries, many sorts of spaces have been considered, about which one can do geometric reasoning in the same way as with Euclidean spaces. In general, they share some properties with Euclidean spaces, but may also have properties that could appear as rather strange. Some of these spaces use Euclidean geometry for their definition, or can be modeled as subspaces of a Euclidean space of higher dimension. When such a space is defined by geometrical axioms, embedding the space in a Euclidean space is a standard way for proving consistency of its definition, or, more precisely for proving that its theory is consistent, if Euclidean geometry is consistent (which cannot be proved).
A Euclidean space is an affine space equipped with a metric. Affine spaces have many other uses in mathematics. In particular, as they are defined over any field, they allow doing geometry in other contexts.
As soon as non-linear questions are considered, it is generally useful to consider affine spaces over the complex numbers as an extension of Euclidean spaces. For example, a circle and a line have always two intersection points (possibly not distinct) in the complex affine space. Therefore, most of algebraic geometry is built in complex affine spaces and affine spaces over algebraically closed fields. The shapes that are studied in algebraic geometry in these affine spaces are therefore called affine algebraic varieties.
Affine spaces over the rational numbers and more generally over algebraic number fields provide a link between (algebraic) geometry and number theory. For example, the Fermat's Last Theorem can be stated "a Fermat curve of degree higher than two has no point in the affine plane over the rationals."
Geometry in affine spaces over a finite fields has also been widely studied. For example, elliptic curves over finite fields are widely used in cryptography.
Originally, projective spaces have been introduced by adding "points at infinity" to Euclidean spaces, and, more generally to affine spaces, in order to make true the assertion "two coplanar lines meet in exactly one point". Projective space share with Euclidean and affine spaces the property of being isotropic, that is, there is no property of the space that allows distinguishing between two points or two lines. Therefore, a more isotropic definition is commonly used, which consists as defining a projective space as the set of the vector lines in a vector space of dimension one more.
As for affine spaces, projective spaces are defined over any field, and are fundamental spaces of algebraic geometry.
"Non-Euclidean geometry" refers usually to geometrical spaces where the parallel postulate is false. They include elliptic geometry, where the sum of the angles of a triangle is more than 180°, and hyperbolic geometry, where this sum is less than 180°. Their introduction in the second half of 19th century, and the proof that their theory is consistent (if Euclidean geometry is not contradictory) is one of the paradoxes that are at the origin of the foundational crisis in mathematics of the beginning of 20th century, and motivated the systematization of axiomatic theories in mathematics.
A manifold is a space that in the neighborhood of each point resembles a Euclidean space. In technical terms, a manifold is a topological space, such that each point has a neighborhood that is homeomorphic to an open subset of a Euclidean space. Manifold can be classified by increasing degree of this "resemblance" into topological manifolds, differentiable manifolds, smooth manifolds, and analytic manifolds. However, none of these types of "resemblance" respect distances and angles, even approximately.
Distances and angles can be defined on a smooth manifold by providing a smoothly varying Euclidean metric on the tangent spaces at the points of the manifold (these tangent are thus Euclidean vector spaces). This results in a Riemannian manifold. Generally, straight lines do not exist in a Riemannian manifold, but their role is played by geodesics, which are the "shortest paths" between two points. This allows defining distances, which are measured along geodesics, and angles between geodesics, which are the angle of their tangents in the tangent space at their intersection. So, Riemannian manifolds behave locally like a Euclidean that has been bended.
Euclidean spaces are trivially Riemannian manifolds. An example illustrating this well is the surface of a sphere. In this case, geodesics are arcs of great circle, which are called orthodromes in the context of navigation. More generally, the spaces of non-Euclidean geometries can be realized as Riemannian manifolds.
The inner product that is defined to define Euclidean spaces is a positive definite bilinear form. If it is replaced by an indefinite quadratic form which is non-degenerate, one gets a pseudo-Euclidean space.
A fundamental example of such a space is the Minkowski space, which is the space-time of Einstein's special relativity. It is a four-dimensional space, where the metric is defined by the quadratic form
where the last coordinate ("t") is temporal, and the other three ("x", "y", "z") are spatial.
To take the gravity into account, general relativity uses a pseudo-Riemannian manifold that has Minkowski spaces as tangent spaces. The curvature of this manifold at a point is a function of the value of the gravitational field at this point. | https://en.wikipedia.org/wiki?curid=9697 |
Edwin Austin Abbey
Edwin Austin Abbey (April 1, 1852August 1, 1911) was an American muralist, illustrator, and painter. He flourished at the beginning of what is now referred to as the "golden age" of illustration, and is best known for his drawings and paintings of Shakespearean and Victorian subjects, as well as for his painting of Edward VII's coronation. His most famous set of murals, "The Quest and Achievement of the Holy Grail", adorns the Boston Public Library.
Abbey was born in Philadelphia in 1852. He studied art at the Pennsylvania Academy of the Fine Arts under Christian Schuessele. Abbey began as an illustrator, producing numerous illustrations and sketches for such magazines as Harper's Weekly (1871–1874) and Scribner's Magazine. His illustrations began appearing in Harper's Weekly at an early age: before Abbey was twenty years old. He moved to New York City in 1871. His illustrations were strongly influenced by French and German black and white art. He also illustrated several best-selling books, including "Christmas Stories" by Charles Dickens (1875), "Selections from the Poetry of Robert Herrick" (1882), and "She Stoops to Conquer" by Oliver Goldsmith (1887). Abbey also illustrated a four-volume set of "The Comedies of Shakespeare" for Harper & Brothers in 1896.
He moved to England in 1878, at the request of his employers, to gather material for illustrations of the poems of Robert Herrick, published in 1882, and he settled permanently there in 1883. In 1883, he was elected to the Royal Institute of Painters in Water-Colours. About this time, he was appraised critically by the American writer, S.G.W. Benjamin:
He also created illustrations for Goldsmith's "She Stoops to Conquer" (1887), for a volume of "Old Songs" (1889), and for the comedies (and a few of the tragedies) of Shakespeare. Among his water-colours are "The Evil Eye" (1877), "The Rose in October" (1879), "An Old Song" (1886), "The Visitors" (1890), and "The Jongleur" (1892). Possibly his best known pastels are "Beatrice," "Phyllis," and "Two Noble Kinsmen."
In 1890 he made his first appearance with an oil painting, "A May Day Morn," at the Royal Academy in London. He exhibited "Richard duke of Gloucester and the Lady Anne" there in 1896, and in that year was elected A.R.A., becoming a full member in 1898. In 1902 he was chosen to paint the coronation of King Edward VII. It was the official painting of the occasion and, hence, resides at Buckingham Palace. He did receive a knighthood, although some say he refused it in 1907. Friendly with other expatriate American artists, he summered at Broadway, Worcestershire, England, where he painted and vacationed alongside John Singer Sargent at the home of Francis Davis Millet.
He completed murals for the Boston Public Library in the 1890s. The frieze for the Library was titled "The Quest and Achievement of the Holy Grail". It took Abbey eleven years to complete this series of murals in his England studio.
In 1904 he painted a mural for the Royal Exchange, London "Reconciliation of the Skinners & Merchant Taylors' Companies by Lord Mayor Billesden, 1484".
In 1908–09, Abbey began an ambitious program of murals and other artworks for the newly completed Pennsylvania State Capitol in Harrisburg, Pennsylvania. These included allegorical medallion murals representing "Science", "Art", "Justice", and "Religion" for the dome of the Rotunda, four large lunette murals beneath the dome, and multiple works for the House and Senate Chambers. For the Senate chamber he finished only one painting, "Von Steuben Training the American Soldiers at Valley Forge", and he was working on the "Reading of the Declaration of Independence" mural in early 1911, when his health began to fail. He was diagnosed with cancer. Studio assistant William Simmonds continued work on the mural with little supervision from Abbey, and with small contributions by John Singer Sergeant.
Abbey died in August 1911. William Simmonds travelled from England to install the completed murals with Abbey's widow Gertrude. The remaining two rooms, which Abbey had been unable to finish, were given to Violet Oakley, who completed the commission using her own designs.
Abbey was elected to the National Academy of Design, in 1902, and The American Academy of Arts and Letters. He was honorary member of the Royal Bavarian Society and the Société Nationale des Beaux-Arts, and was made a chevalier of the French Legion of Honour. He was a prolific illustrator, and attention to detail, including historical accuracy, influenced successive generations of illustrators.
In 1890, Edwin married Gertrude Mead, the daughter of a wealthy New York merchant. Mrs Abbey encouraged her husband to secure more ambitious commissions, although with their marriage commencing when both were in their forties, the couple remained childless. After her husband's death, Gertrude was active in preserving her husband's legacy, writing about his work and giving her substantial collection and archive to Yale. Edwin had been a keen supporter of the newly founded British School at Rome (BSR), so, in his memory, she donated £6000 to assist in building the artists' studio block and, in 1926, founded the Incorporated Edwin Austin Abbey Memorial Scholarships. The scholarships were established to enable British and American painters to pursue their practice. Recipients of Abbey funding – Scholars and, more recently, Fellows – devote their scholarship to working in the studios at the BSR, where there has, ever since, been at least one Abbey-funded artist in residence. Previous award holders include Stephen Farthing, Chantal Joffe and Spartacus Chetwynd. The Abbey Fellowships (formerly 'Awards') were established in their present form in 1990, and the Abbey studios also host the BSR's other fine art residencies, such as the Derek Hill Foundation Scholarship and the Sainsbury Scholarship in Painting and Drawing. A bust of Edwin Abbey, by Sir Thomas Brock, stands in the courtyard of the BSR. Edwin also left bequests of his works to the Metropolitan Museum of Art in New York, to the Museum of Fine Arts, Boston and to the National Gallery in London.
Abbey is buried in the churchyard of Old St Andrew's Church in Kingsbury, London. His grave is Grade II listed. | https://en.wikipedia.org/wiki?curid=9700 |
Evolutionary psychology
Evolutionary psychology is a theoretical approach in the social and natural sciences that examines psychological structure from a modern evolutionary perspective. It seeks to identify which human psychological traits are evolved adaptations – that is, the functional products of natural selection or sexual selection in human evolution. Adaptationist thinking about physiological mechanisms, such as the heart, lungs, and immune system, is common in evolutionary biology. Some evolutionary psychologists apply the same thinking to psychology, arguing that the modularity of mind is similar to that of the body and with different modular adaptations serving different functions. Evolutionary psychologists argue that much of human behavior is the output of psychological adaptations that evolved to solve recurrent problems in human ancestral environments.
Evolutionary psychology is not simply a subdiscipline of psychology but its evolutionary theory can provide a foundational, metatheoretical framework that integrates the entire field of psychology in the same way evolutionary biology has for biology.
Evolutionary psychologists hold that behaviors or traits that occur universally in all cultures are good candidates for evolutionary adaptations including the abilities to infer others' emotions, discern kin from non-kin, identify and prefer healthier mates, and cooperate with others. There have been studies of human social behaviour related to infanticide, intelligence, marriage patterns, promiscuity, perception of beauty, bride price, and parental investment, with impressive findings. The theories and findings of evolutionary psychology have applications in many fields, including economics, environment, health, law, management, psychiatry, politics, and literature.
Criticism of evolutionary psychology involves questions of testability, cognitive and evolutionary assumptions (such as modular functioning of the brain, and large uncertainty about the ancestral environment), importance of non-genetic and non-adaptive explanations, as well as political and ethical issues due to interpretations of research results.
Evolutionary psychology is an approach that views human nature as the product of a universal set of evolved psychological adaptations to recurring problems in the ancestral environment. Proponents suggest that it seeks to integrate psychology into the other natural sciences, rooting it in the organizing theory of biology (evolutionary theory), and thus understanding psychology as a branch of biology. Anthropologist John Tooby and psychologist Leda Cosmides note:
Evolutionary psychology is the long-forestalled scientific attempt to assemble out of the disjointed, fragmentary, and mutually contradictory human disciplines a single, logically integrated research framework for the psychological, social, and behavioral sciences – a framework
that not only incorporates the evolutionary sciences on a full and equal basis, but that systematically works out all of the revisions in existing belief and research practice that such a synthesis requires.
Just as human physiology and evolutionary physiology have worked to identify physical adaptations of the body that represent "human physiological nature," the purpose of evolutionary psychology is to identify evolved emotional and cognitive adaptations that represent "human psychological nature." According to Steven Pinker, it is "not a single theory but a large set of hypotheses" and a term that "has also come to refer to a particular way of applying evolutionary theory to the mind, with an emphasis on adaptation, gene-level selection, and modularity." Evolutionary psychology adopts an understanding of the mind that is based on the computational theory of mind. It describes mental processes as computational operations, so that, for example, a fear response is described as arising from a neurological computation that inputs the perceptional data, e.g. a visual image of a spider, and outputs the appropriate reaction, e.g. fear of possibly dangerous animals. Under this view, any Domain-general learning is impossible because of the Combinatorial explosion. This implies Domain-specific learning. Evolutionary Psychology specifies the domain as the problems of survival and reproduction.
While philosophers have generally considered the human mind to include broad faculties, such as reason and lust, evolutionary psychologists describe evolved psychological mechanisms as narrowly focused to deal with specific issues, such as catching cheaters or choosing mates. The discipline views the human brain as comprising many functional mechanisms, called "psychological adaptations" or evolved cognitive mechanisms or "cognitive modules", designed by the process of natural selection. Examples include language-acquisition modules, incest-avoidance mechanisms, cheater-detection mechanisms, intelligence and sex-specific mating preferences, foraging mechanisms, alliance-tracking mechanisms, agent-detection mechanisms, and others. Some mechanisms, termed "domain-specific", deal with recurrent adaptive problems over the course of human evolutionary history. "Domain-general" mechanisms, on the other hand, are proposed to deal with evolutionary novelty.
Evolutionary psychology has roots in cognitive psychology and evolutionary biology but also draws on behavioral ecology, artificial intelligence, genetics, ethology, anthropology, archaeology, biology, and zoology. It is closely linked to sociobiology, but there are key differences between them including the emphasis on "domain-specific" rather than "domain-general" mechanisms, the relevance of measures of current fitness, the importance of mismatch theory, and psychology rather than behavior. Most of what is now labeled as sociobiological research is now confined to the field of behavioral ecology.
Nikolaas Tinbergen's four categories of questions can help to clarify the distinctions between several different, but complementary, types of explanations. Evolutionary psychology focuses primarily on the "why?" questions, while traditional psychology focuses on the "how?" questions.
Evolutionary psychology is founded on several core premises.
Evolutionary psychology has its historical roots in Charles Darwin's theory of natural selection. In "The Origin of Species", Darwin predicted that psychology would develop an evolutionary basis:
Two of his later books were devoted to the study of animal emotions and psychology; "The Descent of Man, and Selection in Relation to Sex" in 1871 and "The Expression of the Emotions in Man and Animals" in 1872. Darwin's work inspired William James's functionalist approach to psychology. Darwin's theories of evolution, adaptation, and natural selection have provided insight into why brains function the way they do.
The content of evolutionary psychology has derived from, on the one hand, the biological sciences (especially evolutionary theory as it relates to ancient human environments, the study of paleoanthropology and animal behavior) and, on the other, the human sciences, especially psychology.
Evolutionary biology as an academic discipline emerged with the modern synthesis in the 1930s and 1940s. In the 1930s the study of animal behavior (ethology) emerged with the work of the Dutch biologist Nikolaas Tinbergen and the Austrian biologists Konrad Lorenz and Karl von Frisch.
W.D. Hamilton's (1964) papers on inclusive fitness and Robert Trivers's (1972) theories on reciprocity and parental investment helped to establish evolutionary thinking in psychology and the other social sciences. In 1975, Edward O. Wilson combined evolutionary theory with studies of animal and social behavior, building on the works of Lorenz and Tinbergen, in his book "".
In the 1970s, two major branches developed from ethology. Firstly, the study of animal "social" behavior (including humans) generated sociobiology, defined by its pre-eminent proponent Edward O. Wilson in 1975 as "the systematic study of the biological basis of all social behavior" and in 1978 as "the extension of population biology and evolutionary theory to social organization." Secondly, there was behavioral ecology which placed less emphasis on "social" behavior; it focused on the ecological and evolutionary basis of animal and human behavior.
In the 1970s and 1980s university departments began to include the term "evolutionary biology" in their titles. The modern era of evolutionary psychology was ushered in, in particular, by Donald Symons' 1979 book "The Evolution of Human Sexuality" and Leda Cosmides and John Tooby's 1992 book "The Adapted Mind". David Buller observed that the term "evolutionary psychology" is sometimes seen as denoting research based on the specific methodological and theoretical commitments of certain researchers from the Santa Barbara school (University of California), thus some evolutionary psychologists prefer to term their work "human ecology", "human behavioural ecology" or "evolutionary anthropology" instead.
From psychology there are the primary streams of developmental, social and cognitive psychology. Establishing some measure of the relative influence of genetics and environment on behavior has been at the core of behavioral genetics and its variants, notably studies at the molecular level that examine the relationship between genes, neurotransmitters and behavior. Dual inheritance theory (DIT), developed in the late 1970s and early 1980s, has a slightly different perspective by trying to explain how human behavior is a product of two different and interacting evolutionary processes: genetic evolution and cultural evolution. DIT is seen by some as a "middle-ground" between views that emphasize human universals versus those that emphasize cultural variation.
The theories on which evolutionary psychology is based originated with Charles Darwin's work, including his speculations about the evolutionary origins of social instincts in humans. Modern evolutionary psychology, however, is possible only because of advances in evolutionary theory in the 20th century.
Evolutionary psychologists say that natural selection has provided humans with many psychological adaptations, in much the same way that it generated humans' anatomical and physiological adaptations. As with adaptations in general, psychological adaptations are said to be specialized for the environment in which an organism evolved, the environment of evolutionary adaptedness. Sexual selection provides organisms with adaptations related to mating. For male mammals, which have a relatively high maximal potential reproduction rate, sexual selection leads to adaptations that help them compete for females. For female mammals, with a relatively low maximal potential reproduction rate, sexual selection leads to choosiness, which helps females select higher quality mates. Charles Darwin described both natural selection and sexual selection, and he relied on group selection to explain the evolution of altruistic (self-sacrificing) behavior. But group selection was considered a weak explanation, because in any group the less altruistic individuals will be more likely to survive, and the group will become less self-sacrificing as a whole.
In 1964, William D. Hamilton proposed inclusive fitness theory, emphasizing a gene-centered view of evolution. Hamilton noted that genes can increase the replication of copies of themselves into the next generation by influencing the organism's social traits in such a way that (statistically) results in helping the survival and reproduction of other copies of the same genes (most simply, identical copies in the organism's close relatives). According to Hamilton's rule, self-sacrificing behaviors (and the genes influencing them) can evolve if they typically help the organism's close relatives so much that it more than compensates for the individual animal's sacrifice. Inclusive fitness theory resolved the issue of how altruism can evolve. Other theories also help explain the evolution of altruistic behavior, including evolutionary game theory, tit-for-tat reciprocity, and generalized reciprocity. These theories help to explain the development of altruistic behavior, and account for hostility toward cheaters (individuals that take advantage of others' altruism).
Several mid-level evolutionary theories inform evolutionary psychology. The r/K selection theory proposes that some species prosper by having many offspring, while others follow the strategy of having fewer offspring but investing much more in each one. Humans follow the second strategy. Parental investment theory explains how parents invest more or less in individual offspring based on how successful those offspring are likely to be, and thus how much they might improve the parents' inclusive fitness. According to the Trivers–Willard hypothesis, parents in good conditions tend to invest more in sons (who are best able to take advantage of good conditions), while parents in poor conditions tend to invest more in daughters (who are best able to have successful offspring even in poor conditions). According to life history theory, animals evolve life histories to match their environments, determining details such as age at first reproduction and number of offspring. Dual inheritance theory posits that genes and human culture have interacted, with genes affecting the development of culture, and culture, in turn, affecting human evolution on a genetic level (see also the Baldwin effect).
Evolutionary psychology is based on the hypothesis that, just like hearts, lungs, livers, kidneys, and immune systems, cognition has functional structure that has a genetic basis, and therefore has evolved by natural selection. Like other organs and tissues, this functional structure should be universally shared amongst a species, and should solve important problems of survival and reproduction.
Evolutionary psychologists seek to understand psychological mechanisms by understanding the survival and reproductive functions they might have served over the course of evolutionary history. These might include abilities to infer others' emotions, discern kin from non-kin, identify and prefer healthier mates, cooperate with others and follow leaders. Consistent with the theory of natural selection, evolutionary psychology sees humans as often in conflict with others, including mates and relatives. For instance, a mother may wish to wean her offspring from breastfeeding earlier than does her infant, which frees up the mother to invest in additional offspring. Evolutionary psychology also recognizes the role of kin selection and reciprocity in evolving prosocial traits such as altruism. Like chimpanzees and bonobos, humans have subtle and flexible social instincts, allowing them to form extended families, lifelong friendships, and political alliances. In studies testing theoretical predictions, evolutionary psychologists have made modest findings on topics such as infanticide, intelligence, marriage patterns, promiscuity, perception of beauty, bride price and parental investment.
Proponents of evolutionary psychology in the 1990s made some explorations in historical events, but the response from historical experts was highly negative and there has been little effort to continue that line of research. Historian Lynn Hunt says that the historians complained that the researchers:
Hunt states that, "the few attempts to build up a subfield of psychohistory collapsed under the weight of its presuppositions." She concludes that as of 2014 the "'iron curtain' between historians and psychology...remains standing."
Not all traits of organisms are evolutionary adaptations. As noted in the table below, traits may also be exaptations, byproducts of adaptations (sometimes called "spandrels"), or random variation between individuals.
Psychological adaptations are hypothesized to be innate or relatively easy to learn, and to manifest in cultures worldwide. For example, the ability of toddlers to learn a language with virtually no training is likely to be a psychological adaptation. On the other hand, ancestral humans did not read or write, thus today, learning to read and write require extensive training, and presumably involves the repurposing of cognitive capacities that evolved in response to selection pressures unrelated to written language. However, variations in manifest behavior can result from universal mechanisms interacting with different local environments. For example, Caucasians who move from a northern climate to the equator will have darker skin. The mechanisms regulating their pigmentation do not change; rather the input to those mechanisms change, resulting in different output.
One of the tasks of evolutionary psychology is to identify which psychological traits are likely to be adaptations, byproducts or random variation. George C. Williams suggested that an "adaptation is a special and onerous concept that should only be used where it is really necessary." As noted by Williams and others, adaptations can be identified by their improbable complexity, species universality, and adaptive functionality.
A question that may be asked about an adaptation is whether it is generally obligate (relatively robust in the face of typical environmental variation) or facultative (sensitive to typical environmental variation). The sweet taste of sugar and the pain of hitting one's knee against concrete are the result of fairly obligate psychological adaptations; typical environmental variability during development does not much affect their operation. By contrast, facultative adaptations are somewhat like "if-then" statements. For example, adult attachment style seems particularly sensitive to early childhood experiences. As adults, the propensity to develop close, trusting bonds with others is dependent on whether early childhood caregivers could be trusted to provide reliable assistance and attention. The adaptation for skin to tan is conditional to exposure to sunlight; this is an example of another facultative adaptation. When a psychological adaptation is facultative, evolutionary psychologists concern themselves with how developmental and environmental inputs influence the expression of the adaptation.
Evolutionary psychologists hold that behaviors or traits that occur universally in all cultures are good candidates for evolutionary adaptations. Cultural universals include behaviors related to language, cognition, social roles, gender roles, and technology. Evolved psychological adaptations (such as the ability to learn a language) interact with cultural inputs to produce specific behaviors (e.g., the specific language learned).
Basic gender differences, such as greater eagerness for sex among men and greater coyness among women, are explained as sexually dimorphic psychological adaptations that reflect the different reproductive strategies of males and females. Research has also shown participants were able to recognize the facial expression of fear significantly better on a male face than on a female face. Females also recognized fear generally better than males.
Evolutionary psychologists contrast their approach to what they term the "standard social science model," according to which the mind is a general-purpose cognition device shaped almost entirely by culture.
Evolutionary psychology argues that to properly understand the functions of the brain, one must understand the properties of the environment in which the brain evolved. That environment is often referred to as the "environment of evolutionary adaptedness".
The idea of an "environment of evolutionary adaptedness" was first explored as a part of attachment theory by John Bowlby. This is the environment to which a particular evolved mechanism is adapted. More specifically, the environment of evolutionary adaptedness is defined as the set of historically recurring selection pressures that formed a given adaptation, as well as those aspects of the environment that were necessary for the proper development and functioning of the adaptation.
Humans, comprising the genus "Homo", appeared between 1.5 and 2.5 million years ago, a time that roughly coincides with the start of the Pleistocene 2.6 million years ago. Because the Pleistocene ended a mere 12,000 years ago, most human adaptations either newly evolved during the Pleistocene, or were maintained by stabilizing selection during the Pleistocene. Evolutionary psychology therefore proposes that the majority of human psychological mechanisms are adapted to reproductive problems frequently encountered in Pleistocene environments. In broad terms, these problems include those of growth, development, differentiation, maintenance, mating, parenting, and social relationships.
The environment of evolutionary adaptedness is significantly different from modern society. The ancestors of modern humans lived in smaller groups, had more cohesive cultures, and had more stable and rich contexts for identity and meaning. Researchers look to existing hunter-gatherer societies for clues as to how hunter-gatherers lived in the environment of evolutionary adaptedness. Unfortunately, the few surviving hunter-gatherer societies are different from each other, and they have been pushed out of the best land and into harsh environments, so it is not clear how closely they reflect ancestral culture. However, all around the world small-band hunter-gatherers offer a similar developmental system for the young ("hunter-gatherer childhood model," Konner, 2005;
"evolved developmental niche" or "evolved nest;" Narvaez et al., 2013). The characteristics of the niche are largely the same as for social mammals, who evolved over 30 million years ago: soothing perinatal experience, several years of on-request breastfeeding, nearly constant affection or physical proximity, responsiveness to need (mitigating offspring distress), self-directed play, and for humans, multiple responsive caregivers. Initial studies show the importance of these components in early life for positive child outcomes.
Evolutionary psychologists sometimes look to chimpanzees, bonobos, and other great apes for insight into human ancestral behavior.
Since an organism's adaptations were suited to its ancestral environment, a new and different environment can create a mismatch. Because humans are mostly adapted to Pleistocene environments, psychological mechanisms sometimes exhibit "mismatches" to the modern environment. One example is the fact that although about 10,000 people are killed with guns in the US annually, whereas spiders and snakes kill only a handful, people nonetheless learn to fear spiders and snakes about as easily as they do a pointed gun, and more easily than an unpointed gun, rabbits or flowers. A potential explanation is that spiders and snakes were a threat to human ancestors throughout the Pleistocene, whereas guns (and rabbits and flowers) were not. There is thus a mismatch between humans' evolved fear-learning psychology and the modern environment.
This mismatch also shows up in the phenomena of the supernormal stimulus, a stimulus that elicits a response more strongly than the stimulus for which the response evolved. The term was coined by Niko Tinbergen to refer to non-human animal behavior, but psychologist Deirdre Barrett said that supernormal stimulation governs the behavior of humans as powerfully as that of other animals. She explained junk food as an exaggerated stimulus to cravings for salt, sugar, and fats, and she says that television is an exaggeration of social cues of laughter, smiling faces and attention-grabbing action. Magazine centerfolds and double cheeseburgers pull instincts intended for an environment of evolutionary adaptedness where breast development was a sign of health, youth and fertility in a prospective mate, and fat was a rare and vital nutrient. The psychologist Mark van Vugt recently argued that modern organizational leadership is a mismatch. His argument is that humans are not adapted to work in large, anonymous bureaucratic structures with formal hierarchies. The human mind still responds to personalized, charismatic leadership primarily in the context of informal, egalitarian settings. Hence the dissatisfaction and alienation that many employees experience. Salaries, bonuses and other privileges exploit instincts for relative status, which attract particularly males to senior executive positions.
Evolutionary theory is heuristic in that it may generate hypotheses that might not be developed from other theoretical approaches. One of the major goals of adaptationist research is to identify which organismic traits are likely to be adaptations, and which are byproducts or random variations. As noted earlier, adaptations are expected to show evidence of complexity, functionality, and species universality, while byproducts or random variation will not. In addition, adaptations are expected to manifest as proximate mechanisms that interact with the environment in either a generally obligate or facultative fashion (see above). Evolutionary psychologists are also interested in identifying these proximate mechanisms (sometimes termed "mental mechanisms" or "psychological adaptations") and what type of information they take as input, how they process that information, and their outputs. Evolutionary developmental psychology, or "evo-devo," focuses on how adaptations may be activated at certain developmental times (e.g., losing baby teeth, adolescence, etc.) or how events during the development of an individual may alter life history trajectories.
Evolutionary psychologists use several strategies to develop and test hypotheses about whether a psychological trait is likely to be an evolved adaptation. Buss (2011) notes that these methods include:
Evolutionary psychologists also use various sources of data for testing, including experiments, archaeological records, data from hunter-gatherer societies, observational studies, neuroscience data, self-reports and surveys, public records, and human products.
Recently, additional methods and tools have been introduced based on fictional scenarios, mathematical models, and multi-agent computer simulations.
Foundational areas of research in evolutionary psychology can be divided into broad categories of adaptive problems that arise from the theory of evolution itself: survival, mating, parenting, family and kinship, interactions with non-kin, and cultural evolution.
Problems of survival are clear targets for the evolution of physical and psychological adaptations. Major problems the ancestors of present-day humans faced included food selection and acquisition; territory selection and physical shelter; and avoiding predators and other environmental threats.
Consciousness meets George Williams' criteria of species universality, complexity, and functionality, and it is a trait that apparently increases fitness.
In his paper "Evolution of consciousness," John Eccles argues that special anatomical and physical adaptations of the mammalian cerebral cortex gave rise to consciousness. In contrast, others have argued that the recursive circuitry underwriting consciousness is much more primitive, having evolved initially in pre-mammalian species because it improves the capacity for interaction with both social "and" natural environments by providing an energy-saving "neutral" gear in an otherwise energy-expensive motor output machine. Once in place, this recursive circuitry may well have provided a basis for the subsequent development of many of the functions that consciousness facilitates in higher organisms, as outlined by Bernard J. Baars. Richard Dawkins suggested that humans evolved consciousness in order to make themselves the subjects of thought. Daniel Povinelli suggests that large, tree-climbing apes evolved consciousness to take into account one's own mass when moving safely among tree branches. Consistent with this hypothesis, Gordon Gallup found that chimps and orangutans, but not little monkeys or terrestrial gorillas, demonstrated self-awareness in mirror tests.
The concept of consciousness can refer to voluntary action, awareness, or wakefulness. However, even voluntary behavior involves unconscious mechanisms. Many cognitive processes take place in the cognitive unconscious, unavailable to conscious awareness. Some behaviors are conscious when learned but then become unconscious, seemingly automatic. Learning, especially implicitly learning a skill, can take place outside of consciousness. For example, plenty of people know how to turn right when they ride a bike, but very few can accurately explain how they actually do so. Evolutionary psychology approaches self-deception as an adaptation that can improve one's results in social exchanges.
Sleep may have evolved to conserve energy when activity would be less fruitful or more dangerous, such as at night, and especially during the winter season.
Many experts, such as Jerry Fodor, write that the purpose of perception is knowledge, but evolutionary psychologists hold that its primary purpose is to guide action. For example, they say, depth perception seems to have evolved not to help us know the distances to other objects but rather to help us move around in space. Evolutionary psychologists say that animals from fiddler crabs to humans use eyesight for collision avoidance, suggesting that vision is basically for directing action, not providing knowledge.
Building and maintaining sense organs is metabolically expensive, so these organs evolve only when they improve an organism's fitness. More than half the brain is devoted to processing sensory information, and the brain itself consumes roughly one-fourth of one's metabolic resources, so the senses must provide exceptional benefits to fitness. Perception accurately mirrors the world; animals get useful, accurate information through their senses.
Scientists who study perception and sensation have long understood the human senses as adaptations to their surrounding worlds. Depth perception consists of processing over half a dozen visual cues, each of which is based on a regularity of the physical world. Vision evolved to respond to the narrow range of electromagnetic energy that is plentiful and that does not pass through objects. Sound waves go around corners and interact with obstacles, creating a complex pattern that includes useful information about the sources of and distances to objects. Larger animals naturally make lower-pitched sounds as a consequence of their size. The range over which an animal hears, on the other hand, is determined by adaptation. Homing pigeons, for example, can hear very low-pitched sound (infrasound) that carries great distances, even though most smaller animals detect higher-pitched sounds. Taste and smell respond to chemicals in the environment that are thought to have been significant for fitness in the environment of evolutionary adaptedness. For example, salt and sugar were apparently both valuable to the human or pre-human inhabitants of the environment of evolutionary adaptedness, so present day humans have an intrinsic hunger for salty and sweet tastes. The sense of touch is actually many senses, including pressure, heat, cold, tickle, and pain. Pain, while unpleasant, is adaptive. An important adaptation for senses is range shifting, by which the organism becomes temporarily more or less sensitive to sensation. For example, one's eyes automatically adjust to dim or bright ambient light. Sensory abilities of different organisms often coevolve, as is the case with the hearing of echolocating bats and that of the moths that have evolved to respond to the sounds that the bats make.
Evolutionary psychologists contend that perception demonstrates the principle of modularity, with specialized mechanisms handling particular perception tasks. For example, people with damage to a particular part of the brain suffer from the specific defect of not being able to recognize faces (prosopagnosia). Evolutionary psychology suggests that this indicates a so-called face-reading module.
In evolutionary psychology, learning is said to be accomplished through evolved capacities, specifically facultative adaptations. Facultative adaptations express themselves differently depending on input from the environment. Sometimes the input comes during development and helps shape that development. For example, migrating birds learn to orient themselves by the stars during a critical period in their maturation. Evolutionary psychologists believe that humans also learn language along an evolved program, also with critical periods. The input can also come during daily tasks, helping the organism cope with changing environmental conditions. For example, animals evolved Pavlovian conditioning in order to solve problems about causal relationships. Animals accomplish learning tasks most easily when those tasks resemble problems that they faced in their evolutionary past, such as a rat learning where to find food or water. Learning capacities sometimes demonstrate differences between the sexes. In many animal species, for example, males can solve spatial problem faster and more accurately than females, due to the effects of male hormones during development. The same might be true of humans.
Motivations direct and energize behavior, while emotions provide the affective component to motivation, positive or negative. In the early 1970s, Paul Ekman and colleagues began a line of research which suggests that many emotions are universal. He found evidence that humans share at least five basic emotions: fear, sadness, happiness, anger, and disgust. Social emotions evidently evolved to motivate social behaviors that were adaptive in the environment of evolutionary adaptedness. For example, spite seems to work against the individual but it can establish an individual's reputation as someone to be feared. Shame and pride can motivate behaviors that help one maintain one's standing in a community, and self-esteem is one's estimate of one's status.
Motivation has a neurobiological basis in the reward system of the brain. Recently, it has been suggested that reward systems may evolve in such a way that there may be an inherent or unavoidable trade-off in the motivational system for activities of short versus long duration.
Cognition refers to internal representations of the world and internal information processing. From an evolutionary psychology perspective, cognition is not "general purpose," but uses heuristics, or strategies, that generally increase the likelihood of solving problems that the ancestors of present-day humans routinely faced. For example, present day humans are far more likely to solve logic problems that involve detecting cheating (a common problem given humans' social nature) than the same logic problem put in purely abstract terms. Since the ancestors of present-day humans did not encounter truly random events, present day humans may be cognitively predisposed to incorrectly identify patterns in random sequences. "Gamblers' Fallacy" is one example of this. Gamblers may falsely believe that they have hit a "lucky streak" even when each outcome is actually random and independent of previous trials. Most people believe that if a fair coin has been flipped 9 times and Heads appears each time, that on the tenth flip, there is a greater than 50% chance of getting Tails. Humans find it far easier to make diagnoses or predictions using frequency data than when the same information is presented as probabilities or percentages, presumably because the ancestors of present-day humans lived in relatively small tribes (usually with fewer than 150 people) where frequency information was more readily available.
Evolutionary psychology is primarily interested in finding commonalities between people, or basic human psychological nature. From an evolutionary perspective, the fact that people have fundamental differences in personality traits initially presents something of a puzzle. (Note: The field of behavioral genetics is concerned with statistically partitioning differences between people into genetic and environmental sources of variance. However, understanding the concept of heritability can be tricky – heritability refers only to the differences between people, never the degree to which the traits of an individual are due to environmental or genetic factors, since traits are always a complex interweaving of both.)
Personality traits are conceptualized by evolutionary psychologists as due to normal variation around an optimum, due to frequency-dependent selection (behavioral polymorphisms), or as facultative adaptations. Like variability in height, some personality traits may simply reflect inter-individual variability around a general optimum. Or, personality traits may represent different genetically predisposed "behavioral morphs" – alternate behavioral strategies that depend on the frequency of competing behavioral strategies in the population. For example, if most of the population is generally trusting and gullible, the behavioral morph of being a "cheater" (or, in the extreme case, a sociopath) may be advantageous. Finally, like many other psychological adaptations, personality traits may be facultative – sensitive to typical variations in the social environment, especially during early development. For example, later born children are more likely than first borns to be rebellious, less conscientious and more open to new experiences, which may be advantageous to them given their particular niche in family structure. It is important to note that shared environmental influences do play a role in personality and are not always of less importance than genetic factors. However, shared environmental influences often decrease to near zero after adolescence but do not completely disappear.
According to Steven Pinker, who builds on the work by Noam Chomsky, the universal human ability to learn to talk between the ages of 1 – 4, basically without training, suggests that language acquisition is a distinctly human psychological adaptation (see, in particular, Pinker's "The Language Instinct"). Pinker and Bloom (1990) argue that language as a mental faculty shares many likenesses with the complex organs of the body which suggests that, like these organs, language has evolved as an adaptation, since this is the only known mechanism by which such complex organs can develop.
Pinker follows Chomsky in arguing that the fact that children can learn any human language with no explicit instruction suggests that language, including most of grammar, is basically innate and that it only needs to be activated by interaction. Chomsky himself does not believe language to have evolved as an adaptation, but suggests that it likely evolved as a byproduct of some other adaptation, a so-called spandrel. But Pinker and Bloom argue that the organic nature of language strongly suggests that it has an adaptational origin.
Evolutionary psychologists hold that the FOXP2 gene may well be associated with the evolution of human language. In the 1980s, psycholinguist Myrna Gopnik identified a dominant gene that causes language impairment in the KE family of Britain. This gene turned out to be a mutation of the FOXP2 gene. Humans have a unique allele of this gene, which has otherwise been closely conserved through most of mammalian evolutionary history. This unique allele seems to have first appeared between 100 and 200 thousand years ago, and it is now all but universal in humans. However, the once-popular idea that FOXP2 is a 'grammar gene' or that it triggered the emergence of language in "Homo sapiens" is now widely discredited.
Currently several competing theories about the evolutionary origin of language coexist, none of them having achieved a general consensus. Researchers of language acquisition in primates and humans such as Michael Tomasello and Talmy Givón, argue that the innatist framework has understated the role of imitation in learning and that it is not at all necessary to posit the existence of an innate grammar module to explain human language acquisition. Tomasello argues that studies of how children and primates actually acquire communicative skills suggests that humans learn complex behavior through experience, so that instead of a module specifically dedicated to language acquisition, language is acquired by the same cognitive mechanisms that are used to acquire all other kinds of socially transmitted behavior.
On the issue of whether language is best seen as having evolved as an adaptation or as a spandrel, evolutionary biologist W. Tecumseh Fitch, following Stephen J. Gould, argues that it is unwarranted to assume that every aspect of language is an adaptation, or that language as a whole is an adaptation. He criticizes some strands of evolutionary psychology for suggesting a pan-adaptionist view of evolution, and dismisses Pinker and Bloom's question of whether "Language has evolved as an adaptation" as being misleading. He argues instead that from a biological viewpoint the evolutionary origins of language is best conceptualized as being the probable result of a convergence of many separate adaptations into a complex system. A similar argument is made by Terrence Deacon who in "The Symbolic Species" argues that the different features of language have co-evolved with the evolution of the mind and that the ability to use symbolic communication is integrated in all other cognitive processes.
If the theory that language could have evolved as a single adaptation is accepted, the question becomes which of its many functions has been the basis of adaptation. Several evolutionary hypotheses have been posited: that language evolved for the purpose of social grooming, that it evolved as a way to show mating potential or that it evolved to form social contracts. Evolutionary psychologists recognize that these theories are all speculative and that much more evidence is required to understand how language might have been selectively adapted.
Given that sexual reproduction is the means by which genes are propagated into future generations, sexual selection plays a large role in human evolution. Human mating, then, is of interest to evolutionary psychologists who aim to investigate evolved mechanisms to attract and secure mates. Several lines of research have stemmed from this interest, such as studies of mate selection mate poaching, mate retention, mating preferences and conflict between the sexes.
In 1972 Robert Trivers published an influential paper on sex differences that is now referred to as parental investment theory. The size differences of gametes (anisogamy) is the fundamental, defining difference between males (small gametes – sperm) and females (large gametes – ova). Trivers noted that anisogamy typically results in different levels of parental investment between the sexes, with females initially investing more. Trivers proposed that this difference in parental investment leads to the sexual selection of different reproductive strategies between the sexes and to sexual conflict. For example, he suggested that the sex that invests less in offspring will generally compete for access to the higher-investing sex to increase their inclusive fitness (also see Bateman's principle). Trivers posited that differential parental investment led to the evolution of sexual dimorphisms in mate choice, intra- and inter- sexual reproductive competition, and courtship displays. In mammals, including humans, females make a much larger parental investment than males (i.e. gestation followed by childbirth and lactation). Parental investment theory is a branch of life history theory.
Buss and Schmitt's (1993) "Sexual Strategies Theory" proposed that, due to differential parental investment, humans have evolved sexually dimorphic adaptations related to "sexual accessibility, fertility assessment, commitment seeking and avoidance, immediate and enduring resource procurement, paternity certainty, assessment of mate value, and parental investment." Their "Strategic Interference Theory" suggested that conflict between the sexes occurs when the preferred reproductive strategies of one sex interfere with those of the other sex, resulting in the activation of emotional responses such as anger or jealousy.
Women are generally more selective when choosing mates, especially under long term mating conditions. However, under some circumstances, short term mating can provide benefits to women as well, such as fertility insurance, trading up to better genes, reducing risk of inbreeding, and insurance protection of her offspring.
Due to male paternity insecurity, sex differences have been found in the domains of sexual jealousy. Females generally react more adversely to emotional infidelity and males will react more to sexual infidelity. This particular pattern is predicted because the costs involved in mating for each sex are distinct. Women, on average, should prefer a mate who can offer resources (e.g., financial, commitment), thus, a woman risks losing such resources with a mate who commits emotional infidelity. Men, on the other hand, are never certain of the genetic paternity of their children because they do not bear the offspring themselves ("paternity insecurity"). This suggests that for men sexual infidelity would generally be more aversive than emotional infidelity because investing resources in another man's offspring does not lead to propagation of their own genes.
Another interesting line of research is that which examines women's mate preferences across the ovulatory cycle. The theoretical underpinning of this research is that ancestral women would have evolved mechanisms to select mates with certain traits depending on their hormonal status. Known as the ovulatory shift hypothesis, the theory posits that, during the ovulatory phase of a woman's cycle (approximately days 10–15 of a woman's cycle), a woman who mated with a male with high genetic quality would have been more likely, on average, to produce and bear a healthy offspring than a woman who mated with a male with low genetic quality. These putative preferences are predicted to be especially apparent for short-term mating domains because a potential male mate would only be offering genes to a potential offspring. This hypothesis allows researchers to examine whether women select mates who have characteristics that indicate high genetic quality during the high fertility phase of their ovulatory cycles. Indeed, studies have shown that women's preferences vary across the ovulatory cycle. In particular, Haselton and Miller (2006) showed that highly fertile women prefer creative but poor men as short-term mates. Creativity may be a proxy for good genes. Research by Gangestad et al. (2004) indicates that highly fertile women prefer men who display social presence and intrasexual competition; these traits may act as cues that would help women predict which men may have, or would be able to acquire, resources.
Reproduction is always costly for women, and can also be for men. Individuals are limited in the degree to which they can devote time and resources to producing and raising their young, and such expenditure may also be detrimental to their future condition, survival and further reproductive output.
Parental investment is any parental expenditure (time, energy etc.) that benefits one offspring at a cost to parents' ability to invest in other components of fitness (Clutton-Brock 1991: 9; Trivers 1972). Components of fitness (Beatty 1992) include the well-being of existing offspring, parents' future reproduction, and inclusive fitness through aid to kin (Hamilton, 1964). Parental investment theory is a branch of life history theory.
Robert Trivers' theory of parental investment predicts that the sex making the largest investment in lactation, nurturing and protecting offspring will be more discriminating in mating and that the sex that invests less in offspring will compete for access to the higher investing sex (see Bateman's principle). Sex differences in parental effort are important in determining the strength of sexual selection.
The benefits of parental investment to the offspring are large and are associated with the effects on condition, growth, survival and ultimately, on reproductive success of the offspring. However, these benefits can come at the cost of parent's ability to reproduce in the future e.g. through the increased risk of injury when defending offspring against predators, the loss of mating opportunities whilst rearing offspring and an increase in the time to the next reproduction. Overall, parents are selected to maximize the difference between the benefits and the costs, and parental care will likely evolve when the benefits exceed the costs.
The Cinderella effect is an alleged high incidence of stepchildren being physically, emotionally or sexually abused, neglected, murdered, or otherwise mistreated at the hands of their stepparents at significantly higher rates than their genetic counterparts. It takes its name from the fairy tale character Cinderella, who in the story was cruelly mistreated by her stepmother and stepsisters. Daly and Wilson (1996) noted: "Evolutionary thinking led to the discovery of the most important risk factor for child homicide – the presence of a stepparent. Parental efforts and investments are valuable resources, and selection favors those parental psyches that allocate effort effectively to promote fitness. The adaptive problems that challenge parental decision making include both the accurate identification of one's offspring and the allocation of one's resources among them with sensitivity to their needs and abilities to convert parental investment into fitness increments…. Stepchildren were seldom or never so valuable to one's expected fitness as one's own offspring would be, and those parental psyches that were easily parasitized by just any appealing youngster must always have incurred a selective disadvantage"(Daly & Wilson, 1996, pp. 64–65). However, they note that not all stepparents will "want" to abuse their partner's children, or that genetic parenthood is any insurance against abuse. They see step parental care as primarily "mating effort" towards the genetic parent.
Inclusive fitness is the sum of an organism's classical fitness (how many of its own offspring it produces and supports) and the number of equivalents of its own offspring it can add to the population by supporting others. The first component is called classical fitness by Hamilton (1964).
From the gene's point of view, evolutionary success ultimately depends on leaving behind the maximum number of copies of itself in the population. Until 1964, it was generally believed that genes only achieved this by causing the individual to leave the maximum number of viable offspring. However, in 1964 W. D. Hamilton proved mathematically that, because close relatives of an organism share some identical genes, a gene can also increase its evolutionary success by promoting the reproduction and survival of these related or otherwise similar individuals. Hamilton concluded that this leads natural selection to favor organisms that would behave in ways that maximize their inclusive fitness. It is also true that natural selection favors behavior that maximizes personal fitness.
Hamilton's rule describes mathematically whether or not a gene for altruistic behavior will spread in a population:
where
The concept serves to explain how natural selection can perpetuate altruism. If there is an "altruism gene" (or complex of genes) that influences an organism's behavior to be helpful and protective of relatives and their offspring, this behavior also increases the proportion of the altruism gene in the population, because relatives are likely to share genes with the altruist due to common descent. Altruists may also have some way to recognize altruistic behavior in unrelated individuals and be inclined to support them. As Dawkins points out in "The Selfish Gene" (Chapter 6) and "The Extended Phenotype", this must be distinguished from the green-beard effect.
Although it is generally true that humans tend to be more altruistic toward their kin than toward non-kin, the relevant proximate mechanisms that mediate this cooperation have been debated (see kin recognition), with some arguing that kin status is determined primarily via social and cultural factors (such as co-residence, maternal association of sibs, etc.), while others have argued that kin recognition can also be mediated by biological factors such as facial resemblance and immunogenetic similarity of the major histocompatibility complex (MHC). For a discussion of the interaction of these social and biological kin recognition factors see Lieberman, Tooby, and Cosmides (2007) (PDF).
Whatever the proximate mechanisms of kin recognition there is substantial evidence that humans act generally more altruistically to close genetic kin compared to genetic non-kin.
Although interactions with non-kin are generally less altruistic compared to those with kin, cooperation can be maintained with non-kin via mutually beneficial reciprocity as was proposed by Robert Trivers. If there are repeated encounters between the same two players in an evolutionary game in which each of them can choose either to "cooperate" or "defect," then a strategy of mutual cooperation may be favored even if it pays each player, in the short term, to defect when the other cooperates. Direct reciprocity can lead to the evolution of cooperation only if the probability, w, of another encounter between the same two individuals exceeds the cost-to-benefit ratio of the altruistic act:
Reciprocity can also be indirect if information about previous interactions is shared. Reputation allows evolution of cooperation by indirect reciprocity. Natural selection favors strategies that base the decision to help on the reputation of the recipient: studies show that people who are more helpful are more likely to receive help. The calculations of indirect reciprocity are complicated and only a tiny fraction of this universe has been uncovered, but again a simple rule has emerged. Indirect reciprocity can only promote cooperation if the probability, q, of knowing someone's reputation exceeds the cost-to-benefit ratio of the altruistic act:
One important problem with this explanation is that individuals may be able to evolve the capacity to obscure their reputation, reducing the probability, q, that it will be known.
Trivers argues that friendship and various social emotions evolved in order to manage reciprocity. Liking and disliking, he says, evolved to help present day humans' ancestors form coalitions with others who reciprocated and to exclude those who did not reciprocate. Moral indignation may have evolved to prevent one's altruism from being exploited by cheaters, and gratitude may have motivated present day humans' ancestors to reciprocate appropriately after benefiting from others' altruism. Likewise, present day humans feel guilty when they fail to reciprocate. These social motivations match what evolutionary psychologists expect to see in adaptations that evolved to maximize the benefits and minimize the drawbacks of reciprocity.
Evolutionary psychologists say that humans have psychological adaptations that evolved specifically to help us identify nonreciprocators, commonly referred to as "cheaters." In 1993, Robert Frank and his associates found that participants in a prisoner's dilemma scenario were often able to predict whether their partners would "cheat," based on a half-hour of unstructured social interaction. In a 1996 experiment, for example, Linda Mealey and her colleagues found that people were better at remembering the faces of people when those faces were associated with stories about those individuals cheating (such as embezzling money from a church).
Humans may have an evolved set of psychological adaptations that predispose them to be more cooperative than otherwise would be expected with members of their tribal in-group, and, more nasty to members of tribal out groups. These adaptations may have been a consequence of tribal warfare. Humans may also have predispositions for "altruistic punishment" – to punish in-group members who violate in-group rules, even when this altruistic behavior cannot be justified in terms of helping those you are related to (kin selection), cooperating with those who you will interact with again (direct reciprocity), or cooperating to better your reputation with others (indirect reciprocity).
Though evolutionary psychology has traditionally focused on individual-level behaviors, determined by species-typical psychological adaptations, considerable work has been done on how these adaptations shape and, ultimately govern, culture (Tooby and Cosmides, 1989). Tooby and Cosmides (1989) argued that the mind consists of many domain-specific psychological adaptations, some of which may constrain what cultural material is learned or taught. As opposed to a domain-general cultural acquisition program, where an individual passively receives culturally-transmitted material from the group, Tooby and Cosmides (1989), among others, argue that: "the psyche evolved to generate adaptive rather than repetitive behavior, and hence critically analyzes the behavior of those surrounding it in highly structured and patterned ways, to be used as a rich (but by no means the only) source of information out of which to construct a 'private culture' or individually tailored adaptive system; in consequence, this system may or may not mirror the behavior of others in any given respect." (Tooby and Cosmides 1989).
According to Paul Baltes, the benefits granted by evolutionary selection decrease with age. Natural selection has not eliminated many harmful conditions and nonadaptive characteristics that appear among older adults, such as Alzheimer disease. If it were a disease that killed 20-year-olds instead of 70-year-olds this may have been a disease that natural selection could have eliminated ages ago. Thus, unaided by evolutionary pressures against nonadaptive conditions, modern humans suffer the aches, pains, and infirmities of aging and as the benefits of evolutionary selection decrease with age, the need for modern technological mediums against non-adaptive conditions increases.
As humans are a highly social species, there are many adaptive problems associated with navigating the social world (e.g., maintaining allies, managing status hierarchies, interacting with outgroup members, coordinating social activities, collective decision-making). Researchers in the emerging field of evolutionary social psychology have made many discoveries pertaining to topics traditionally studied by social psychologists, including person perception, social cognition, attitudes, altruism, emotions, group dynamics, leadership, motivation, prejudice, intergroup relations, and cross-cultural differences.
When endeavouring to solve a problem humans at an early age show determination while chimpanzees have no comparable facial expression. Researchers suspect the human determined expression evolved because when a human is determinedly working on a problem other people will frequently help.
Adaptationist hypotheses regarding the etiology of psychological disorders are often based on analogies between physiological and psychological dysfunctions, as noted in the table below. Prominent theorists and evolutionary psychiatrists include Michael T. McGuire, Anthony Stevens, and Randolph M. Nesse. They, and others, suggest that mental disorders are due to the interactive effects of both nature and nurture, and often have multiple contributing causes.
Evolutionary psychologists have suggested that schizophrenia and bipolar disorder may reflect a side-effect of genes with fitness benefits, such as increased creativity. (Some individuals with bipolar disorder are especially creative during their manic phases and the close relatives of people with schizophrenia have been found to be more likely to have creative professions.) A 1994 report by the American Psychiatry Association found that people suffered from schizophrenia at roughly the same rate in Western and non-Western cultures, and in industrialized and pastoral societies, suggesting that schizophrenia is not a disease of civilization nor an arbitrary social invention. Sociopathy may represent an evolutionarily stable strategy, by which a small number of people who cheat on social contracts benefit in a society consisting mostly of non-sociopaths. Mild depression may be an adaptive response to withdraw from, and re-evaluate, situations that have led to disadvantageous outcomes (the "analytical rumination hypothesis") (see Evolutionary approaches to depression).
Some of these speculations have yet to be developed into fully testable hypotheses, and a great deal of research is required to confirm their validity.
Evolutionary psychology has been applied to explain criminal or otherwise immoral behavior as being adaptive or related to adaptive behaviors. Males are generally more aggressive than females, who are more selective of their partners because of the far greater effort they have to contribute to pregnancy and child-rearing. Males being more aggressive is hypothesized to stem from the more intense reproductive competition faced by them. Males of low status may be especially vulnerable to being childless. It may have been evolutionary advantageous to engage in highly risky and violently aggressive behavior to increase their status and therefore reproductive success. This may explain why males are generally involved in more crimes, and why low status and being unmarried is associated with criminality. Furthermore, competition over females is argued to have been particularly intensive in late adolescence and young adulthood, which is theorized to explain why crime rates are particularly high during this period.
Many conflicts that result in harm and death involve status, reputation, and seemingly trivial insults. Steven Pinker in his book "The Blank Slate" argues that in non-state societies without a police it was very important to have a credible deterrence against aggression. Therefore, it was important to be perceived as having a credible reputation for retaliation, resulting in humans to develop instincts for revenge as well as for protecting reputation ("honor"). Pinker argues that the development of the state and the police have dramatically reduced the level of violence compared to the ancestral environment. Whenever the state breaks down, which can be very locally such as in poor areas of a city, humans again organize in groups for protection and aggression and concepts such as violent revenge and protecting honor again become extremely important.
Rape is theorized to be a reproductive strategy that facilitates the propagation of the rapist's progeny. Such a strategy may be adopted by men who otherwise are unlikely to be appealing to women and therefore cannot form legitimate relationships, or by high status men on socially vulnerable women who are unlikely to retaliate to increase their reproductive success even further. The sociobiological theories of rape are highly controversial, as traditional theories typically do not consider rape to be a behavioral adaptation, and objections to this theory are made on ethical, religious, political, as well as scientific grounds.
Adaptationist perspectives on religious belief suggest that, like all behavior, religious behaviors are a product of the human brain. As with all other organ functions, cognition's functional structure has been argued to have a genetic foundation, and is therefore subject to the effects of natural selection and sexual selection. Like other organs and tissues, this functional structure should be universally shared amongst humans and should have solved important problems of survival and reproduction in ancestral environments. However, evolutionary psychologists remain divided on whether religious belief is more likely a consequence of evolved psychological adaptations, or a byproduct of other cognitive adaptations.
Coalitional psychology is an approach to explain political behaviors between different coalitions and the conditionality of these behaviors in evolutionary psychological perspective. This approach assumes that since human beings appeared on the earth, they have evolved to live in groups instead of living as individuals to achieve benefits such as more mating opportunities and increased status. Human beings thus naturally think and act in a way that manages and negotiates group dynamics.
Coalitional psychology offers falsifiable ex ante prediction by positing five hypotheses on how these psychological adaptations operate:
Critics of evolutionary psychology accuse it of promoting genetic determinism, panadaptionism (the idea that all behaviors and anatomical features are adaptations), unfalsifiable hypotheses, distal or ultimate explanations of behavior when proximate explanations are superior, and malevolent political or moral ideas.
Critics have argued that evolutionary psychology might be used to justify existing social hierarchies and reactionary policies. It has also been suggested by critics that evolutionary psychologists' theories and interpretations of empirical data rely heavily on ideological assumptions about race and gender.
In response to such criticism, evolutionary psychologists often caution against committing the naturalistic fallacy – the assumption that "what is natural" is necessarily a moral good. However, their caution against committing the naturalistic fallacy has been criticized as means to stifle legitimate ethical discussions.
Some criticisms of evolutionary psychology point at contradictions between different aspects of adaptive scenarios posited by evolutionary psychology. One example is the evolutionary psychology model of extended social groups selecting for modern human brains, a contradiction being that the synaptic function of modern human brains require high amounts of many specific essential nutrients so that such a transition to higher requirements of the same essential nutrients being shared by all individuals in a population would decrease the possibility of forming large groups due to bottleneck foods with rare essential nutrients capping group sizes. It is mentioned that some insects have societies with different ranks for each individual and that monkeys remain socially functioning after removal of most of the brain as additional arguments against big brains promoting social networking. The model of males as both providers and protectors is criticized for the impossibility of being in two places at once, the male cannot both protect his family at home and be out hunting at the same time. In the case of the claim that a provider male could buy protection service for his family from other males by bartering food that he had hunted, critics point at the fact that the most valuable food (the food that contained the rarest essential nutrients) would be different in different ecologies and as such vegetable in some geographical areas and animal in others, making it impossible for hunting styles relying on physical strength or risk taking to be universally of similar value in bartered food and instead making it inevitable that in some parts of Africa, food gathered with no need for major physical strength would be the most valuable to barter for protection. A contradiction between evolutionary psychology's claim of men needing to be more sexually visual than women for fast speed of assessing women's fertility than women needed to be able to assess the male's genes and its claim of male sexual jealousy guarding against infidelity is also pointed at, as it would be pointless for a male to be fast to assess female fertility if he needed to assess the risk of there being a jealous male mate and in that case his chances of defeating him before mating anyway (pointlessness of assessing one necessary condition faster than another necessary condition can possibly be assessed).
Evolutionary psychology has been entangled in the larger philosophical and social science controversies related to the debate on nature versus nurture. Evolutionary psychologists typically contrast evolutionary psychology with what they call the standard social science model (SSSM). They characterize the SSSM as the "blank slate", "relativist", "social constructionist", and "cultural determinist" perspective that they say dominated the social sciences throughout the 20th century and assumed that the mind was shaped almost entirely by culture.
Critics have argued that evolutionary psychologists created a false dichotomy between their own view and the caricature of the SSSM. Other critics regard the SSSM as a rhetorical device or a straw man and suggest that the scientists whom evolutionary psychologists associate with the SSSM did not believe that the mind was a blank state devoid of any natural predispositions.
Some critics view evolutionary psychology as a form of genetic reductionism and genetic determinism, a common critique being that evolutionary psychology does not address the complexity of individual development and experience and fails to explain the influence of genes on behavior in individual cases. Evolutionary psychologists respond that they are working within a nature-nurture interactionist framework that acknowledges that many psychological adaptations are facultative (sensitive to environmental variations during individual development). The discipline is generally not focused on proximate analyses of behavior, but rather its focus is on the study of distal/ultimate causality (the evolution of psychological adaptations). The field of behavioral genetics is focused on the study of the proximate influence of genes on behavior.
A frequent critique of the discipline is that the hypotheses of evolutionary psychology are frequently arbitrary and difficult or impossible to adequately test, thus questioning its status as an actual scientific discipline, for example because many current traits probably evolved to serve different functions than they do now. Thus because there are a potentially infinite number of alternative explanations for why a trait evolved, critics contend that it is impossible to determine the exact explanation. While evolutionary psychology hypotheses are difficult to test, evolutionary psychologists assert that it is not impossible. Part of the critique of the scientific base of evolutionary psychology includes a critique of the concept of the Environment of Evolutionary Adaptation (EEA). Some critics have argued that researchers know so little about the environment in which "Homo sapiens" evolved that explaining specific traits as an adaption to that environment becomes highly speculative. Evolutionary psychologists respond that they do know many things about this environment, including the facts that present day humans' ancestors were hunter-gatherers, that they generally lived in small tribes, etc. Edward Hagen argues that the human past environments were not radically different in the same sense as the Carboniferous or Jurassic periods and that the animal and plant taxa of the era were similar to those of the modern world, as was the geology and ecology. Hagen argues that few would deny that other organs evolved in the EEA (for example, lungs evolving in an oxygen rich atmopshere) yet critics question whether or not the brain's EEA is truly knowable, which he argues constitutes selective scepticism. Hagen also argues that most evolutionary psychology research is based on the fact that females can get pregnant and males cannot, which Hagen observes was also true in the EEA.
John Alcock describes this as the "No Time Machine Argument", as critics are arguing that since it is not possible to travel back in time to the EEA, then it cannot be determined what was going on there and thus what was adaptive. Alcock argues that present day evidence allows researchers to be reasonably confident about the conditions of the EEA and that the fact that so many human behaviours are adaptive in the "current" environment is evidence that the ancestral environment of humans had much in common with the present one, as these behaviours would have evolved in the ancestral environment. Thus Alcock concludes that researchers can make predictions on the adaptive value of traits. Similarly, Dominic Murphy argues that alternative explanations cannot just be forwarded but instead need their own evidence and predictions - if one explanation makes predictions that the others cannot, it is reasonable to have confidence in that explanation. In addition, Murphy argues that other historical sciences also make predictions about modern phenomena to come up with explanations about past phenomena, for example cosmologists look for evidence for what we would expect to see in the modern day if the Big Bang was true, while geologists make predictions about modern phenomena to determine if an asteroid wiped out the dinosaurs. Murphy argues that if other historical disciplines can conduct tests without a time machine, then the onus is on the critics to show why evolutionary psychology is untestable if other historical disciplines are not, as "methods should be judged across the board, not singled out for ridicule in one context."
Evolutionary psychologists generally presume that, like the body, the mind is made up of many evolved modular adaptations, although there is some disagreement within the discipline regarding the degree of general plasticity, or "generality," of some modules. It has been suggested that modularity evolves because, compared to non-modular networks, it would have conferred an advantage in terms of fitness and because connection costs are lower.
In contrast, some academics argue that it is unnecessary to posit the existence of highly domain specific modules, and, suggest that the neural anatomy of the brain supports a model based on more domain general faculties and processes. Moreover, empirical support for the domain-specific theory stems almost entirely from performance on variations of the Wason selection task which is extremely limited in scope as it only tests one subtype of deductive reasoning.
Cecilia Heyes has argued that the picture presented by some evolutionary psychology of the human mind as a collection of cognitive instinctsorgans of thought shaped by genetic evolution over very long time periodsdoes not fit research results. She posits instead that humans have cognitive gadgets"special-purpose organs of thought" built in the course of development through social interaction.
Evolutionary psychologists have addressed many of their critics (see, for example, books by Segerstråle (2000), "Defenders of the Truth: The Battle for Science in the Sociobiology Debate and Beyond," Barkow (2005), "Missing the Revolution: Darwinism for Social Scientists," and Alcock (2001), "The Triumph of Sociobiology"). Among their rebuttals are that some criticisms are straw men, are based on an incorrect nature versus nurture dichotomy, are based on misunderstandings of the discipline, etc. Robert Kurzban suggested that "...critics of the field, when they err, are not slightly missing the mark. Their confusion is deep and profound. It’s not like they are marksmen who can’t quite hit the center of the target; they’re holding the gun backwards." | https://en.wikipedia.org/wiki?curid=9703 |
Languages of Europe
Most languages of Europe belong to the Indo-European language family.
Out of a total European population of 744 million as of 2018, some 94% are native speakers of an Indo-European language; within Indo-European, the three largest phyla are "Romance", "Germanic", and "Slavic" with more than 200 million speakers each, between them accounting for close to 90% of Europeans. Smaller phyla of Indo-European found in Europe include Hellenic (Greek, 13 million), Baltic ( 7 million), Albanian ( 5 million), Indo-Aryan (Romani, 1.5 million), and Celtic (including Welsh, 1 million).
Of the approximately 45 million Europeans speaking non-Indo-European languages, most speak languages within either the "Uralic" or "Turkic" families. Still smaller groups (such as Basque and various languages of the Caucasus) account for less than 1% of the European population between them. Immigration has added sizeable communities of speakers of African and Asian languages, amounting to about 4% of the population, with Arabic being the most widely spoken of them.
Five languages have more than 50 million native speakers in Europe: French, Italian, German, English, and Russian. While Russian has the largest number of native speakers (more than 100 million in Europe), English has the largest number of speakers in total, including some 200 million speakers of English as a second or foreign language. ("See English language in Europe".)
The Indo-European language family is descended from Proto-Indo-European, which is believed to have been spoken thousands of years ago. Early speakers of Indo-European daughter languages most likely expanded into Europe with the incipient Bronze Age, around 4,000 years ago (Bell-Beaker culture).
Roughly 215 million Europeans (primarily in Western and Southern Europe) are native speakers of Romance languages, the largest groups including
French ( 72 million),
Italian ( 65 million),
Spanish (Castilian) ( 40 million),
Romanian ( 24 million),
Portuguese ( 10 million),
Catalan ( 9 million),
Sicilian ( 5 million, also subsumed under Italian),
Venetian language ( 4 million),
Galician ( 2 million),
Sardinian ( 1 million),
Occitan ( 500,000), besides numerous smaller communities.
The Romance languages are descended from varieties of Vulgar Latin spoken in the various parts of the Roman Empire in Late Antiquity. Latin was itself part of the (otherwise extinct) Italic branch of Indo-European.
Romance languages are divided phylogenetically into "Italo-Western", "Eastern Romance" (including Romanian) and "Sardinian". The Romance-speaking area of Europe is occasionally referred to as "Latin Europe".
We can further break down Italo-Western into the "Italo-Dalmatian languages" (sometimes grouped with Eastern Romance), including the Tuscan-derived Italian and numerous local Romance lects in Italy as well as Dalmatian, and the "Western Romance languages". The Western Romance languages in turn separate into the Gallo-Romance languages, including French and its varieties (Langues d'oïl), the Rhaeto-Romance languages and the Gallo-Italic languages; the Occitano-Romance languages, grouped with either Gallo-Romance or East Iberian, including Occitan, Catalan and Aragonese; and finally the West Iberian languages (Spanish-Portuguese), including the Astur-Leonese languages, Galician-Portuguese, and Castilian.
The Germanic languages make up the predominant language family in Western, Northern and Central Europe. An estimated 210 million Europeans are native speakers of Germanic languages, the largest groups being German ( 95 million), English ( 70 million), Dutch ( 24 million), Swedish ( 10 million), Danish ( 6 million), and Norwegian ( 5 million).
There are two extant major sub-divisions: "West Germanic" and "North Germanic". A third group, East Germanic, is now extinct; the only known surviving East Germanic texts are written in the Gothic language. West Germanic is divided into Anglo-Frisian (including English), Low German, and Low Franconian (including Dutch) and High German (including Standard German).
German is spoken throughout Germany, Austria, Liechtenstein, much of Switzerland (including the northeast areas bordering on Germany and Austria), northern Italy (South Tyrol), Luxembourg, and the East Cantons of Belgium.
There are several groups of German dialects:
Low German (including Low Saxon) is spoken in various regions throughout Northern Germany and the northern and eastern parts of the Netherlands. It is an official language in Germany. It may be separated into Low Saxon (West Low German) and East Low German.
Dutch is spoken throughout the Netherlands, the northern half of Belgium, as well as the Nord-Pas de Calais region of France, and around Düsseldorf in Germany. In Belgian and French contexts, Dutch is sometimes referred to as Flemish. Dutch dialects are varied and cut across national borders.
The Anglo-Frisian language family is now mostly represented by English (Anglic), descended from the Old English language spoken by the Anglo-Saxons:
The Frisian languages are spoken by about 500,000 Frisians, who live on the southern coast of the North Sea in the Netherlands and Germany. These languages include West Frisian, Saterlandic, and North Frisian.
The "North Germanic languages" are spoken in Scandinavian countries and include Danish (Denmark), Norwegian (Norway), Swedish (Sweden and parts of Finland), or Elfdalian (in a small part of central Sweden), Faroese (Faroe Islands), and Icelandic (Iceland).
English has a long history of contact with Scandinavian languages, given the immigration of Scandinavians early in the history of Britain, and shares various features with the Scandinavian languages.
Slavic languages are spoken in large areas of Southern, Central and Eastern Europe. An estimated 250 million Europeans are native speakers of Slavic languages, the largest groups being
Russian ( 110 million in European Russia and adjacent parts of Eastern Europe, Russian forming the largest linguistic community in Europe),
Polish ( 55 million),
Ukrainian ( 40 million),
Serbo-Croatian ( 21 million),
Czech ( 11 million),
Bulgarian ( 9 million),
Slovak ( 5 million)
Belarusian and Slovene ( 3 million each)
and Macedonian ( 2 million).
Phylogenetically, Slavic is divided into three subgroups:
Uralic is native to northern Eurasia. Finno-Ugric groups the Uralic languages other than Samoyedic.
Finnic languages include Finnish ( 5 million) and Estonian ( 1 million). The Sami languages ( 30,000) are closely related to Finnic.
The Ugric languages are represented in Europe with the Hungarian language ( 13 million), historically introduced with the Hungarian conquest of the Carpathian Basin of the 9th century.
The Samoyedic Nenets language is spoken in Nenets Autonomous Okrug of Russia, located in the far northeastern corner of Europe (as delimited by the Ural Mountains).
In the Middle Ages the two most important defining elements of Europe were "Christianitas" and "Latinitas".
The earliest dictionaries were glossaries: more or less structured lists of lexical pairs (in alphabetical order or according to conceptual fields). The Latin-German (Latin-Bavarian) "Abrogans" was among the first. A new wave of lexicography can be seen from the late 15th century onwards (after the introduction of the printing press, with the growing interest in standardisation of languages).
The concept of the nation state began to emerge in the early modern period. Nations adopted particular dialects as their national language. This, together with improved communications, led to official efforts to standardise the national language, and a number of language academies were established: 1582 "Accademia della Crusca" in Florence, 1617 "Fruchtbringende Gesellschaft" in Weimar, 1635 "Académie française" in Paris, 1713 "Real Academia Española" in Madrid. Language became increasingly linked to nation as opposed to culture, and was also used to promote religious and ethnic identity: e.g. different Bible translations in the same language for Catholics and Protestants.
The first languages whose standardisation was promoted included Italian ("questione della lingua": Modern Tuscan/Florentine vs. Old Tuscan/Florentine vs. Venetian → Modern Florentine + archaic Tuscan + Upper Italian), French (the standard is based on Parisian), English (the standard is based on the London dialect) and (High) German (based on the dialects of the chancellery of Meissen in Saxony, Middle German, and the chancellery of Prague in Bohemia ("Common German")). But several other nations also began to develop a standard variety in the 16th century.
Europe has had a number of languages that were considered linguae francae over some ranges for some periods according to some historians. Typically in the rise of a national language the new language becomes a lingua franca to peoples in the range of the future nation until the consolidation and unification phases. If the nation becomes internationally influential, its language may become a lingua franca among nations that speak their own national languages. Europe has had no lingua franca ranging over its entire territory spoken by all or most of its populations during any historical period. Some linguae francae of past and present over some of its regions for some of its populations are:
Historical attitudes towards linguistic diversity are illustrated by two French laws: the Ordonnance de Villers-Cotterêts (1539), which said that every document in France should be written in French (neither in Latin nor in Occitan) and the Loi Toubon (1994), which aimed to eliminate anglicisms from official documents. States and populations within a state have often resorted to war to settle their differences. There have been attempts to prevent such hostilities: two such initiatives were promoted by the Council of Europe, founded in 1949, which affirms the right of minority language speakers to use their language fully and freely. The Council of Europe is committed to protecting linguistic diversity. Currently all European countries except France, Andorra and Turkey have signed the Framework Convention for the Protection of National Minorities, while Greece, Iceland and Luxembourg have signed it, but have not ratified it; this framework entered into force in 1998. Another European treaty, the European Charter for Regional or Minority Languages, was adopted in 1992 under the auspices of the Council of Europe: it entered into force in 1998, and while it is legally binding for 24 countries, France, Iceland, Italy, North Macedonia, Moldova and Russia have chosen to sign without ratifying the convention.
The main scripts used in Europe today are the Latin and Cyrillic.
The Greek alphabet was derived from the Phoenician alphabet, and Latin was derived from the Greek via the Old Italic alphabet. In the Early Middle Ages, Ogham was used in Ireland and runes (derived the Old Italic script) in Scandinavia. Both were replaced in general use by the Latin alphabet by the Late Middle Ages. The Cyrillic script was derived from the Greek with the first texts appearing around 940 AD.
Around 1900 there were mainly two typeface variants of the Latin alphabet used in Europe: Antiqua and Fraktur. Fraktur was used most for German, Estonian, Latvian, Norwegian and Danish whereas Antiqua was used for Italian, Spanish, French, Polish, Portuguese, English, Romanian, Swedish and Finnish. The Fraktur variant was banned by Hitler in 1941, having been described as "Schwabacher Jewish letters". Other scripts have historically been in use in Europe, including Phoenician, from which modern Latin letters descend, Ancient Egyptian hieroglyphs on Egyptian artefacts traded during Antiquity various runic systems used in Northern Europe preceding Christianisation, and Arabic during the era of the Ottoman Empire.
Hungarian rovás was used by the Hungarian people in the early Middle Ages, but it was gradually replaced with the Latin-based Hungarian alphabet when Hungary became a kingdom, though it was revived in the 20th century and has certain marginal, but growing area of usage since then.
The European Union (as of 2016) had 28 member states accounting for a population of 510 million, or about 69% of the population of Europe.
The European Union has designated by agreement with the member states 24 languages as "official and working": Bulgarian, Croatian, Czech, Danish, Dutch, English, Estonian, Finnish, French, German, Greek, Hungarian, Irish, Italian, Latvian, Lithuanian, Maltese, Polish, Portuguese, Romanian, Slovak, Slovenian, Spanish and Swedish. This designation provides member states with two "entitlements": the member state may communicate with the EU in any of the designated languages, and view "EU regulations and other legislative documents" in that language.
The European Union and the Council of Europe have been collaborating in education of member populations in languages for "the promotion of plurilingualism" among EU member states. The joint document, "Common European Framework of Reference for Languages: Learning, Teaching, Assessment (CEFR)", is an educational standard defining "the competencies necessary for communication" and related knowledge for the benefit of educators in setting up educational programs.
In a 2005 independent survey requested by the EU's Directorate-General for Education and Culture regarding the extent to which major European languages were spoken in member states. The results were published in a 2006 document, "Europeans and Their Languages", or "Eurobarometer 243". In this study, statistically relevant samples of the population in each country were asked to fill out a survey form concerning the languages that they spoke with sufficient competency "to be able to have a conversation".
The following is a table of European languages. The number of speakers as a first or second language (L1 and L2 speakers) listed are speakers in Europe only; see list of languages by number of native speakers and list of languages by total number of speakers for global estimates on numbers of speakers.
The list is intended to include any language variety with an ISO 639 code. However, it omits sign languages. Because the ISO-639-2 and ISO-639-3 codes have different definitions, this means that some communities of speakers may be listed more than once. For instance, speakers of Austro-Bavarian are listed both under "Bavarian" (ISO-639-3 code "bar") as well as under "German" (ISO-639-2 code "de").
There are various definitions of Europe, which may or may not include all or parts of Turkey, Cyprus, Armenia, Azerbaijan, and Georgia. For convenience, the languages and associated statistics for all five of these countries are grouped together on this page, as they are usually presented at a national, rather than subnational, level.
Recent (post–1945) immigration to Europe introduced substantial communities of speakers of non-European languages.
The largest such communities include Arabic speakers (see Arabs in Europe)
and Turkish speakers (beyond European Turkey and the historical sphere of influence of the Ottoman Empire, see Turks in Europe).
Armenians, Berbers, and Kurds have diaspora communities of 1–2 million each. The various languages of Africa and languages of India form numerous smaller diaspora communities.
Various sign languages are also used in Europe. The most widespread sign language family in Europe is the French Sign Language family, but others include the BANZSL family, the Danish Sign Language family and the Swedish Sign Language family. There are also language isolates, most notably Spanish Sign Language.
The three most used sign languages in Europe, according to Ethnologue, are French Sign Language (spoken in France and Switzerland, and in countries outside Europe), British Sign Language (in the United Kingdom) and German Sign Language (in Germany).
The EU and several other European countries afford legal recognition for various sign languages. | https://en.wikipedia.org/wiki?curid=9705 |
Eindhoven University of Technology
The Eindhoven University of Technology (), abbr. TU/e, is a technical university in the Netherlands, operating in English.
The University has been placed in the top 400 universities in the world by five major ranking tables. The 2019 QS World University Rankings place Eindhoven 99th in the world, 34th in Europe, and 3rd in the Netherlands - TU/e has moved up 59 places in this world ranking since 2012 (in two other main world rankings it is 167th and 51-75th).
TU/e is the Dutch member of the EuroTech Universities Alliance, a strategic partnership of universities of science & technology in Europe: Technical University of Denmark (DTU), École Polytechnique Fédérale de Lausanne (EPFL), École Polytechnique (L’X), The Technion, Eindhoven University of Technology (TU/e), and Technical University of Munich (TUM).
The Eindhoven University of Technology was founded as the "Technische Hogeschool Eindhoven" (THE) on 23 June 1956 by the Dutch government. It was the second institute of its kind in the Netherlands, preceded only by the Delft University of Technology.
Undergraduate education was given in four- or five-year programs until 2002, styled along the lines of the German system of education; graduates of these programs were granted an engineering title and allowed to prefix their name with the title "ir." (an abbreviation of ingenieur; not to be confused with graduates of technical "hogescholen", who were engineers abbreviated "ing."). Starting in 2002, following the entry into force of the Bologna Accords, the university switched to the bachelor/master structure (students graduating in 2002 were given both an old-style engineering title and a new master's title). The undergraduate programs are now split into two parts, a three-year bachelor program and a two-year master program.
On 3 January 2011, the university's strategic vision document for the period up to 2020, the "Strategic Plan 2020", was presented. Despite the economic crisis and the budget cutbacks announced by the Dutch government for the period up to 2014, the university set itself an ambitious strategic vision for the period up to 2020. This vision included establishing a University College to foster both depth, breadth, and societal relevance in engineering education; establishing a combined Graduate School to manage the graduate programs; an increase of the student body by 50 percent; a 50 percent increase in the number of annual PhDs awarded; an increase of knowledge "valorisation" (exploitation by industry and society) to a campus-wide score of 4.2; increasing the international position of the university to within the top-100 universities; and increasing the embedding of the university within the city and the Brainport region by transforming the campus into a high-grade science park with laboratories, housing facilities for 700 students and researchers and supporting facilities. The science park was one of the more costly elements of the plan.
The Eindhoven University of Technology is a public university of the Netherlands. As such its general structure and management is determined by the "Wet op het Hoger Onderwijs en Wetenschappelijk Onderzoek" (English: "Law on Higher Education and Scientific Research"). Between that law and the statutes of the university itself, the management of the university is organized according to the following chart:
The day-to-day running of the university is in the hands of the Executive College (Dutch: "College van Bestuur"). The College provides oversight for the departments, the service organizations and the Innovation Lab, plus the local activities of the Stan Ackermans Institute. The College consists of three people, plus a secretary:
There are two bodies that provide oversight over the Executive College:
Most of the work at the university is done in the departments and the service organizations.
Both for the departments and the service organizations, the staff (and students) are involved with the running of the body. For that reason both types of bodies have advisory councils which have advisory and co-decision authorities.
Over the past two decades, the TU/e has increasingly developed commercial interests and off-campus ties. These include commercial agreements and contracts directly between the university and external companies, but also interests in spinoff companies. In order to manage these kinds of contractual obligations the university started the TU/e Holding B.V. in 1997. The Holding is a limited company, dedicated to the commercial exploitation of scientific knowledge.
The scientific departments (or faculties; Dutch: "faculteiten") are the primary vehicles for teaching and research in the university. They employ the majority of the academic staff, are responsible for teaching and sponsor the research schools and institutions.
The departments also offer Ph.D programs (Dutch: "promotiefase") whereby a qualified master may earn a Ph.D. Unlike in anglo-saxon countries these are not educational programs, however; rather, a person working towards obtaining the Ph.D is a research employee of the university.
The TU/e has nine departments:
The university offers honors programs aimed at both bachelor and master students. At the bachelor level it consists of intensive study within eight possible areas or tracks. At the master level it consists of personal leadership and professional development components, over and above the normal masters study.
In 1986, the university started a number of programs for a postgraduate doctorate of engineering (PDEng) together with two other Dutch technological universities (TU Delft and University of Twente). These programs are managed by the Stan Ackermans Institute on behalf of the 4TU Federation. Each program is two years in length. Ten programs are available at the TU/e.
Nationally, more than 3,500 students have earned the postgraduate PDEng degree through this program. On February 13, Ravi Thakkar was awarded 3000th PDEng diploma at TU/e
The university hosts a number of other educational programs that are in some way related to the main educational programs. These include the teacher's program and an MBA program.
The TU/e participates in a large number of research institutes which balance in different ways between pure science and applied science research. Some of these institutes are bound strictly to the university, others combine research across different universities.
The TU/e is among the world’s ten best-performing research universities in terms of research cooperation with industry in 2011 (Number 1 in 2009). Ten to 20 percent of the scientific publications of these ten universities in the period 2006–2008 were the result of partnerships with researchers in industry. As well as TU/e and Delft University of Technology, the top 10 also includes two universities in Japan (Tokyo Institute of Technology and Keio University in Tokyo), two in Sweden (CTH Chalmers University of Technology and KTH Royal Institute of Technology in Stockholm), and one each in Denmark (DTU Technical University of Denmark in Lyngby), Finland (University of Helsinki), Norway (Norwegian University of Science and Technology in Trondheim) and the USA (Rensselaer Polytechnic Institute in Troy, New York).
The TU/e plays a central role in the academic, economic and social life of Eindhoven and the surrounding region. In addition the university maintains relations with institutions far beyond that region as well and participates in national and international events (sometimes through the student body).
The TU/e is enormously important to the economy of the Eindhoven region, as well as the wider areas of BrabantStad and the Samenwerkingsverband Regio Eindhoven. It provides highly skilled labor for the local knowledge economy and is a knowledge and research partner for technology companies in the area.
The historic basis for the university's role as an economy and research motor was the interaction with Philips. The university was founded primarily to address the need of Philips for local personnel with academic levels of education in electronics, physics, chemistry and later computer science. Later that interest spread to DAF and Royal Dutch Shell (which became the primary employer for graduates of the chemistry department). There was also a synergy with these companies in that senior personnel were hired from them to form the academic staff of the university (which led to the Eindhoven joke that the university trains the engineers and Philips trains the professors).
Changing economic times and business strategies changed the relationship during the 1980s and 1990s. As Philips started moving away from the region, its importance to the region and the university decreased. A struggle for economic survival forced the university to seek closer ties with the city and region of Eindhoven in the 1989–1995 period, resulting in the creation of the Brainport initiative to draw high tech business and industry to the region. The university started expending more effort in knowledge valorisation, in incubating technology startups, in providing direct knowledge support for local technology companies. Also the academic interests of the research shifted with the times, with more effort going into energy efficiency research, green technologies, and other areas of interest driven by social relevance (the call for better technology in the medical field, for example, led to cooperation with the Catharina Hospital and the University of Maastricht medical department and finally the creation of the Biomedical Technology department).
The TU/e is host (and in some cases also commissioner) of a number of highly successful research schools, including the ESI and the DPI. These research institutes are a source of high-tech knowledge for high-tech companies in the area, such as ASML, NXP and FEI. The university also plays a large role as knowledge and personnel supplier to other companies in the High Tech Campus Eindhoven and helps incubate startups through the Eindhoven Twinning Center. It is also a knowledge supporter of the automotive industry in the Helmond region.
In the extended region, the TU/e is part of the backbone of the Eindhoven-Leuven-Aachen triangle. This economic cooperation agreement between three cities in three countries has created one of the most innovative regions in the European Union (measured in terms of money invested in technology and knowledge economy); the agreement is based on the cooperative triangle that connects the three technical universities in those cities.
As of the summer of 2010, the TU/e is host to the Eindhoven Energy Institute (EEI). The EEI is a virtual research institute (meaning that it doesn't have any actual offices or facilities), which manages and coordinates the activities of a large number of groups and subinstitutes in the general area of sustainable and alternative energy technologies.
The scientific director of the institute is prof.dr.ir. David Smeulders. He is pro forma head of the research department, which is split into four key areas: "Built Environment" (energy usage and patterns in building, headed by prof.dr.ir. Jan Hensen from the Department of the Built Environment), "Future Fuels" (headed by prof.dr. Philip de Goey of Mechanical Engineering), "Energy Conversion" (headed by prof.dr.ir. René Janssen from Chemical Engineering) and "Fusion and Plasma" (headed by prof.dr. Niek Lopes Cardozo from Physics). The EEI also incorporates the Graduate School on Sustainable Energy, which the TU/e had already established together with the TU Munich and DTU Lyngby. Secretarial services will be provided by the Center Technology for Sustainable Development (TDO) which also already existed at the TU/e (since 1994).
Energy research at the TU/e is among the best in academic Europe (a February 2010 study by Reed Elsevier puts it second only to Imperial College London). This fact, as well as the unique attention to energy in the built-up environment, drew the attention of the European Institute of Innovation and Technology. The EEI is now a full co-location of EIT's KIC on Sustainable Energy (InnoEnergy).
The TU/e maintains active academic cooperation with sister institutions in many different countries, for example:
The TU/e also provides education to an increasing number of foreign students and graduates. According to the 2009 annual report in the academic year 2008–2009 there were 490 exchange students, 103 foreign nationals registered in a bachelor program, 430 in a master program, 158 in a professional doctorate program (79% of the total). In 2009 the university employed 37 foreign professors (15.9% of the total) and 16 foreign associate professors (12.8%). Overall, 29.5% of the university staff was non-Dutch.
In 2011/2012, the TU/e has Erasmus bilateral agreements with many universities in 30 countries across Europe in a diverse range of subjects for student exchange.
In addition to the "regular" types of sports practiced among the student body and by the staff, the TU/e collaborates with the student body in a number of "technology sporting efforts". These usually take the form of cross-department projects, which makes them multidisciplinary efforts. Some examples include:
There university is more than just the departments, research bodies and the students. There are several ancillary activities necessary to the running of the university, activities that cross the boundaries and interests of the different departments. These activities are carried out by the universities' service organizations.
The university has the following service organizations:
Eindhoven is currently (2018) ranked between 51 and 141 in the world (the university itself provides a survey), and a top ten technical university in Europe.
In a 2003 European Commission report, TU/e was ranked as third among European research universities (after Cambridge and Oxford, at equality with TU Munich and thus making it the highest ranked Technical University in Europe), based on the impact of its scientific research.
In 2011 Academic Ranking of World Universities (ARWU) rankings, TU/e was placed at the 52-75 bucket internationally in Engineering/Technology and Computer Science ( ENG ) category and at 34th place internationally in the Computer Science subject field. | https://en.wikipedia.org/wiki?curid=9706 |
Electronegativity
Electronegativity, symbol "χ", is a concept that describes the tendency of an atom to attract a shared pair of electrons (or electron density) towards itself. An atom's electronegativity is affected by both its atomic number and the distance at which its valence electrons reside from the charged nucleus. The higher the associated electronegativity number, the more an atom or a substituent group attracts electrons towards itself.
On the most basic level, electronegativity is determined by factors like the nuclear charge (the more protons an atom has, the more "pull" it will have on electrons) and the number/location of other electrons present in the atomic shells (the more electrons an atom has, the farther from the nucleus the valence electrons will be, and as a result the less positive charge they will experience—both because of their increased distance from the nucleus, and because the other electrons in the lower energy core orbitals will act to shield the valence electrons from the positively charged nucleus).
The opposite of electronegativity is electropositivity: a measure of an element's ability to donate electrons.
The term "electronegativity" was introduced by Jöns Jacob Berzelius in 1811,
though the concept was known even before that and was studied by many chemists including Avogadro.
In spite of its long history, an accurate scale of electronegativity was not developed until 1932, when Linus Pauling proposed an electronegativity scale, which depends on bond energies, as a development of valence bond theory. It has been shown to correlate with a number of other chemical properties. Electronegativity cannot be directly measured and must be calculated from other atomic or molecular properties. Several methods of calculation have been proposed, and although there may be small differences in the numerical values of the electronegativity, all methods show the same periodic trends between elements.
The most commonly used method of calculation is that originally proposed by Linus Pauling. This gives a dimensionless quantity, commonly referred to as the Pauling scale ("χ"r), on a relative scale running from 0.79 to 3.98 (hydrogen = 2.20). When other methods of calculation are used, it is conventional (although not obligatory) to quote the results on a scale that covers the same range of numerical values: this is known as an electronegativity in "Pauling units".
As it is usually calculated, electronegativity is not a property of an atom alone, but rather a property of an atom in a molecule. Properties of a free atom include ionization energy and electron affinity. It is to be expected that the electronegativity of an element will vary with its chemical environment, but it is usually considered to be a transferable property, that is to say that similar values will be valid in a variety of situations.
Caesium is the least electronegative element in the periodic table (= 0.79), while fluorine is most electronegative (= 3.98). Francium and caesium were originally both assigned 0.7; caesium's value was later refined to 0.79, but no experimental data allows a similar refinement for francium. However, francium's ionization energy is known to be slightly higher than caesium's, in accordance with the relativistic stabilization of the 7s orbital, and this in turn implies that francium is in fact more electronegative than caesium.
Pauling first proposed the concept of electronegativity in 1932 as an explanation of the fact that the covalent bond between two different atoms (A–B) is stronger than would be expected by taking the average of the strengths of the A–A and B–B bonds. According to valence bond theory, of which Pauling was a notable proponent, this "additional stabilization" of the heteronuclear bond is due to the contribution of ionic canonical forms to the bonding.
The difference in electronegativity between atoms A and B is given by:
where the dissociation energies, "E"d, of the A–B, A–A and B–B bonds are expressed in electronvolts, the factor (eV)− being included to ensure a dimensionless result. Hence, the difference in Pauling electronegativity between hydrogen and bromine is 0.73 (dissociation energies: H–Br, 3.79 eV; H–H, 4.52 eV; Br–Br 2.00 eV)
As only differences in electronegativity are defined, it is necessary to choose an arbitrary reference point in order to construct a scale. Hydrogen was chosen as the reference, as it forms covalent bonds with a large variety of elements: its electronegativity was fixed first at 2.1, later revised to 2.20. It is also necessary to decide which of the two elements is the more electronegative (equivalent to choosing one of the two possible signs for the square root). This is usually done using "chemical intuition": in the above example, hydrogen bromide dissolves in water to form H+ and Br− ions, so it may be assumed that bromine is more electronegative than hydrogen. However, in principle, since the same electronegativities should be obtained for any two bonding compounds, the data are in fact overdetermined, and the signs are unique once a reference point is fixed (usually, for H or F).
To calculate Pauling electronegativity for an element, it is necessary to have data on the dissociation energies of at least two types of covalent bond formed by that element. A. L. Allred updated Pauling's original values in 1961 to take account of the greater availability of thermodynamic data, and it is these "revised Pauling" values of the electronegativity that are most often used.
The essential point of Pauling electronegativity is that there is an underlying, quite accurate, semi-empirical formula for dissociation energies, namely:
or sometimes, a more accurate fit
This is an approximate equation, but holds with good accuracy. Pauling obtained it by noting that a bond can be approximately represented as a quantum mechanical superposition of a covalent bond and two ionic bond-states. The covalent energy of a bond is approximately, by quantum mechanical calculations, the geometric mean of the two energies of covalent bonds of the same molecules, and there is an additional energy that comes from ionic factors, i.e. polar character of the bond.
The geometric mean is approximately equal to the arithmetic mean - which is applied in the first formula above - when the energies are of the similar value, e.g., except for the highly electropositive elements, where there is a larger difference of two dissociation energies; the geometric mean is more accurate and almost always gives a positive excess energy, due to ionic bonding. The square root of this excess energy, Pauling notes, is approximately additive, and hence one can introduce the electronegativity. Thus, it is this semi-empirical formula for bond energy that underlies Pauling electronegativity concept.
The formulas are approximate, but this rough approximation is in fact relatively good and gives the right intuition, with the notion of polarity of the bond and some theoretical grounding in quantum mechanics. The electronegativities are then determined to best fit the data.
In more complex compounds, there is additional error since electronegativity depends on the molecular environment of an atom. Also, the energy estimate can be only used for single, not for multiple bonds. The energy of formation of a molecule containing only single bonds then can be approximated from an electronegativity table, and depends on the constituents and sum of squares of differences of electronegativities of all pairs of bonded atoms. Such a formula for estimating energy typically has relative error of order of 10%, but can be used to get a rough qualitative idea and understanding of a molecule.
Robert S. Mulliken proposed that the arithmetic mean of the first ionization energy (Ei) and the electron affinity (Eea) should be a measure of the tendency of an atom to attract electrons. As this definition is not dependent on an arbitrary relative scale, it has also been termed absolute electronegativity, with the units of kilojoules per mole or electronvolts.
However, it is more usual to use a linear transformation to transform these absolute values into values that resemble the more familiar Pauling values. For ionization energies and electron affinities in electronvolts,
and for energies in kilojoules per mole,
The Mulliken electronegativity can only be calculated for an element for which the electron affinity is known, fifty-seven elements as of 2006.
The Mulliken electronegativity of an atom is sometimes said to be the negative of the chemical potential. By inserting the energetic definitions of the ionization potential and electron affinity into the Mulliken electronegativity, it is possible to show that the Mulliken chemical potential is a finite difference approximation of the electronic energy with respect to the number of electrons., i.e.,
A. Louis Allred and Eugene G. Rochow considered that electronegativity should be related to the charge experienced by an electron on the "surface" of an atom: The higher the charge per unit area of atomic surface the greater the tendency of that atom to attract electrons. The effective nuclear charge, "Z"eff, experienced by valence electrons can be estimated using Slater's rules, while the surface area of an atom in a molecule can be taken to be proportional to the square of the covalent radius, "r"cov. When "r"cov is expressed in picometres,
R.T. Sanderson has also noted the relationship between Mulliken electronegativity and atomic size, and has proposed a method of calculation based on the reciprocal of the atomic volume. With a knowledge of bond lengths, Sanderson's model allows the estimation of bond energies in a wide range of compounds. Sanderson's model has also been used to calculate molecular geometry, "s"-electrons energy, NMR spin-spin constants and other parameters for organic compounds. This work underlies the concept of electronegativity equalization, which suggests that electrons distribute themselves around a molecule to minimize or to equalize the Mulliken electronegativity. This behavior is analogous to the equalization of chemical potential in macroscopic thermodynamics.
Perhaps the simplest definition of electronegativity is that of Leland C. Allen, who has proposed that it is related to the average energy of the valence electrons in a free atom,
where "ε"s,p are the one-electron energies of s- and p-electrons in the free atom and "n"s,p are the number of s- and p-electrons in the valence shell. It is usual to apply a scaling factor, 1.75×10−3 for energies expressed in kilojoules per mole or 0.169 for energies measured in electronvolts, to give values that are numerically similar to Pauling electronegativities.
The one-electron energies can be determined directly from spectroscopic data, and so electronegativities calculated by this method are sometimes referred to as spectroscopic electronegativities. The necessary data are available for almost all elements, and this method allows the estimation of electronegativities for elements that cannot be treated by the other methods, e.g. francium, which has an Allen electronegativity of 0.67. However, it is not clear what should be considered to be valence electrons for the d- and f-block elements, which leads to an ambiguity for their electronegativities calculated by the Allen method.
In this scale neon has the highest electronegativity of all elements, followed by fluorine, helium, and oxygen.
The wide variety of methods of calculation of electronegativities, which all give results that correlate well with one another, is one indication of the number of chemical properties which might be affected by electronegativity. The most obvious application of electronegativities is in the discussion of bond polarity, for which the concept was introduced by Pauling. In general, the greater the difference in electronegativity between two atoms the more polar the bond that will be formed between them, with the atom having the higher electronegativity being at the negative end of the dipole. Pauling proposed an equation to relate "ionic character" of a bond to the difference in electronegativity of the two atoms, although this has fallen somewhat into disuse.
Several correlations have been shown between infrared stretching frequencies of certain bonds and the electronegativities of the atoms involved: however, this is not surprising as such stretching frequencies depend in part on bond strength, which enters into the calculation of Pauling electronegativities. More convincing are the correlations between electronegativity and chemical shifts in NMR spectroscopy or isomer shifts in Mössbauer spectroscopy (see figure). Both these measurements depend on the s-electron density at the nucleus, and so are a good indication that the different measures of electronegativity really are describing "the ability of an atom in a molecule to attract electrons to itself".
In general, electronegativity increases on passing from left to right along a period, and decreases on descending a group. Hence, fluorine is the most electronegative of the elements (not counting noble gases), whereas caesium is the least electronegative, at least of those elements for which substantial data is available. This would lead one to believe that caesium fluoride is the compound whose bonding features the most ionic character.
There are some exceptions to this general rule. Gallium and germanium have higher electronegativities than aluminium and silicon, respectively, because of the d-block contraction. Elements of the fourth period immediately after the first row of the transition metals have unusually small atomic radii because the 3d-electrons are not effective at shielding the increased nuclear charge, and smaller atomic size correlates with higher electronegativity (see Allred-Rochow electronegativity, Sanderson electronegativity above). The anomalously high electronegativity of lead, in particular when compared to thallium and bismuth, is an artifact of electronegativity varying with oxidation state: its electronegativity conforms better to trends if it is quoted for the +2 state instead of the +4 state.
In inorganic chemistry it is common to consider a single value of the electronegativity to be valid for most "normal" situations. While this approach has the advantage of simplicity, it is clear that the electronegativity of an element is "not" an invariable atomic property and, in particular, increases with the oxidation state of the element.
Allred used the Pauling method to calculate separate electronegativities for different oxidation states of the handful of elements (including tin and lead) for which sufficient data was available. However, for most elements, there are not enough different covalent compounds for which bond dissociation energies are known to make this approach feasible. This is particularly true of the transition elements, where quoted electronegativity values are usually, of necessity, averages over several different oxidation states and where trends in electronegativity are harder to see as a result.
The chemical effects of this increase in electronegativity can be seen both in the structures of oxides and halides and in the acidity of oxides and oxoacids. Hence CrO3 and Mn2O7 are acidic oxides with low melting points, while Cr2O3 is amphoteric and Mn2O3 is a completely basic oxide.
The effect can also be clearly seen in the dissociation constants of the oxoacids of chlorine. The effect is much larger than could be explained by the negative charge being shared among a larger number of oxygen atoms, which would lead to a difference in p"K"a of log10() = –0.6 between hypochlorous acid and perchloric acid. As the oxidation state of the central chlorine atom increases, more electron density is drawn from the oxygen atoms onto the chlorine, reducing the partial negative charge on the oxygen atoms and increasing the acidity.
The electronegativity of an atom changes depending on the hybridization of the orbital employed in bonding. Electrons in s orbitals are held more tightly than electrons in p orbitals. Hence, a bond to an atom that employs an sp"x" hybrid orbital for bonding will be more heavily polarized to that atom when the hybrid orbital has more s character. That is, when electronegativities are compared for different hybridization schemes of a given element, the order χ(sp3) < χ(sp2) < χ(sp) holds (the trend should apply to non-integer hybridization indices as well). While this holds true in principle for any main-group element, values for the hybridization-specific electronegativity are most frequently cited for carbon. In organic chemistry, these electronegativities are frequently invoked to predict or rationalize bond polarities in organic compounds containing double and triple bonds to carbon.
In organic chemistry, electronegativity is associated more with different functional groups than with individual atoms. The terms group electronegativity and substituent electronegativity are used synonymously. However, it is common to distinguish between the inductive effect and the resonance effect, which might be described as σ- and π-electronegativities, respectively. There are a number of linear free-energy relationships that have been used to quantify these effects, of which the Hammett equation is the best known. Kabachnik parameters are group electronegativities for use in organophosphorus chemistry.
Electropositivity is a measure of an element's ability to donate electrons, and therefore form positive ions; thus, it is opposed to electronegativity.
Mainly, this is an attribute of metals, meaning that, in general, the greater the metallic character of an element the greater the electropositivity. Therefore, the alkali metals are most electropositive of all. This is because they have a single electron in their outer shell and, as this is relatively far from the nucleus of the atom, it is easily lost; in other words, these metals have low ionization energies.
While electronegativity increases along periods in the periodic table, and decreases down groups, electropositivity "decreases" along periods (from left to right) and "increases" down groups. This means that elements in the upper right of the periodic table of elements (oxygen, sulfur, chlorine, etc.) will have the greatest electronegativity, and those in the lower left (rubidium, cesium, and francium) the greatest electropositivity. | https://en.wikipedia.org/wiki?curid=9707 |
European Charter for Regional or Minority Languages
The European Charter for Regional or Minority Languages (ECRML) is a European treaty (CETS 148) adopted in 1992 under the auspices of the Council of Europe to protect and promote historical regional and minority languages in Europe. The preparation for the charter was undertaken by the predecessor to the current Congress of Local and Regional Authorities, the Standing Conference of Local and Regional Authorities of Europe because involvement of local and regional government was essential. The actual charter was written in the Parliamentary Assembly based on the Congress' Recommendations. It only applies to languages traditionally used by the nationals of the State Parties (thus excluding languages used by recent immigrants from other states, see immigrant languages), which significantly differ from the majority or official language (thus excluding what the state party wishes to consider as mere local dialects of the official or majority language) and that either have a territorial basis (and are therefore traditionally spoken by populations of regions or areas within the State) or are used by linguistic minorities within the State as a whole (thereby including such languages as Yiddish, Romani and Lemko, which are used over a wide geographic area).
Some states, such as Ukraine and Sweden, have tied the status of minority language to the recognized national minorities, which are defined by ethnic, cultural and/or religious criteria, thereby circumventing the Charter's notion of linguistic minority.
Languages that are official within regions, provinces or federal units within a State (for example Catalan in Spain) are not classified as official languages of the State and may therefore benefit from the Charter. On the other hand, Ireland has not been able to sign the Charter on behalf of the Irish language (although a minority language) as it is defined as the first official language of the state. The United Kingdom has ratified the Charter in respect to (among other languages) Welsh in Wales, Scots and Gaelic in Scotland, and Irish in Northern Ireland. France, although a signatory, has been constitutionally blocked from ratifying the Charter in respect to the languages of France.
The charter provides many actions state parties can take to protect and promote historical regional and minority languages. There are two levels of protection—all signatories must apply the lower level of protection to qualifying languages. Signatories may further declare that a qualifying language or languages will benefit from the higher level of protection, which lists a range of actions from which states must agree to undertake at least 35.
Countries can ratify the charter in respect of its minority languages based on Part II or Part III of the charter, which contain varying principles. Countries can treat languages differently under the charter, for example, in the United Kingdom, the Welsh language is ratified under the general Part II principles as well as the more specific Part III commitments, while the Cornish language is ratified only under Part II.
Part II of the Charter details eight main principles and objectives upon which States must base their policies and legislation. They are seen as a framework for the preservation of the languages concerned.
Part III details comprehensive rules, across a number of sectors, by which states agree to abide. Each language to which Part III of the Charter is applied must be named specifically by the government. States must select at least thirty-five of the undertakings in respect to each language. Many provisions contain several options, of varying degrees of stringency, one of which has to be chosen “according to the situation of each language”. The areas from which these specific undertakings must be chosen are as follows:
Countries that have ratified the Charter, and languages for which the ratification was made: | https://en.wikipedia.org/wiki?curid=9708 |
English Civil War
The English Civil War (1642–1651) was a series of civil wars and political machinations between Parliamentarians ("Roundheads") and Royalists ("Cavaliers") principally over the manner of England's governance and part of the wider Wars of the Three Kingdoms. The first (1642–1646) and second (1648–1649) wars pitted the supporters of King Charles I against the supporters of the Long Parliament, while the third (1649–1651) saw fighting between supporters of King Charles II and supporters of the Rump Parliament. The war ended with Parliamentarian victory at the Battle of Worcester on 3 September 1651.
Unlike other civil wars in England, which were mainly fought over "who" should rule, these conflicts were also concerned with "how" the three kingdoms of England, Scotland, and Ireland were to be governed. The outcome was threefold: the trial and execution of Charles I (1649); the exile of his son, Charles II (1651); and the replacement of English monarchy with, at first, the Commonwealth of England (1649–1653) and then the Protectorate, which as the Commonwealth of England, Scotland, and Ireland unified the British Isles under the personal rule of Oliver Cromwell (1653–1658) and briefly his son Richard (1658–1659). The execution of Charles I was particularly notable given that an English king had never been executed before. In England, the monopoly of the Church of England on Christian worship was ended, while in Ireland the victors consolidated the established Protestant Ascendancy. Constitutionally, the wars established the precedent that an English monarch cannot govern without Parliament's consent, although the idea of Parliamentary sovereignty was only legally established as part of the Glorious Revolution in 1688.
The term "English Civil War" appears most often in the singular, although historians often divide the conflict into two or three separate wars. These were not restricted to England, as Wales was part of the Kingdom of England and affected accordingly. The conflicts also involved wars with Scotland and Ireland, and civil wars within them.
The wars spanning all four countries are known as the Wars of the Three Kingdoms. In the early 19th century, Sir Walter Scott referred to it as "the Great Civil War".
The 1911 "Encyclopædia Britannica" called the series of conflicts the "Great Rebellion", while some historians – notably Marxists such as Christopher Hill (1912–2003) – long favoured the term "English Revolution".
Each side had a geographical stronghold, such that minority elements were silenced or fled. The Royalist areas included the countryside, the shires, the cathedral city of Oxford, and the less economically developed areas of northern and western England. Parliament's strengths spanned the industrial centres, ports, and economically advanced regions of southern and eastern England, including the remaining cathedral cities (except York, Chester, Worcester). Lacey Baldwin Smith says, "the words "populous, rich, and rebellious" seemed to go hand in hand".
Many officers and veteran soldiers had fought in European wars, notably the Eighty Years' War between the Spanish and the Dutch, which began in 1568.
The main battle tactic came to be known as pike and shot infantry. The two sides would line up opposite one another, with infantry brigades of musketeers in the centre. These carried matchlock muskets, an inaccurate weapon which nevertheless could be lethal at a range of up to 300 yards. Musketeers would assemble three rows deep, the first kneeling, second crouching, and third standing, allowing all to fire a volley simultaneously. At times, troops divided into two groups, allowing one to reload while the other fired. Among the musketeers were pike men, carrying pikes of to long, whose main purpose was to protect the musketeers from cavalry charges. Positioned on each side of the infantry were cavalry, with a right wing led by the lieutenant-general and left by the commissary general. Its main aim was to rout the opponents' cavalry, then turn and overpower their infantry.
The Royalist cavaliers' skill and speed on horseback led to many early victories. Prince Rupert, commanding the king's cavalry, used a tactic learned while fighting in the Dutch army, where cavalry would charge at full speed into the opponent's infantry, firing their pistols just before impact.
However, with Oliver Cromwell and the introduction of the more disciplined New Model Army, a group of disciplined pike men would stand its ground, which could have a devastating effect.
The Royalist cavalry had a tendency to chase down individual targets after the initial charge, leaving their forces scattered and tired, whereas Cromwell's cavalry was slower but better disciplined. Trained to operate as a single unit, it went on to win many decisive victories.
The English Civil War broke out in 1642, less than 40 years after the death of Queen Elizabeth I. Elizabeth had been succeeded by her first cousin twice-removed, King James VI of Scotland, as James I of England, creating the first personal union of the Scottish and English kingdoms. As King of Scots, James had become accustomed to Scotland's weak parliamentary tradition since assuming control of the Scottish government in 1583, so that upon assuming power south of the border, the new King of England was affronted by the constraints the English Parliament attempted to place on him in exchange for money. In spite of this, James's personal extravagance meant he was perennially short of money and had to resort to extra-parliamentary sources of income.
This extravagance was tempered by James's peaceful disposition, so that by the succession of his son Charles I in 1625 the two kingdoms had both experienced relative peace, internally and in their relations with each other, for as long as anyone could remember. Charles followed his father's dream in hoping to unite the kingdoms of England, Scotland and Ireland into a single kingdom. Many English Parliamentarians were suspicious of such a move, fearing that such a new kingdom might destroy old English traditions that had bound the English monarchy. As Charles shared his father's position on the power of the crown (James had described kings as "little gods on Earth", chosen by God to rule in accordance with the doctrine of the "Divine Right of Kings"), the suspicions of the Parliamentarians had some justification.
At the time, the Parliament of England did not have a large permanent role in the English system of government. Instead, it functioned as a temporary advisory committee and was summoned only if and when the monarch saw fit. Once summoned, a Parliament's continued existence was at the king's pleasure since it was subject to dissolution by him at any time.
Yet in spite of this limited role, Parliament had acquired over the centuries "de facto" powers of enough significance that monarchs could not simply ignore them indefinitely. For a monarch, Parliament's most indispensable power was its ability to raise tax revenues far in excess of all other sources of revenue at the Crown's disposal. By the 17th century, Parliament's tax-raising powers had come to be derived from the fact that the gentry was the only stratum of society with the ability and authority to collect and remit the most meaningful forms of taxation then available at the local level. So if the king wanted to ensure smooth revenue collection, he needed gentry co-operation. For all of the Crown's legal authority, its resources were limited by any modern standard to an extent that if the gentry refused to collect the king's taxes on a national scale, the Crown lacked a practical means of compelling them.
From the thirteenth century, monarchs ordered the election of representatives to sit in the House of Commons, with most voters being the owners of property, although in some potwalloper boroughs every male householder could vote. When assembled along with the House of Lords, these elected representatives formed a Parliament. So the concept of Parliaments allowed representatives of the property-owning class to meet, primarily, at least from the point of view of the monarch, to sanction whatever taxes the monarch wished to collect. In the process, the representatives could debate and enact statutes, or acts. However, Parliament lacked the power to force its will upon the monarch; its only leverage was the threat of withholding the financial means required to implement his plans.
Many concerns were raised over Charles's marriage in 1625 to a Roman Catholic French princess: Henrietta Maria. Parliament refused to assign him the traditional right to collect customs duties for his entire reign, deciding instead to grant it only on a provisional basis and negotiate with him.
Charles, meanwhile, decided to send an expeditionary force to relieve the French Huguenots, whom French royal troops held besieged in La Rochelle. Such military support for Protestants on the Continent potentially alleviated concerns about the King's marriage to a Catholic. However, Charles's insistence on giving command of the English force to his unpopular royal favourite George Villiers, the Duke of Buckingham, undermined that support. Unfortunately for Charles and Buckingham, the relief expedition proved a fiasco (1627), and Parliament, already hostile to Buckingham for his monopoly on royal patronage, opened impeachment proceedings against him. Charles responded by dissolving Parliament. This saved Buckingham but confirmed the impression that Charles wanted to avoid Parliamentary scrutiny of his ministers.
Having dissolved Parliament and unable to raise money without it, the king assembled a new one in 1628. (The elected members included Oliver Cromwell, John Hampden, and Edward Coke.) The new Parliament drew up a Petition of Right, which Charles accepted as a concession to obtain his subsidy. The Petition made reference to Magna Carta, but did not grant him the right of tonnage and poundage, which Charles had been collecting without Parliamentary authorisation since 1625. Several more active members of the opposition were imprisoned, which caused outrage; one, John Eliot, subsequently died in prison and came to be seen as a martyr for the rights of Parliament.
Charles avoided calling a Parliament for the next decade, a period known as the "personal rule of Charles I", or the "Eleven Years' Tyranny". During this period, Charles's policies were determined by his lack of money. First and foremost, to avoid Parliament, the King needed to avoid war. Charles made peace with France and Spain, effectively ending England's involvement in the Thirty Years' War. However, that in itself was far from enough to balance the Crown's finances.
Unable to raise revenue without Parliament and unwilling to convene it, Charles resorted to other means. One was to revive conventions, often outdated. For example, a failure to attend and receive knighthood at Charles's coronation became a finable offence with the fine paid to the Crown. The King also tried to raise revenue through ship money, demanding in 1634–1636 that the inland English counties pay a tax for the Royal Navy to counter the threat of privateers and pirates in the English Channel. Established law supported the policy of coastal counties and inland ports such as London paying ship money in times of need, but it had not been applied to inland counties before. Authorities had ignored it for centuries, and many saw it as yet another extra-Parliamentary, illegal tax, which prompted some prominent men to refuse to pay it. Charles issued a writ against John Hampden for his failure to pay, and although five judges including Sir George Croke supported Hampden, seven judges found in favour of the King in 1638. The fines imposed on people who refused to pay ship money and standing out against its illegality aroused widespread indignation.
During his "Personal Rule", Charles aroused most antagonism through his religious measures. He believed in High Anglicanism, a sacramental version of the Church of England, theologically based upon Arminianism, a creed shared with his main political adviser, Archbishop William Laud. In 1633, Charles appointed Laud Archbishop of Canterbury and started making the Church more ceremonial, replacing the wooden communion tables with stone altars. Puritans accused Laud of reintroducing Catholicism; when they complained, he had them arrested. In 1637, John Bastwick, Henry Burton, and William Prynne had their ears cut off for writing pamphlets attacking Laud's views — a rare penalty for gentlemen, and one that aroused anger. Moreover, the Church authorities revived statutes from the time of Elizabeth I about church attendance and fined Puritans for not attending Anglican services.
The end of Charles's independent governance came when he attempted to apply the same religious policies in Scotland. The Church of Scotland, reluctantly episcopal in structure, had independent traditions. Charles wanted one uniform Church throughout Britain and introduced a new, High Anglican version of the English Book of Common Prayer to Scotland in the middle of 1637. This was violently resisted. A riot broke out in Edinburgh, which may have been started in St Giles' Cathedral, according to legend, by Jenny Geddes. In February 1638, the Scots formulated their objections to royal policy in the National Covenant. This document took the form of a "loyal protest", rejecting all innovations not first tested by free Parliaments and General Assemblies of the Church.
In the spring of 1639, King Charles I accompanied his forces to the Scottish border to end the rebellion known as the Bishops' War, but after an inconclusive campaign, he accepted the offered Scottish truce: the Pacification of Berwick. This truce proved temporary, and a second war followed in mid-1640. A Scots army defeated Charles's forces in the north, then captured Newcastle. Charles eventually agreed not to interfere in Scotland's religion and paid the Scots' war expenses.
Charles needed to suppress the rebellion in Scotland, but had insufficient funds to do so. He needed to seek money from a newly elected English Parliament in 1640. Its majority faction, led by John Pym, used this appeal for money as a chance to discuss grievances against the Crown and oppose the idea of an English invasion of Scotland. Charles took exception to this "lèse-majesté" (offense against the ruler) and dissolved the Parliament after only a few weeks; hence its name, "the Short Parliament".
Without Parliament's support, Charles attacked Scotland again, breaking the truce at Berwick, and suffered comprehensive defeat. The Scots went on to invade England, occupying Northumberland and Durham. Meanwhile, another of Charles's chief advisers, Thomas Wentworth, 1st Viscount Wentworth, had risen to the role of Lord Deputy of Ireland in 1632, and brought in much-needed revenue for Charles by persuading the Irish Catholic gentry to pay new taxes in return for promised religious concessions.
In 1639, Charles had recalled Wentworth to England and in 1640 made him Earl of Strafford, attempting to have him achieve similar results in Scotland. This time he proved less successful and the English forces fled the field at their second encounter with the Scots in 1640. Almost the whole of Northern England was occupied and Charles forced to pay £850 per day to keep the Scots from advancing. Had he not done so they would have pillaged and burnt the cities and towns of Northern England.
All this put Charles in a desperate financial state. As King of Scots, he had to find money to pay the Scottish army in England; as King of England, he had to find money to pay and equip an English army to defend England. His means of raising English revenue without an English Parliament fell critically short of achieving this. Against this backdrop, and according to advice from the Magnum Concilium (the House of Lords, but without the Commons, so not a Parliament), Charles finally bowed to pressure and summoned another English Parliament in November 1640.
The new Parliament proved even more hostile to Charles than its predecessor. It immediately began to discuss grievances against him and his government, with Pym and Hampden (of ship money fame) in the lead. They took the opportunity presented by the King's troubles to force various reforming measures — including many with strong "anti-Papist" themes — upon him. The members passed a law stating that a new Parliament would convene at least once every three years — without the King's summons if need be. Other laws passed making it illegal for the king to impose taxes without Parliamentary consent and later gave Parliament control over the king's ministers. Finally, the Parliament passed a law forbidding the King to dissolve it without its consent, even if the three years were up. Ever since this Parliament has been known as the Long Parliament. However, Parliament did attempt to avert conflict by requiring all adults to sign The Protestation, an oath of allegiance to Charles.
Early in the Long Parliament, the house overwhelmingly accused Thomas Wentworth, Earl of Strafford of high treason and other crimes and misdemeanors.
Henry Vane the Younger supplied evidence of Strafford's claimed improper use of the army in Ireland, alleging that he had encouraged the King to use his Ireland-raised forces to threaten England into compliance. This evidence was obtained from Vane's father, Henry Vane the Elder, a member of the King's Privy council, who refused to confirm it in Parliament out of loyalty to Charles. On 10 April 1641, Pym's case collapsed, but Pym made a direct appeal to Henry Vane the Younger to produce a copy of the notes from the King's Privy council, discovered by the younger Vane and secretly turned over to Pym, to the great anguish of the Elder Vane. These notes contained evidence that Strafford had told the King, "Sir, you have done your duty, and your subjects have failed in theirs; and therefore you are absolved from the rules of government, and may supply yourself by extraordinary ways; you have an army in Ireland, with which you may reduce the kingdom."
Pym immediately launched a Bill of Attainder stating Strafford's guilt and demanding that he be put to death. Unlike a guilty verdict in a court case, attainder did not require a legal burden of proof, but it did require the king's approval. Charles, however, guaranteed Strafford that he would not sign the attainder, without which the bill could not be passed. Furthermore, the Lords opposed the severity of a death sentence on Strafford. Yet increased tensions and a plot in the army to support Strafford began to sway the issue. On 21 April, the Commons passed the Bill (204 in favour, 59 opposed, and 250 abstained), and the Lords acquiesced. Charles, still incensed over the Commons' handling of Buckingham, refused his assent. Strafford himself, hoping to head off the war he saw looming, wrote to the king and asked him to reconsider. Charles, fearing for the safety of his family, signed on 10 May. Strafford was beheaded two days later. In the meantime both Parliament and the King agreed to an independent investigation into the king's involvement in Strafford's plot.
The Long Parliament then passed the Triennial Act, also known as the Dissolution Act in May 1641, to which the Royal Assent was readily granted. The Triennial Act required Parliament to be summoned at least once in three years. When the King failed to issue a proper summons, the members could assemble on their own. This act also forbade ship money without Parliament's consent, fines in distraint of knighthood, and forced loans. Monopolies were cut back sharply, the Courts of the Star Chamber and High Commission abolished by the Habeas Corpus Act 1640, and the Triennial Act respectively. All remaining forms of taxation were legalised and regulated by the Tonnage and Poundage Act. On 3 May, Parliament decreed The Protestation, attacking the 'wicked counsels' of Charles's government, whereby those who signed the petition undertook to defend 'the true reformed religion', Parliament, and the king's person, honour and estate. Throughout May, the House of Commons launched several bills attacking bishops and Episcopalianism in general, each time defeated in the Lords.
Charles and his Parliament hoped that the execution of Strafford and the Protestation would end the drift towards war, but in fact, they encouraged it. Charles and his supporters continued to resent Parliament's demands, and Parliamentarians continued to suspect Charles of wanting to impose episcopalianism and unfettered royal rule by military force. Within months, the Irish Catholics, fearing a resurgence of Protestant power, struck first, and all Ireland soon descended into chaos. Rumors circulated that the King supported the Irish, and Puritan members of the Commons soon started murmuring that this exemplified the fate that Charles had in store for them all.
In early January 1642, Charles, accompanied by 400 soldiers, attempted to arrest five members of the House of Commons on a charge of treason. This attempt failed. When the troops marched into Parliament, Charles enquired of William Lenthall, the Speaker, as to the whereabouts of the five. Lenthall replied, "May it please your Majesty, I have neither eyes to see nor tongue to speak in this place but as the House is pleased to direct me, whose servant I am here." So the Speaker proclaimed himself a servant of Parliament, rather than the King.
In the summer of 1642, these national troubles helped to polarise opinion, ending indecision about which side to support or what action to take. Opposition to Charles also arose from many local grievances. For example, imposed drainage schemes in The Fens disrupted the livelihood of thousands after the King awarded a number of drainage contracts. Many saw the King as indifferent to public welfare, and this played a role in bringing much of eastern England into the Parliamentarian camp. This sentiment brought with it such people as the Earl of Manchester and Oliver Cromwell, each a notable wartime adversary of the King. Conversely, one of the leading drainage contractors, the Earl of Lindsey, was to die fighting for the King at the Battle of Edgehill.
In early January 1642, a few days after failing to capture five members of the House of Commons, Charles feared for the safety of his family and retinue and left the London area for the north country. Further frequent negotiations by letter between the King and the Long Parliament, through to early summer, proved fruitless. As the summer progressed, cities and towns declared their sympathies for one faction or the other: for example, the garrison of Portsmouth commanded by Sir George Goring declared for the King, but when Charles tried to acquire arms from Kingston upon Hull, the weaponry depository used in the previous Scottish campaigns, Sir John Hotham, the military governor appointed by Parliament in January, refused to let Charles enter the town, and when Charles returned with more men later, Hotham drove them off. Charles issued a warrant for Hotham's arrest as a traitor but was powerless to enforce it. Throughout the summer, tensions rose and there was brawling in several places, the first death from the conflict taking place in Manchester.
At the outset of the conflict, much of the country remained neutral, though the Royal Navy and most English cities favoured Parliament, while the King found marked support in rural communities. Historians estimate that both sides had only about 15,000 men between them, but the war quickly spread and eventually involved every level of society. Many areas attempted to remain neutral. Some formed bands of Clubmen to protect their localities from the worst excesses of the armies of both sides, but most found it impossible to withstand both King and Parliament. On one side, the King and his supporters fought for traditional government in church and state, while on the other, most Parliamentarians initially took up arms to defend what they saw as a traditional balance of government in church and state, which the bad advice the King received from his advisers had undermined before and during the "Eleven Years' Tyranny". The views of the members of Parliament ranged from unquestioning support of the King — at one point during the First Civil War, more members of the Commons and Lords gathered in the King's Oxford Parliament than at Westminster — through to radicals who sought major reforms in religious independence and redistribution of power at a national level. However, even the most radical Parliamentarian supporters still favoured keeping Charles on the throne.
After the debacle at Hull, Charles moved on to Nottingham, raising the royal standard there on 22 August 1642. At the time, Charles had with him about 2,000 cavalries and a small number of Yorkshire infantrymen, and using the archaic system of a Commission of Array, his supporters started to build a larger army around the standard. Charles moved in a westerly direction, first to Stafford, then on to Shrewsbury, as support for his cause seemed particularly strong in the Severn valley area and in North Wales. While passing through Wellington, he declared in what became known as the "Wellington Declaration" that he would uphold the "Protestant religion, the laws of England, and the liberty of Parliament".
The Parliamentarians who opposed the King did not remain passive in this pre-war period. As in Hull, they took measures to secure strategic towns and cities by appointing to office men sympathetic to their cause. On 9 June they voted to raise an army of 10,000 volunteers and appointed Robert Devereux, 3rd Earl of Essex its commander three days later. He received orders "to rescue His Majesty's person, and the persons of the Prince [of Wales] and the Duke of York [James II] out of the hands of those desperate persons who were about them." The Lords Lieutenant whom Parliament appointed used the Militia Ordinance to order the militia to join Essex's army.
Two weeks after the King had raised his standard at Nottingham, Essex led his army north towards Northampton, picking up support along the way (including a detachment of Huntingdonshire cavalry raised and commanded by Oliver Cromwell). By mid-September Essex's forces had grown to 21,000 infantry and 4,200 cavalries and dragoons. On 14 September he moved his army to Coventry and then to the north of the Cotswolds, a strategy that placed it between the Royalists and London. With the size of both armies now in the tens of thousands and only Worcestershire between them, it was inevitable that cavalry reconnaissance units would meet sooner or later. This happened in the first major skirmish of the Civil War, when a troop of about 1,000 Royalist cavalry under Prince Rupert, a German nephew of the King and one of the outstanding cavalry commanders of the war, defeated a Parliamentary cavalry detachment under Colonel John Brown at the Battle of Powick Bridge, which crossed the River Teme close to Worcester.
Rupert withdrew to Shrewsbury, where a council-of-war discussed two courses of action: whether to advance towards Essex's new position near Worcester, or march down the now open road towards London. The Council decided on the London route, but not to avoid a battle, for the Royalist generals wanted to fight Essex before he grew too strong, and the temper of both sides made it impossible to postpone the decision. In the Earl of Clarendon's words, "it was considered more counsellable to march towards London, it being morally sure that the earl of Essex would put himself in their way." So the army left Shrewsbury on 12 October, gaining two days' start on the enemy, and moved south-east. This had the desired effect of forcing Essex to move to intercept them.
The first pitched battle of the war, at Edgehill on 23 October 1642, proved inconclusive, both Royalists and Parliamentarians claiming victory. The second field action, the stand-off at Turnham Green, saw Charles forced to withdraw to Oxford, which would serve as his base for the rest of the war.
In 1643, Royalist forces won at Adwalton Moor, gaining control of most of Yorkshire. In the Midlands, a Parliamentary force under Sir John Gell besieged and captured the cathedral city of Lichfield, after the death of the original commander, Lord Brooke. This group then joined forces with Sir John Brereton at the inconclusive Battle of Hopton Heath (19 March 1643), where the Royalist commander, the Earl of Northampton, was killed. John Hampden died after being wounded in the Battle of Chalgrove Field (18 June 1643). Subsequent battles in the west of England at Lansdowne and Roundway Down also went to the Royalists. Prince Rupert could then take Bristol. In the same year, however, Cromwell formed his troop of "Ironsides", a disciplined unit that demonstrated his military leadership ability. With their assistance he won a victory at the Battle of Gainsborough in July.
At this stage, from 7 to 9 August 1643, there were some popular demonstrations in London — both for and against war. They were protesting at Westminster. A peace demonstration by London women, which turned violent, was suppressed by William Waller's regiment of horse. Some women were beaten and even killed, and many arrested.
After these August events, the representative of Venice in England reported to the doge that the London government took considerable measures to stifle dissent.
In general, the early part of the war went well for the Royalists. The turning point came in the late summer and early autumn of 1643, when the Earl of Essex's army forced the king to raise the Siege of Gloucester and then brushed the Royalists aside at the First Battle of Newbury (20 September 1643), to return triumphantly to London. Parliamentarian forces led by the Earl of Manchester besieged the port of King's Lynn, Norfolk, which under Sir Hamon L'Estrange held out until September. Other forces won the Battle of Winceby, giving them control of Lincoln. Political manœuvring to gain an advantage in numbers led Charles to negotiate a ceasefire in Ireland, freeing up English troops to fight on the Royalist side in England, while Parliament offered concessions to the Scots in return for aid and assistance.
Helped by the Scots, Parliament won at Marston Moor (2 July 1644), gaining York and the north of England. Cromwell's conduct in the battle proved decisive, and showed his potential as a political and as an important military leader. The defeat at the Battle of Lostwithiel in Cornwall, however, marked a serious reverse for Parliament in the south-west of England. Subsequent fighting around Newbury (27 October 1644), though tactically indecisive, strategically gave another check to Parliament.
In 1645, Parliament reaffirmed its determination to fight the war to a finish. It passed the Self-denying Ordinance, by which all members of either House of Parliament laid down their commands and re-organized its main forces into the New Model Army, under the command of Sir Thomas Fairfax, with Cromwell as his second-in-command and Lieutenant-General of Horse. In two decisive engagements — the Battle of Naseby on 14 June and the Battle of Langport on 10 July — the Parliamentarians effectively destroyed Charles's armies.
In the remains of his English realm, Charles tried to recover a stable base of support by consolidating the Midlands. He began to form an axis between Oxford and Newark-on-Trent in Nottinghamshire. These towns had become fortresses and showed more reliable loyalty to him than others. He took Leicester, which lies between them, but found his resources exhausted. Having little opportunity to replenish them, in May 1646 he sought shelter with a Presbyterian Scottish army at Southwell in Nottinghamshire. Charles was eventually handed over to the English Parliament by the Scots and imprisoned. This marked the end of the First English Civil War.
The end of the First Civil War, in 1646, left a partial power vacuum in which any combination of the three English factions, Royalists, Independents of the New Model Army ("the Army"), and Presbyterians of the English Parliament, as well as the Scottish Parliament allied with the Scottish Presbyterians (the "Kirk"), could prove strong enough to dominate the rest. Armed political Royalism was at an end, but despite being a prisoner, Charles I was considered by himself and his opponents (almost to the last) as necessary to ensure the success of whichever group could come to terms with him. Thus he passed successively into the hands of the Scots, the Parliament and the Army. The King attempted to reverse the verdict of arms by "coquetting" with each in turn. On 3 June 1647, Cornet George Joyce of Thomas Fairfax's horse seized the King for the Army, after which the English Presbyterians and the Scots began to prepare for a fresh civil war, less than two years after the conclusion of the first, this time against "Independency", as embodied in the Army. After making use of the Army's sword, its opponents attempted to disband it, to send it on foreign service and to cut off its arrears of pay. The result was that the Army leadership was exasperated beyond control, and, remembering not merely their grievances but also the principle for which the Army had fought, it soon became the most powerful political force in the realm. From 1646 to 1648 the breach between Army and Parliament widened day by day until finally the Presbyterian party, combined with the Scots and the remaining Royalists, felt itself strong enough to begin a Second Civil War.
Charles I took advantage of the deflection of attention away from himself to negotiate on 28 December 1647 a secret treaty with the Scots, again promising church reform. Under the agreement, called the "Engagement", the Scots undertook to invade England on Charles's behalf and restore him to the throne on condition of the establishment of Presbyterianism within three years.
A series of Royalist uprisings throughout England and a Scottish invasion occurred in the summer of 1648. Forces loyal to Parliament put down most of those in England after little more than a skirmish, but uprisings in Kent, Essex and Cumberland, the rebellion in Wales, and the Scottish invasion involved pitched battles and prolonged sieges.
In the spring of 1648, unpaid Parliamentarian troops in Wales changed sides. Colonel Thomas Horton defeated the Royalist rebels at the Battle of St Fagans (8 May) and the rebel leaders surrendered to Cromwell on 11 July after a protracted two-month siege of Pembroke. Sir Thomas Fairfax defeated a Royalist uprising in Kent at the Battle of Maidstone on 1 June. Fairfax, after his success at Maidstone and the pacification of Kent, turned north to reduce Essex, where, under an ardent, experienced and popular leader, Sir Charles Lucas, the Royalists had taken up arms in great numbers. Fairfax soon drove the enemy into Colchester, but his first attack on the town met with a repulse and he had to settle down to a long siege.
In the North of England, Major-General John Lambert fought a successful campaign against several Royalist uprisings, the largest being that of Sir Marmaduke Langdale in Cumberland. Thanks to Lambert's successes, the Scottish commander, the Duke of Hamilton, had to take a western route through Carlisle in his pro-Royalist Scottish invasion of England. The Parliamentarians under Cromwell engaged the Scots at the Battle of Preston (17–19 August). The battle took place largely at Walton-le-Dale near Preston, Lancashire, and resulted in a victory for Cromwell's troops over the Royalists and Scots commanded by Hamilton. This victory marked the end of the Second English Civil War.
Nearly all the Royalists who had fought in the First Civil War had given their word not to bear arms against Parliament, and many, like Lord Astley, were therefore bound by oath not to take any part in the second conflict. So the victors in the Second Civil War showed little mercy to those who had brought war into the land again. On the evening of the surrender of Colchester, Parliamentarians had Sir Charles Lucas and Sir George Lisle shot. Parliamentary authorities sentenced the leaders of the Welsh rebels, Major-General Rowland Laugharne, Colonel John Poyer and Colonel Rice Powel to death, but executed only Poyer (25 April 1649), having selected him by lot. Of five prominent Royalist peers who had fallen into Parliamentary hands, three – the Duke of Hamilton, the Earl of Holland, and Lord Capel, one of the Colchester prisoners and a man of high character – were beheaded at Westminster on 9 March.
Charles's secret pacts and encouragement of supporters to break their parole caused Parliament to debate whether to return the King to power at all. Those who still supported Charles's place on the throne, such as the army leader and moderate Fairfax, tried again to negotiate with him. The Army, furious that Parliament continued to countenance Charles as a ruler, then marched on Parliament and conducted "Pride's Purge" (named after the commanding officer of the operation, Thomas Pride) in December 1648. Troops arrested 45 members and kept 146 out of the chamber. They allowed only 75 members in, and then only at the Army's bidding. This Rump Parliament received orders to set up, in the name of the people of England, a High Court of Justice for the trial of Charles I for treason. Fairfax, a constitutional monarchist and moderate, declined to have anything to do with the trial. He resigned as head of the army, so clearing Cromwell's road to power.
At the end of the trial the 59 Commissioners (judges) found Charles I guilty of high treason as a "tyrant, traitor, murderer and public enemy". His beheading took place on a scaffold in front of the Banqueting House of the Palace of Whitehall on 30 January 1649. After the Restoration in 1660, nine of the surviving regicides not living in exile were executed and most others sentenced to life imprisonment.
After the regicide, Charles as the eldest son was publicly proclaimed King Charles II in the Royal Square of St. Helier, Jersey, on 17 February 1649 (after a first such proclamation in Edinburgh on 5 February 1649).
Ireland had undergone continual war since the rebellion of 1641, with most of the island controlled by the Irish Confederates. Increasingly threatened by the armies of the English Parliament after Charles I's arrest in 1648, the Confederates signed a treaty of alliance with the English Royalists. The joint Royalist and Confederate forces under the Duke of Ormonde tried to eliminate the Parliamentary army holding Dublin by laying siege, but their opponents routed them at the Battle of Rathmines (2 August 1649). As the former Member of Parliament Admiral Robert Blake blockaded Prince Rupert's fleet in Kinsale, Cromwell could land at Dublin on 15 August 1649 with an army to quell the Royalist alliance.
Cromwell's suppression of the Royalists in Ireland in 1649 is still remembered by many Irish people. After the Siege of Drogheda, the massacre of nearly 3,500 people — around 2,700 Royalist soldiers and 700 others, including civilians, prisoners and Catholic priests (Cromwell claimed all had carried arms) — became one of the historical memories that has driven Irish-English and Catholic-Protestant strife during the last three centuries. The Parliamentarian conquest of Ireland ground on for another four years until 1653, when the last Irish Confederate and Royalist troops surrendered. In the wake of the conquest, the victors confiscated almost all Irish Catholic-owned land and distributed it to Parliament's creditors, to Parliamentary soldiers who served in Ireland, and to English who had settled there before the war.
The execution of Charles I altered the dynamics of the Civil War in Scotland, which had raged between Royalists and Covenanters since 1644. By 1649, the struggle had left the Royalists there in disarray and their erstwhile leader, the Marquess of Montrose, had gone into exile. At first, Charles II encouraged Montrose to raise a Highland army to fight on the Royalist side. However, when the Scottish Covenanters (who did not agree with the execution of Charles I and who feared for the future of Presbyterianism under the new Commonwealth) offered him the crown of Scotland, Charles abandoned Montrose to his enemies. However, Montrose, who had raised a mercenary force in Norway, had already landed and could not abandon the fight. He did not succeed in raising many Highland clans and the Covenanters defeated his army at the Battle of Carbisdale in Ross-shire on 27 April 1650. The victors captured Montrose shortly afterwards and took him to Edinburgh. On 20 May the Scottish Parliament sentenced him to death and had him hanged the next day.
Charles II landed in Scotland at Garmouth in Morayshire on 23 June 1650 and signed the 1638 National Covenant and the 1643 Solemn League and Covenant shortly after coming ashore. With his original Scottish Royalist followers and his new Covenanter allies, Charles II became the greatest threat facing the new English republic. In response to the threat, Cromwell left some of his lieutenants in Ireland to continue the suppression of the Irish Royalists and returned to England.
He arrived in Scotland on 22 July 1650 and proceeded to lay siege to Edinburgh. By the end of August, disease and a shortage of supplies had reduced his army, and he had to order a retreat towards his base at Dunbar. A Scottish army under the command of David Leslie tried to block the retreat, but Cromwell defeated them at the Battle of Dunbar on 3 September. Cromwell's army then took Edinburgh, and by the end of the year his army had occupied much of southern Scotland.
In July 1651, Cromwell's forces crossed the Firth of Forth into Fife and defeated the Scots at the Battle of Inverkeithing (20 July 1651). The New Model Army advanced towards Perth, which allowed Charles, at the head of the Scottish army, to move south into England. Cromwell followed Charles into England, leaving George Monck to finish the campaign in Scotland. Monck took Stirling on 14 August and Dundee on 1 September. The next year, 1652, saw a mopping up of the remnants of Royalist resistance, and under the terms of the "Tender of Union", the Scots received 30 seats in a united Parliament in London, with General Monck as the military governor of Scotland.
Although Cromwell's New Model Army had defeated a Scottish army at Dunbar, Cromwell could not prevent Charles II from marching from Scotland deep into England at the head of another Royalist army. They marched to the west of England where English Royalist sympathies were strongest, but although some English Royalists joined the army, they were far fewer in number than Charles and his Scottish supporters had hoped. Cromwell finally engaged and defeated the new Scottish king at Worcester on 3 September 1651.
After the Royalist defeat at Worcester, Charles II escaped via safe houses and a famous oak tree to France, and Parliament was left in "de facto" control of England. Resistance continued for a time in the Channel Islands, Ireland and Scotland, but with the pacification of England, resistance elsewhere did not threaten the military supremacy of the New Model Army and its Parliamentary paymasters.
During the Wars, the Parliamentarians established a number of successive committees to oversee the war effort. The first, the Committee of Safety set up in July 1642, comprised 15 members of Parliament. After the Anglo-Scottish alliance against the Royalists, the Committee of Both Kingdoms replaced the Committee of Safety between 1644 and 1648. Parliament dissolved the Committee of Both Kingdoms when the alliance ended, but its English members continued to meet as the Derby House Committee. A second Committee of Safety then replaced it.
During the English Civil War, the role of bishops as wielders of political power and upholders of the established church became a matter of heated political controversy. John Calvin formulated a doctrine of Presbyterianism, which held that the offices of "presbyter" and "episkopos" in the New Testament were identical; he rejected the doctrine of apostolic succession. Calvin's follower John Knox brought Presbyterianism to Scotland when the Scottish church was reformed in 1560. In practice, Presbyterianism meant that committees of lay elders had a substantial voice in church government, as opposed to merely being subjects to a ruling hierarchy.
This vision of at least partial democracy in ecclesiology paralleled the struggles between Parliament and the King. A body within the Puritan movement in the Church of England sought to abolish the office of bishop and remake the Church of England along Presbyterian lines. The Martin Marprelate tracts (1588–1589), applying the pejorative name of "prelacy" to the church hierarchy, attacked the office of bishop with satire that deeply offended Elizabeth I and her Archbishop of Canterbury John Whitgift. The vestments controversy also related to this movement, seeking further reductions in church ceremony, and labelling the use of elaborate vestments as "unedifying" and even idolatrous.
King James I, reacting against the perceived contumacy of his Presbyterian Scottish subjects, adopted "No Bishop, no King" as a slogan; he tied the hierarchical authority of the bishop to the absolute authority he sought as King, and viewed attacks on the authority of the bishops as attacks on his authority. Matters came to a head when Charles I appointed William Laud as Archbishop of Canterbury; Laud aggressively attacked the Presbyterian movement and sought to impose the full Book of Common Prayer. The controversy eventually led to Laud's impeachment for treason by a bill of attainder in 1645 and subsequent execution. Charles also attempted to impose episcopacy on Scotland; the Scots' violent rejection of bishops and liturgical worship sparked the Bishops' Wars in 1639–1640.
During the height of Puritan power under the Commonwealth and the Protectorate, episcopacy was formally abolished in the Church of England on 9 October 1646. The Church of England remained Presbyterian until the Restoration of the monarchy under Charles II in 1660.
During the English Civil War, the English overseas possessions became highly involved. In the Channel Islands, the island of Jersey and Castle Cornet in Guernsey supported the King until a surrender with honour in December 1651.
Although the newer, Puritan settlements in North America, notably Massachusetts, were dominated by Parliamentarians, the older colonies sided with the Crown. Friction between Royalists and Puritans in Maryland came to a head in the Battle of the Severn. The Virginia Company's settlements, Bermuda and Virginia, as well as Antigua and Barbados were conspicuous in their loyalty to the Crown. Bermuda's Independent Puritans were expelled, settling the Bahamas under William Sayle as the Eleutheran Adventurers. Parliament passed An Act for prohibiting Trade with the Barbadoes, Virginia, Bermuda and Antego in October, 1650, which stated that
The Act also authorised Parliamentary privateers to act against English vessels trading with the rebellious colonies:
The Parliament began assembling a fleet to invade the Royalist colonies, but many of the English islands in the Caribbean were captured by the Dutch and French in 1651 during the Second Anglo-Dutch War. Far to the North, Bermuda's regiment of Militia and its coastal batteries prepared to resist an invasion that never came. Built-up inside the natural defence of a nearly impassable barrier reef, to fend off the might of Spain, these defences were too powerful for the Parliamentary fleet sent in 1651 under the command of Admiral Sir George Ayscue, which was forced instead to blockade Bermuda for several months 'til the Bermudians negotiated a separate peace that respected the internal status quo. The Parliament of Bermuda avoided the Parliament of England's fate during The Protectorate, becoming one of the oldest continuous legislatures in the world.
Virginia's population swelled with Cavaliers during and after the English Civil War. Even so, Virginia Puritan Richard Bennett was made Governor answering to Cromwell in 1652, followed by two more nominal "Commonwealth Governors". The loyalty of Virginia's Cavaliers to the Crown was rewarded after the 1660 Restoration of the Monarchy when Charles II dubbed it the "Old Dominion".
Figures for casualties during this period are unreliable, but some attempt has been made to provide rough estimates.
In England, a conservative estimate is that roughly 100,000 people died from war-related disease during the three civil wars. Historical records count 84,830 dead from the wars themselves. Counting in accidents and the two Bishops' wars, an estimate of 190,000 dead is achieved, out of a total population of about five million.
Figures for Scotland are less reliable and should be treated with caution. Casualties include the deaths of prisoners-of-war in conditions that accelerated their deaths, with estimates of 10,000 prisoners not surviving or not returning home (8,000 captured during and immediately after the Battle of Worcester were deported to New England, Bermuda and the West Indies to work for landowners as indentured labourers). There are no figures to calculate how many died from war-related diseases, but if the same ratio of disease to battle deaths from English figures is applied to the Scottish figures, a not unreasonable estimate of 60,000 people is achieved, from a population of about one million.
Figures for Ireland are described as "miracles of conjecture". Certainly the devastation inflicted on Ireland was massive, with the best estimate provided by Sir William Petty, the father of English demography. Petty estimated that 112,000 Protestants and 504,000 Catholics were killed through plague, war and famine, giving an estimated total of 616,000 dead, out of a pre-war population of about one and a half million. Although Petty's figures are the best available, they are still acknowledged as tentative; they do not include an estimated 40,000 driven into exile, some of whom served as soldiers in European continental armies, while others were sold as indentured servants to New England and the West Indies. Many of those sold to landowners in New England eventually prospered, but many sold to landowners in the West Indies were worked to death.
These estimates indicate that England suffered a 4 percent loss of population, Scotland a loss of 6 percent, while Ireland suffered a loss of 41 percent of its population. Putting these numbers into the context of other catastrophes helps to understand the devastation of Ireland in particular. The Great Hunger of 1845–1852 resulted in a loss of 16 percent of the population, while during the Second World War the population of the Soviet Union fell by 16 percent.
Ordinary people took advantage of the dislocation of civil society in the 1640s to gain personal advantages. The contemporary guild democracy movement won its greatest successes among London's transport workers, notably the Thames watermen. Rural communities seized timber and other resources on the sequestrated estates of Royalists and Catholics, and on the estates of the royal family and church hierarchy. Some communities improved their conditions of tenure on such estates. The old "status quo" began a retrenchment after the end of the First Civil War in 1646, and more especially after the Restoration in 1660, but some gains were long-term. The democratic element introduced into the watermen's company in 1642, for example, survived with vicissitudes until 1827.
The wars left England, Scotland, and Ireland among the few countries in Europe without a monarch. In the wake of victory, many of the ideals (and many idealists) became sidelined. The republican government of the Commonwealth of England ruled England (and later all of Scotland and Ireland) from 1649 to 1653 and from 1659 to 1660. Between the two periods, and due to in-fighting among various factions in Parliament, Oliver Cromwell ruled over the Protectorate as Lord Protector (effectively a military dictator) until his death in 1658.
On Oliver Cromwell's death, his son Richard became Lord Protector, but the Army had little confidence in him. After seven months the Army removed Richard, and in May 1659 it re-installed the Rump. However, since the Rump Parliament acted as though nothing had changed since 1653 and as though it could treat the Army as it liked, military force shortly afterward dissolved this as well. After the second dissolution of the Rump, in October 1659, the prospect of a total descent into anarchy loomed as the Army's pretense of unity finally dissolved into factions.
Into this atmosphere General George Monck, Governor of Scotland under the Cromwells, marched south with his army from Scotland. On 4 April 1660, in the Declaration of Breda, Charles II made known the conditions of his acceptance of the Crown of England. Monck organised the Convention Parliament, which met for the first time on 25 April 1660. On 8 May 1660, it declared that Charles II had reigned as the lawful monarch since the execution of Charles I in January 1649. Charles returned from exile on 23 May 1660. On 29 May 1660, the populace in London acclaimed him as king. His coronation took place at Westminster Abbey on 23 April 1661. These events became known as the "Restoration".
Although the monarchy was restored, it was still only with the consent of Parliament. So the civil wars effectively set England and Scotland on course towards a parliamentary monarchy form of government. The outcome of this system was that the future Kingdom of Great Britain, formed in 1707 under the Acts of Union, managed to forestall the kind of revolution typical of European republican movements which generally resulted in total abolition of monarchy. Thus the United Kingdom was spared the wave of revolutions that occurred in Europe in the 1840s. Specifically, future monarchs became wary of pushing Parliament too hard, and Parliament effectively chose the line of royal succession in 1688 with the Glorious Revolution and in the 1701 Act of Settlement.
In the early decades of the 20th century, the Whig school was the dominant theoretical view. It explained the Civil War as resulting from centuries of struggle between Parliament (notably the House of Commons) and the Monarchy, with Parliament defending the traditional rights of Englishmen, while the Stuart monarchy continually attempted to expand its right to dictate law arbitrarily. The major Whig historian, S. R. Gardiner, popularised the idea that the English Civil War was a "Puritan Revolution", which challenged the repressive Stuart Church and prepared the way for religious toleration. So Puritanism was seen as the natural ally of a people preserving their traditional rights against arbitrary monarchical power.
The Whig view was challenged and largely superseded by the Marxist school, which became popular in the 1940s, and saw the English Civil War as a bourgeois revolution. According to Marxist historian Christopher Hill:
In the 1970s, revisionist historians challenged both the Whig and the Marxist theories, notably in the 1973 anthology "The Origins of the English Civil War" (Conrad Russell ed.). These historians focused on the minutiae of the years immediately before the civil war, returning to the contingency-based historiography of Clarendon's famous "History of the Rebellion and Civil Wars in England". This, it was claimed, demonstrated that patterns of war allegiance did not fit either Whig or Marxist theories. Parliament was not inherently progressive, nor the events of 1640 a precursor for the Glorious Revolution. Furthermore, Puritans did not necessarily ally themselves with Parliamentarians. Many members of the bourgeoisie fought for the King, while many landed aristocrats supported Parliament.
From the 1990s, a number of historians replaced the historical title "English Civil War" with "Wars of the Three Kingdoms" and "British Civil Wars", positing that the civil war in England cannot be understood apart from events in other parts of Britain and Ireland. King Charles I remains crucial, not just as King of England, but through his relationship with the peoples of his other realms. For example, the wars began when Charles forced an Anglican Prayer Book upon Scotland, and when this was met with resistance from the Covenanters, he needed an army to impose his will. However, this need of military funds forced Charles I to call an English Parliament, which was not willing to grant the needed revenue unless he addressed their grievances. By the early 1640s, Charles was left in a state of near-permanent crisis management, confounded by the demands of the various factions. For example, Charles finally made terms with the Covenanters in August 1641, but although this might have weakened the position of the English Parliament, the Irish Rebellion of 1641 broke out in October 1641, largely negating the political advantage he had obtained by relieving himself of the cost of the Scottish invasion.
Thomas Hobbes gave a much earlier historical account of the English Civil War in his "Behemoth", written in 1668 and published in 1681. He assessed the causes of the war to be the conflicting political doctrines of the time. "Behemoth" offered a uniquely historical and philosophical approach to naming the catalysts for the war. It also attempted to explain why Charles I could not hold his throne and maintain peace in his kingdom.
Hobbes analysed in turn the following aspects of English thought during the war: the opinions of divinity and politics that spurred rebellion; rhetoric and doctrine used by the rebels against the king; and how opinions about "taxation, the conscription of soldiers, and military strategy" affected the outcomes of battles and shifts of sovereignty.
Hobbes attributed the war to the novel theories of intellectuals and divines spread for their own pride of reputation. He held that clerical pretensions had contributed significantly to the troubles — "whether those of puritan fundamentalists, papal supremacists or divine right Episcopalians". Hobbes wanted to abolish the independence of the clergy and bring it under the control of the civil state.
Some scholars suggest that "Behemoth" has not received its due as an academic work, being comparatively overlooked and under-rated in the shadow of Hobbes' "Leviathan". Its scholarly reputation may have suffered because it takes the form of a dialogue, which, while common in philosophy, is rarely adopted by historians. Other factors that hindered its success include Charles II's refusing its publication and Hobbes' lack of empathy with views different from his own.
Two large historical societies exist, The Sealed Knot and The English Civil War Society, which regularly re-enact events and battles of the Civil War in full period costume.
Attribution: | https://en.wikipedia.org/wiki?curid=9709 |
Elementary algebra
Elementary algebra encompasses some of the basic concepts of algebra, one of the main branches of mathematics. It is typically taught to secondary school students and builds on their understanding of arithmetic. Whereas arithmetic deals with specified numbers, algebra introduces quantities without fixed values, known as variables. This use of variables entails use of algebraic notation and an understanding of the general rules of the operators introduced in arithmetic. Unlike abstract algebra, elementary algebra is not concerned with algebraic structures outside the realm of real and complex numbers.
The use of variables to denote quantities allows general relationships between quantities to be formally and concisely expressed, and thus enables solving a broader scope of problems. Many quantitative relationships in science and mathematics are expressed as algebraic equations.
Algebraic notation describes the rules and conventions for writing mathematical expressions, as well as the terminology used for talking about parts of expressions. For example, the expression formula_1 has the following components:
A "coefficient" is a numerical value, or letter representing a numerical constant, that multiplies a variable (the operator is omitted). A "term" is an addend or a summand, a group of coefficients, variables, constants and exponents that may be separated from the other terms by the plus and minus operators. Letters represent variables and constants. By convention, letters at the beginning of the alphabet (e.g. formula_2) are typically used to represent constants, and those toward the end of the alphabet (e.g. formula_3 and ) are used to represent variables. They are usually written in italics.
Algebraic operations work in the same way as arithmetic operations, such as addition, subtraction, multiplication, division and exponentiation. and are applied to algebraic variables and terms. Multiplication symbols are usually omitted, and implied when there is no space between two variables or terms, or when a coefficient is used. For example, formula_4 is written as formula_5, and formula_6 may be written formula_7.
Usually terms with the highest power (exponent), are written on the left, for example, formula_8 is written to the left of . When a coefficient is one, it is usually omitted (e.g. formula_9 is written formula_8). Likewise when the exponent (power) is one, (e.g. formula_11 is written formula_12). When the exponent is zero, the result is always 1 (e.g. formula_13 is always rewritten to ). However formula_14, being undefined, should not appear in an expression, and care should be taken in simplifying expressions in which variables may appear in exponents.
Other types of notation are used in algebraic expressions when the required formatting is not available, or can not be implied, such as where only letters and symbols are available. As an illustration of this, while exponents are usually formatted using superscripts, e.g., formula_8, in plain text, and in the TeX mark-up language, the caret symbol "^" represents exponentiation, so formula_8 is written as "x^2"., as well as some programming languages such as Lua. In programming languages such as Ada, Fortran, Perl, Python and Ruby, a double asterisk is used, so formula_8 is written as "x**2". Many programming languages and calculators use a single asterisk to represent the multiplication symbol, and it must be explicitly used, for example, formula_12 is written "3*x".
Elementary algebra builds on and extends arithmetic by introducing letters called variables to represent general (non-specified) numbers. This is useful for several reasons.
Algebraic expressions may be evaluated and simplified, based on the basic properties of arithmetic operations (addition, subtraction, multiplication, division and exponentiation). For example,
An equation states that two expressions are equal using the symbol for equality, (the equals sign). One of the best-known equations describes Pythagoras' law relating the length of the sides of a right angle triangle:
This equation states that formula_39, representing the square of the length of the side that is the hypotenuse, the side opposite the right angle. That is equal to the sum (addition) of the squares of the other two sides whose lengths are represented by and .
An equation is the claim that two expressions have the same value and are equal. Some equations are true for all values of the involved variables (such as formula_40); such equations are called identities. Conditional equations are true for only some values of the involved variables, e.g. formula_41 is true only for formula_42 and formula_43. The values of the variables which make the equation true are the solutions of the equation and can be found through equation solving.
Another type of equation is inequality. Inequalities are used to show that one side of the equation is greater, or less, than the other. The symbols used for this are: formula_44 where formula_45 represents 'greater than', and formula_46 where formula_47 represents 'less than'. Just like standard equality equations, numbers can be added, subtracted, multiplied or divided. The only exception is that when multiplying or dividing by a negative number, the inequality symbol must be flipped.
By definition, equality is an equivalence relation, meaning it has the properties (a) reflexive (i.e. formula_48), (b) symmetric (i.e. if formula_49 then formula_50) (c) transitive (i.e. if formula_49 and formula_52 then formula_53). It also satisfies the important property that if two symbols are used for equal things, then one symbol can be substituted for the other in any true statement about the first and the statement will remain true. This implies the following properties:
The relations "less than" formula_47 and greater than formula_45 have the property of transitivity:
By reversing the inequation, formula_47 and formula_45 can be swapped, for example:
Substitution is replacing the terms in an expression to create a new expression. Substituting 3 for in the expression makes a new expression with meaning . Substituting the terms of a statement makes a new statement. When the original statement is true independently of the values of the terms, the statement created by substitutions is also true. Hence, definitions can be made in symbolic terms and interpreted through substitution: if formula_81 is meant as the definition of formula_82 as the product of with itself, substituting for informs the reader of this statement that formula_83 means . Often it's not known whether the statement is true independently of the values of the terms. And, substitution allows one to derive restrictions on the possible values, or show what conditions the statement holds under. For example, taking the statement , if is substituted with , this implies , which is false, which implies that if then cannot be .
If and are integers, rationals, or real numbers, then implies or . Consider . Then, substituting for and for , we learn or . Then we can substitute again, letting and , to show that if then or . Therefore, if , then or ( or ), so implies or or .
If the original fact were stated as " implies or ", then when saying "consider ," we would have a conflict of terms when substituting. Yet the above logic is still valid to show that if then or or if, instead of letting and , one substitutes for and for (and with , substituting for and for ). This shows that substituting for the terms in a statement isn't always the same as letting the terms from the statement equal the substituted terms. In this situation it's clear that if we substitute an expression into the term of the original equation, the substituted does not refer to the in the statement " implies or ."
The following sections lay out examples of some of the types of algebraic equations that may be encountered.
Linear equations are so-called, because when they are plotted, they describe a straight line. The simplest equations to solve are linear equations that have only one variable. They contain only constant numbers and a single variable without an exponent. As an example, consider:
To solve this kind of equation, the technique is add, subtract, multiply, or divide both sides of the equation by the same number in order to isolate the variable on one side of the equation. Once the variable is isolated, the other side of the equation is the value of the variable. This problem and its solution are as follows:
In words: the child is 4 years old.
The general form of a linear equation with one variable, can be written as: formula_85
Following the same procedure (i.e. subtract from both sides, and then divide by ), the general solution is given by formula_86
A linear equation with two variables has many (i.e. an infinite number of) solutions. For example:
That cannot be worked out by itself. If the son's age was made known, then there would no longer be two unknowns (variables). The problem then becomes a linear equation with just one variable, that can be solved as described above.
To solve a linear equation with two variables (unknowns), requires two related equations. For example, if it was also revealed that:
Now there are two related linear equations, each with two unknowns, which enables the production of a linear equation with just one variable, by subtracting one from the other (called the elimination method):
In other words, the son is aged 12, and since the father 22 years older, he must be 34. In 10 years, the son will be 22, and the father will be twice his age, 44. This problem is illustrated on the associated plot of the equations.
For other ways to solve this kind of equations, see below, System of linear equations.
A quadratic equation is one which includes a term with an exponent of 2, for example, formula_30, and no term with higher exponent. The name derives from the Latin "quadrus", meaning square. In general, a quadratic equation can be expressed in the form formula_92, where is not zero (if it were zero, then the equation would not be quadratic but linear). Because of this a quadratic equation must contain the term formula_93, which is known as the quadratic term. Hence formula_94, and so we may divide by and rearrange the equation into the standard form
where formula_96 and formula_97. Solving this, by a process known as completing the square, leads to the quadratic formula
where the symbol "±" indicates that both
are solutions of the quadratic equation.
Quadratic equations can also be solved using factorization (the reverse process of which is expansion, but for two linear terms is sometimes denoted foiling). As an example of factoring:
which is the same thing as
It follows from the zero-product property that either formula_102 or formula_103 are the solutions, since precisely one of the factors must be equal to zero. All quadratic equations will have two solutions in the complex number system, but need not have any in the real number system. For example,
has no real number solution since no real number squared equals −1.
Sometimes a quadratic equation has a root of multiplicity 2, such as:
For this equation, −1 is a root of multiplicity 2. This means −1 appears twice, since the equation can be rewritten in factored form as
All quadratic equations have exactly two solutions in complex numbers (but they may be equal to each other), a category that includes real numbers, imaginary numbers, and sums of real and imaginary numbers. Complex numbers first arise in the teaching of quadratic equations and the quadratic formula. For example, the quadratic equation
has solutions
Since formula_109 is not any real number, both of these solutions for "x" are complex numbers.
An exponential equation is one which has the form formula_110 for formula_111, which has solution
when formula_113. Elementary algebraic techniques are used to rewrite a given equation in the above way before arriving at the solution. For example, if
then, by subtracting 1 from both sides of the equation, and then dividing both sides by 3 we obtain
whence
or
A logarithmic equation is an equation of the form formula_118 for formula_111, which has solution
For example, if
then, by adding 2 to both sides of the equation, followed by dividing both sides by 4, we get
whence
from which we obtain
A radical equation is one that includes a radical sign, which includes square roots, formula_125 cube roots, formula_126, and "n"th roots, formula_127. Recall that an "n"th root can be rewritten in exponential format, so that formula_127 is equivalent to formula_129. Combined with regular exponents (powers), then formula_130 (the square root of cubed), can be rewritten as formula_131. So a common form of a radical equation is formula_132 (equivalent to formula_133) where and are integers. It has real solution(s):
For example, if:
then
and thus
There are different methods to solve a system of linear equations with two variables.
An example of solving a system of linear equations is by using the elimination method:
Multiplying the terms in the second equation by 2:
Adding the two equations together to get:
which simplifies to
Since the fact that formula_102 is known, it is then possible to deduce that formula_143 by either of the original two equations (by using "2" instead of ) The full solution to this problem is then
This is not the only way to solve this specific system; could have been resolved before .
Another way of solving the same system of linear equations is by substitution.
An equivalent for can be deduced by using one of the two equations. Using the second equation:
Subtracting formula_147 from each side of the equation:
and multiplying by −1:
Using this value in the first equation in the original system:
Adding "2" on each side of the equation:
which simplifies to
Using this value in one of the equations, the same solution as in the previous method is obtained.
This is not the only way to solve this specific system; in this case as well, could have been solved before .
In the above example, a solution exists. However, there are also systems of equations which do not have any solution. Such a system is called inconsistent. An obvious example is
As 0≠2, the second equation in the system has no solution. Therefore, the system has no solution.
However, not all inconsistent systems are recognized at first sight. As an example, consider the system
Multiplying by 2 both sides of the second equation, and adding it to the first one results in
which clearly has no solution.
There are also systems which have infinitely many solutions, in contrast to a system with a unique solution (meaning, a unique pair of values for and ) For example:
Isolating in the second equation:
And using this value in the first equation in the system:
The equality is true, but it does not provide a value for . Indeed, one can easily verify (by just filling in some values of ) that for any there is a solution as long as formula_160. There is an infinite number of solutions for this system.
Systems with more variables than the number of linear equations are called underdetermined. Such a system, if it has any solutions, does not have a unique one but rather an infinitude of them. An example of such a system is
When trying to solve it, one is led to express some variables as functions of the other ones if any solutions exist, but cannot express "all" solutions numerically because there are an infinite number of them if there are any.
A system with a higher number of equations than variables is called overdetermined. If an overdetermined system has any solutions, necessarily some equations are linear combinations of the others. | https://en.wikipedia.org/wiki?curid=9710 |
Excalibur
Excalibur () is the legendary sword of King Arthur, sometimes also attributed with magical powers or associated with the rightful sovereignty of Britain. Excalibur and the Sword in the Stone (the proof of Arthur's lineage) are sometimes said to be the same weapon, but in most versions they are considered separate. Excalibur was associated with the Arthurian legend very early on. In Welsh, it is called "Caledfwlch"; in Cornish, "Calesvol" (in Modern Cornish: "Kalesvolgh"); in Breton, "Kaledvoulc'h"; and in Latin, "Caliburnus".
The name "Excalibur" ultimately derives from the Welsh "Caledfwlch" (and Breton "Kaledvoulc'h", Middle Cornish "Calesvol"), which is a compound of ' "hard" and ' "breach, cleft". Caledfwlch appears in several early Welsh works, including the prose tale "Culhwch and Olwen". The name was later used in Welsh adaptations of foreign material such as the "Brut"s (chronicles), which were based on Geoffrey of Monmouth. It is often considered to be related to the phonetically similar "Caladbolg", a sword borne by several figures from Irish mythology, although a borrowing of "Caledfwlch" from Irish "Caladbolg" has been considered unlikely by Rachel Bromwich and D. Simon Evans. They suggest instead that both names "may have similarly arisen at a very early date as generic names for a sword". This sword then became exclusively the property of Arthur in the British tradition.
Geoffrey of Monmouth, in his "Historia Regum Britanniae" ("The History of the Kings of Britain", c. 1136), Latinised the name of Arthur's sword as "Caliburnus" (potentially influenced by the Medieval Latin spelling "calibs" of Classical Latin "chalybs", from Greek "chályps" ["χάλυψ"] "steel"). Most Celticists consider Geoffrey's "Caliburnus" to be derivative of a lost Old Welsh text in which "bwlch" (Old Welsh "bulc[h]") had not yet been lenited to "fwlch" (Middle Welsh "vwlch" or "uwlch"). In the late 15th/early 16th-century Middle Cornish play "Beunans Ke", Arthur's sword is called "Calesvol", which is etymologically an exact Middle Cornish cognate of the Welsh "Caledfwlch". It is unclear if the name was borrowed from the Welsh (if so, it must have been an early loan, for phonological reasons), or represents an early, pan-Brittonic traditional name for Arthur's sword.
In Old French sources this then became "Escalibor", "Excalibor", and finally the familiar "Excalibur". Geoffrey Gaimar, in his Old French "L'Estoire des Engleis" (1134-1140), mentions Arthur and his sword: "this Constantine was the nephew of Arthur, who had the sword Caliburc" (""Cil Costentin, li niès Artur, Ki out l'espée Caliburc""). In Wace's "Roman de Brut" (c. 1150-1155), an Old French translation and versification of Geoffrey's "Historia", the sword is called "Calabrum", "Callibourc", "Chalabrun", and "Calabrun" (with variant spellings such as "Chalabrum", "Calibore", "Callibor", "Caliborne", "Calliborc", and "Escaliborc", found in various manuscripts of the "Brut").
In Chrétien de Troyes' late 12th-century Old French "Perceval", Arthur's knight Gawain carries the sword "Escalibor" and it is stated, "for at his belt hung Escalibor, the finest sword that there was, which sliced through iron as through wood" (""Qu'il avoit cainte Escalibor, la meillor espee qui fust, qu'ele trenche fer come fust""). This statement was probably picked up by the author of the "Estoire Merlin", or Vulgate Merlin, where the author (who was fond of fanciful folk etymologies) asserts that Escalibor "is a Hebrew name which means in French 'cuts iron, steel, and wood'" (""c'est non Ebrieu qui dist en franchois trenche fer & achier et fust""; note that the word for "steel" here, achier, also means "blade" or "sword" and comes from medieval Latin "aciarium", a derivative of "acies" "sharp", so there is no direct connection with Latin "chalybs" in this etymology). It is from this fanciful etymological musing that Thomas Malory got the notion that Excalibur meant "cut steel" (""the name of it,' said the lady, 'is Excalibur, that is as moche to say, as Cut stele"").
In Arthurian romance, a number of explanations are given for Arthur's possession of Excalibur. In Robert de Boron's "Merlin", the first tale to mention the "sword in the stone" motif c. 1200, Arthur obtained the British throne by pulling a sword from an anvil sitting atop a stone that appeared in a churchyard on Christmas Eve. In this account, as foretold by Merlin, the act could not be performed except by "the true king," meaning the divinely appointed king or true heir of Uther Pendragon. As Malory related in his most famous English-language version of the Arthurian tales, the 15th-century "Le Morte d'Arthur": ""Whoso pulleth out this sword of this stone and anvil, is rightwise king born.""
The identity of this sword as Excalibur is made explicit in the Prose "Merlin", part of the Lancelot-Grail cycle of French romances (the Vulgate Cycle). In the Vulgate "Mort Artu", when Arthur at the brink of death he orders Griflet to throw the sword into the enchanted lake; after two failed attempts (as he felt such a great sword should not be thrown away), Griflet finally complies with the wounded king's request and a hand emerges from the lake to catch it. This tale becomes attached to Bedivere instead of Griflet in Malory and the English tradition.
However, in the Post-Vulgate Cycle and consequently Malory, early in his reign Arthur breaks the Sword from the Stone known as Caliburn while in combat against King Pellinore, and is given Excalibur by a Lady of the Lake in exchange for a later boon for her (some time later, she arrives at Arthur's court to demand the head of Sir Balin). Malory records both versions of the legend in his "Le Morte d'Arthur", naming both swords as Excalibur.
In Welsh legends, Arthur's sword is known as "Caledfwlch". In "Culhwch and Olwen", it is one of Arthur's most valuable possessions and is used by Arthur's warrior Llenlleawg the Irishman to kill the Irish king Diwrnach while stealing his magical cauldron. Irish mythology mentions a weapon "Caladbolg", the sword of Fergus mac Róich, which was also known for its incredible power and was carried by some of Ireland's greatest heroes. The name, which can also mean "hard cleft" in Irish, appears in the plural, "caladbuilc", as a generic term for "great swords" in "Togail Troi" ("The Destruction of Troy"), a 10th-century Irish translation of the classical tale.
Though not named as Caledfwlch, Arthur's sword is described vividly in "The Dream of Rhonabwy", one of the tales associated with the "Mabinogion" (as translated by Jeffrey Gantz): "Then they heard Cadwr Earl of Cornwall being summoned, and saw him rise with Arthur's sword in his hand, with a design of two chimeras on the golden hilt; when the sword was unsheathed what was seen from the mouths of the two chimeras was like two flames of fire, so dreadful that it was not easy for anyone to look."
Geoffrey's "Historia" is the first non-Welsh source to speak of the sword. Geoffrey says the sword was forged in Avalon and Latinises the name "Caledfwlch" as "Caliburnus". When his influential pseudo-history made it to Continental Europe, writers altered the name further until it finally took on the popular form "Excalibur" (various spellings in the medieval Arthurian romance and chronicle tradition include: "Calabrun", "Calabrum", "Calibourne", "Callibourc", "Calliborc", "Calibourch", "Escaliborc", and "Escalibor"). The legend was expanded upon in the Vulgate Cycle and in the Post-Vulgate Cycle which emerged in its wake. Both included the work known as the "Prose Merlin", but the Post-Vulgate authors left out the "Merlin" continuation from the earlier cycle, choosing to add an original account of Arthur's early days including a new origin for Excalibur.
In several early French works, such as Chrétien de Troyes' "Perceval, the Story of the Grail" and the Vulgate "Lancelot Proper" section, Excalibur is used by Gawain, Arthur's nephew and one of his best knights. This is in contrast to later versions, where Excalibur belongs solely to the king. The challenge of drawing a sword from a stone also appears in the later Arthurian stories of Galahad, whose achievement of the task indicates that he is destined to find the Holy Grail.
In many versions, Excalibur's blade was engraved with phrases on opposite sides: "Take me up" and "Cast me away" (or similar). In addition, when Excalibur was first drawn, in the first battle testing Arthur's sovereignty, its blade blinded his enemies. Malory writes: ""thenne he drewe his swerd Excalibur, but it was so breyght in his enemyes eyen that it gaf light lyke thirty torchys.""
In some tellings, Excalibur's scabbard was said to have powers of its own, as the one wearing it would not lose even a drop of blood. Any wounds received while wearing the scabbard would not bleed at all, thus preventing the death of the wearer. For this reason, Merlin chides Arthur for preferring the sword over the scabbard, saying that the latter was the greater treasure. In the later romance tradition, including "Le Morte d'Arthur", the scabbard is stolen from Arthur by his half-sister Morgan le Fay in revenge for the death of her beloved Accolon and thrown into a lake, never to be found again. This act later enables the death of Arthur at the Battle of Camlann.
Other weapons have been associated with Arthur. Welsh tradition also knew of a dagger named Carnwennan and a spear named Rhongomyniad that belonged to him. Carnwennan ("little white-hilt") first appears in "Culhwch and Olwen", where Arthur uses it to slice the witch Orddu in half. Rhongomyniad ("spear" + "striker, slayer") is also mentioned in "Culhwch", although only in passing; it appears as simply "Ron" ("spear") in Geoffrey's "Historia". Geoffrey also names Arthur's shield as "Pridwen", but in "Culhwch", "Prydwen" ("fair face") is the name of Arthur's ship while his shield is named "Wynebgwrthucher" ("face of evening").
The Alliterative "Morte Arthure", a Middle English poem, mentions Clarent, a sword of peace meant for knighting and ceremonies as opposed to battle, which Mordred stole and then used to kill Arthur at Camlann. The Prose "Lancelot" of the Vulgate Cycle mentions a sword called Seure, or Secace in some manuscripts, which belonged to the king but was used by Lancelot in one battle.
In Welsh mythology, the "Dyrnwyn" ("White-Hilt"), one of the Thirteen Treasures of the Island of Britain, is said to be a powerful sword belonging to Rhydderch Hael, one of the Three Generous Men of Britain mentioned in the Welsh Triads. When drawn by a worthy or well-born man, the entire blade would blaze with fire. Rhydderch was never reluctant to hand the weapon to anyone, hence his nickname Hael "the Generous", but the recipients, as soon as they had learned of its peculiar properties, always rejected the sword.
There are other similar weapons described in other mythologies. In particular, Claíomh Solais, which is an Irish term meaning "Sword of Light", or "Shining Sword", appears in a number of orally transmitted Irish folk-tales. The Sword in the Stone has an analogue in some versions of the story of Sigurd, whose father, Sigmund, draws the sword Gram out of the tree Barnstokkr where it is embedded by the Norse god Odin. A sword in the stone legend is also associated with the 12th-century Italian Saint Galgano. | https://en.wikipedia.org/wiki?curid=9717 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.