DirtyText
stringlengths
3
1.42M
CleanText
stringlengths
5
225k
{{distinguish|Leonardo Vinci}} {{Infobox Artist | name = Leonardo da Vinci | image = Francesco Melzi - Portrait of Leonardo - WGA14795.jpg<!-- Don't change this image. This was drawn by his pupil. It is the ONLY certain likeness. --> | caption = Portrait of Leonardo by Melzi | birth_name = Leonardo di ser Piero da Vinci | birth_date = {{birth date|1452|4|15|df=yes}} | birth_place = Vinci, Republic of [[Florence]] | death_date = {{death date and age|1519|5|2|1452|4|15|df=yes}} | death_place = Amboise, Kingdom of [[France]] | nationality = Italian | field = Arts and sciences | training = Verrocchio | movement = [[Renaissance|High Renaissance]] | works = {{plainlist|All 8 of his paintings which include: * ''[[The Adoration of the Magi]]''<br/> (by {{circa}} 1482) * ''[[Saint Jerome in the Wilderness]]''<br/> (by {{circa}} 1490) * ''[[Virgin of the Rocks]]'' (Louvre version)<br/> (by {{circa}} 1493) * ''[[The Last Supper]]''<br/> (by {{circa}} 1498) * ''[[Salla delle Asse]]''<br/> (by {{circa}} 1499) * ''[[The Virgin and Child with Saint Anne and Saint John the Baptist]]'' (by {{circa}} 1508) * ''[[Mona Lisa]]''<br/> (by {{circa}} 1517) * ''[[The Virgin and Child with Saint Anne]]''<br/> (by {{circa}} 1519) }} | patrons = King of France | awards = }} '''Leonardo Da Vinci''' (15 April 1452 – 2 May 1519) was an [[Italy|Italian]] [[polymath]] who lived during the [[Renaissance]]. He is famous for his [[painting]]s.<ref name="genius">Vasari, Boltraffio, Castiglione, "Anonimo" Gaddiano, Berensen, Taine, Fuseli, Rio, Bortolon, etc as quoted in della Chiesa, see Bibliography</ref> He was also a [[scientist]], [[mathematician]], [[engineer]], [[inventor]], [[anatomy|anatomist]], [[sculptor]], [[architect]], [[botanist]], [[musician]], and [[writer]]. Leonardo wanted to know everything about [[nature]], and wanted to know how everything worked. He was [[renaissance man|very good at studying]], as well as designing and making all sorts of inventions.<ref name="HG">{{citation| first = Helen| last = Gardner| title = Art through the Ages| year = 1970| publisher = Harcourt, Brace and World}}</ref> The [[Art history|art historian]] Helen Gardner said that no one has ever been quite like him because he was interested in so many things that he seems to have had the mind of a giant, and yet what he was like as a person is still a mystery.<ref>"...His mind and personality seem to us superhuman, the man himself mysterious and remote". (Helen Gardner)</ref> Leonardo was born in [[Vinci, Tuscany|Vinci]], a small town near [[Florence]], [[Italy]]. He was trained to be an artist by the sculptor and painter [[Andrea del Verrocchio|Verrocchio]]. He spent most of his life working for rich Italian [[Nobility|noblemen]]. In his last years, he lived in a expensive home given to him by the [[Francis I of France|King of France]]. Two of his paintings are among the best-known in the world: the ''[[Mona Lisa]]'' and ''[[The Last Supper]]''. He did many drawings. His best-known drawing is ''[[Vitruvian Man]]''. Leonardo often thought of new [[invention]]s. He kept notebooks with notes and drawings of these ideas. Most of his inventions were never made. Some of his ideas were a [[helicopter]], a [[tank]], a [[calculator]], a [[parachute]], a [[robot]], a [[telephone]], [[evolution]], and [[solar energy|solar power]]. [[File:Why was Leonardo da Vinci that famous?.webm|thumb|260x260px|A simple video presentation of Leonardo da Vinci.]] == Life == === Childhood, 1452–1466 === [[File:Study of a Tuscan Landscape.jpg|thumb|Leonardo's earliest known drawing, the Arno Valley, 1473. It is in the [[Uffizi|Uffizi Gallery]].|259x259px]] Leonardo was born on the 15th of April, 1452,<ref name="notablebiographies.com">Leonardo da Vinci Biography,[http://www.notablebiographies.com/Ki-Lo/Leonardo-da-Vinci.html] accessdate: December 26, 2015</ref> in the [[Tuscany|Tuscan]] hill town of Vinci, in the valley of the [[Arno|Arno River]]. His grandfather, Antonio da Vinci, wrote down the details of the birth. Leonardo's parents were not married. His father was a [[lawyer|Notary]], Ser Piero da Vinci.<ref name=AV> {{citation | first = Alessandro | last = Vezzosi | title = Leonardo da Vinci: Renaissance Man }}</ref><ref name=Chiesa>Angela Ottino della Chiesa, ''The Complete Paintings of Leonardo da Vinci''</ref> His mother, Caterina, was a servant. She may have been a [[slavery|slave]] from the [[Middle East]],<ref>According to Alessandro Vezzosi, Head of the Leonardo Museum in Vinci, Piero may have owned a Middle Eastern slave called Caterina. A study of Leonardo's fingerprint suggests that he may have had Middle Eastern blood.[http://www.foxnews.com/wires/2006Dec01/0,4670,LeonardoapossFingerprint,00.html Experts Reconstruct Leonardo Fingerprint]" December 12, 2001</ref><ref> {{citation | url = https://www.washingtonpost.com/wp-dyn/content/article/2006/12/01/AR2006120100961_pf.html | title = Experts Reconstruct Leonardo Fingerprint | publisher = The Associated Press | accessdate = 2007-12-14 }}</ref> or from China.<ref>Angelo Paratico ''Leonardo Da Vinci. A Chinese Scholar Lost in Renaissance Italy'' Lascar Publishing, Hong Kong, 2015.{{ISBN|9881419808}}</ref> His father later took custody of Leonardo, and his mother remarried and had 5 more children.<ref>{{Cite web |url=http://legacy.mos.org/leonardo/bio.html |title=Renaissance Man |publisher=legacy.mos.org |access-date=2015-04-23 |archive-date=2015-04-18 |archive-url=https://web.archive.org/web/20150418010743/http://legacy.mos.org/Leonardo/bio.html |url-status=dead }}</ref> Leonardo's full name was "Leonardo di ser Piero da Vinci", which means "Leonardo, the son of Messer (Mister) Piero da Vinci". Leonardo spent his first five years living in a farmhouse with his mother. After that, he lived in Vinci with his father, his father's new wife Albiera, his grandparents, and uncle Francesco.<ref name= LB /> When Leonardo grew up, he only wrote down two memories from his childhood. He remembered that when he was lying outside in his cradle, a large bird flew from the sky and hovered over him. Its tail feathers brushed his face.<ref name= LB> {{citation | first = Liana | last = Bortolon | title = The Life and Times of Leonardo | publisher = Paul Hamlyn | publication-place=London | year = 1967 }}</ref> Leonardo's other important memory was when he found a cave while exploring in the mountains. He was terrified that some great [[monster]] might be hiding there, but he was also very excited and wanted to find out what was inside.<ref name= LB /> Leonardo started painting major works when he was about 29 or 30 years old. His first major painting was the [[Adoration of the Magi]], and his last major painting was [[The Virgin and Child with Saint Anne]], painted when he was about 66 or 67 years old. [[Giorgio Vasari]] wrote about Leonardo's life shortly after his death. He tells many interesting stories about how clever Leonardo was. He says that Leonardo painted a round wooden [[shield]] with a picture of snakes spitting fire. Messer Piero took his son's painting to Florence and sold it to an art dealer.<ref name= Vasari>Giorgio Vasari, ''Lives of the Artists'', 1568; this edition Penguin Classics, trans. George Bull 1965</ref> === Verrocchio's workshop, 1466–1476 === In 1466, when Leonardo was fourteen, his father took him to [[Florence]], to be an [[apprenticeship|apprentice]] to the artist [[Andrea del Verrocchio|Verrocchio]]. [[Florence]] was an exciting place for a young person who wanted to be an artist. Many famous artists had lived in [[Florence]], starting with [[Cimabue]] and [[Giotto di Bondone|Giotto]] in the 1200s. Everywhere a person looked, there were famous and beautiful artworks. The huge cathedral had an enormous new dome. The church of St John had doors that gleamed with gold and were said to be the most beautiful doors in the world. Another church had statues all around it by the most famous sculptors, including one by Leonardo's teacher Verrocchio. If an artist was lucky, they would find a rich patron who would buy lots of their paintings. The richest family in Florence were the [[House of Medici|Medici]]. They had built themselves the finest palace in [[Florence]], and liked buying paintings, statues and other beautiful things. They were also interested in the study of [[literature]] and [[philosophy]]. Many young artists hoped to get work from the [[House of Medici|Medici]] and their friends. [[Andrea del Verrocchio|Verrocchio]] had a big workshop that was one of the busiest in [[Florence]]. Leonardo was learning to be an artist, so he had to learn [[drawing]], [[painting]], [[sculpture|sculpting]] and model-making. While he was at the workshop, he learned many other useful skills: [[chemistry]], [[metallurgy]], [[metal]] working, [[plaster]] casting, [[leather]] working, [[mechanics]] and [[carpenter|carpentry]].<ref name=AM>{{citation | first = Andrew | last = Martindale | title = The Rise of the Artist | year = 1972 | publisher = Thames and Hudson | isbn = 0-500-56006-4 }}</ref><ref>Theophilus ''On Divers Arts'', translators:J.G.Hawthorne and C.S. Smith, University of Chicago Press, 1963; reprinted New York: Dover Publications 1979. This is a Medieval practical handbook of skills for the artisan, and includes a brief instruction for mixing oil paint.</ref><ref>Cennino d’A. Cennini ''Il Libro dell’ Arte'', ed. D. V. Thompson Jr. (1933) New Haven: Yale University Press. A practical handbook of painting written in the early 15th century.</ref> Leonardo was not the only young painter at [[Andrea del Verrocchio|Verrocchio]]'s workshop. Many other painters trained there, or often visited. Some of them later became famous: [[Domenico Ghirlandaio|Ghirlandaio]], [[Pietro Perugino|Perugino]] and [[Sandro Botticelli|Botticelli]]. These artists were all a few years older than Leonardo.<ref name= LB /><ref name=DA/> [[Giorgio Vasari]] tells an interesting story from this time in Leonardo's life. [[Andrea del Verrocchio|Verrocchio]] was painting a large picture of ''the Baptism of Christ''. He gave Leonardo the job of painting one of the angels holding [[Jesus]]' robe on the left side of the picture. [[Giorgio Vasari|Vasari]] said that Leonardo painted the angel so beautifully that [[Andrea del Verrocchio|Verrocchio]] put down his brush and never painted again.<ref name= Vasari/> When the painting is examined closely, it is possible to see that many other parts of the picture, such as the rocks, the brown stream and the background may have been painted by Leonardo as well.<ref name=Chiesa/> [[Andrea del Verrocchio|Verrocchio]] made a bronze statue of ''[[David]]'' at this time. It is believed that he used Leonardo as his model.<ref name=Chiesa/> In about 1472, when he was twenty, Leonardo joined the [[Guild of St Luke]], an organization of artists and doctors of medicine. Even after his father set him up in his own workshop, Leonardo still enjoyed working at Verrocchio's workshop.<ref name= LB /> Leonardo's earliest known work is a drawing in pen and ink of the [[Arno River]] valley. It has the date 5 August, 1473. It is now in the [[Uffizi]] Gallery.<ref name= DA /> === Working life 1476–1499 === [[File:Leonardo Magi.jpg|thumb|250px|''The Adoration of the Magi'', (1481) in the [[Uffizil]] Gallery]] When [[Giorgio Vasari|Vasari]] writes about Leonardo, he uses words like "noble", "generous", "graceful," and "beautiful". [[Giorgio Vasari|Vasari]] tells us that as an adult, Leonardo was a tall handsome man. He was so strong that he could bend [[horseshoe]]s with his bare hands. His voice was so beautiful that it charmed everyone that heard it. Almost everyone wanted to be his friend. He loved animals, was a [[vegetarian]] and would buy birds at the [[market]] to set free.<ref>Eugene Muntz, ''Leonardo da Vinci Artist, Thinker, and Man of Science'' (1898), quoted at [http://www.ivu.org/history/davinci/hurwitz.html Leonardo da Vinci's Ethical Vegetarianism]</ref> Very little is known about Leonardo's life and work between 1472 and 1481. People think he was busy in [[Florence]].<ref name=Chiesa/> In 1478, he had an important [[contract|commission]] to paint an [[altarpiece]] for the Monks of San Donato a Scopeto. The painting was to be the ''[[Christmas|Adoration of the Magi]]'' (''The Three Wise Men''). The painting was never finished because Leonardo was sent away to [[Milan]]. Leonardo was a very talented musician.<ref>Winternitz, Emanuel and Libin, Laurence "Leonardo da Vinci." Grove Music Online. Oxford Music Online. Oxford University Press, {{subscription required}}, accessed December 26, 2015, http://www.oxfordmusiconline.com/subscriber/article/grove/music/16426</ref> In 1482, he made a silver [[lyre]] (a [[musical instrument]]) in the shape of a horse's head. At that time there was a new ruler in the city of [[Milan]], in the north of [[Italy]]. Duke [[Ludovico il Moro]] was making other rulers nervous. [[Medici|Lorenzo Medici]] sent Leonardo to Milan as an [[ambassador]]. [[Lorenzo de' Medici]] wanted Leonardo to give Ludovico the lyre as a present from him.<ref> {{cite book |last=Rossi |first=Paolo |title=The Birth of Modern Science |url=https://archive.org/details/birthofmodernsci0000ross |year=2001 |publisher=Blackwell Publishing |isbn=0631227113 |pages=[https://archive.org/details/birthofmodernsci0000ross/page/33 33] }}</ref> Leonardo wrote a letter to the Duke of Milan, telling him about all the clever and useful things that he could do, like making war machines. He wrote in the letter that he could "also paint". Leonardo did not know at the time that it was for painting that he would be mostly remembered.<ref name=DA/><ref> {{cite web | title =Leonardo's Letter to Ludovico Sforza | publisher =Leonardo-history | url =http://www.leonardo-history.com/life.htm?Section=S5 | accessdate =2007-09-28 }}</ref> Leonardo stayed in [[Milan]] and worked for the Duke between 1482 and 1499. Part of his work was to design [[festival]]s and [[carnival]] processions. In Leonardo's notebooks are drawings of theatre costumes, amazing [[helmet]]s and scenes that might be for the [[theatre]]. [[File:Leonardo Da Vinci - Vergine delle Rocce (Louvre).jpg|thumb|left|''Virgin of the Rocks'', [[Louvre]], possibly 1505–1508.]] Leonardo, like most other well-known artists of his time, had servants, young students and older assistants in his workshop. One of his young students was a boy whose name was Gian Giacomo Caprotti da Oreno. He was a handsome boy with beautiful long golden curls. He looked perfect as an artist's model for an angel. But he was such a difficult and dishonest boy that Leonardo called him "Salai" or "Salaino" which means "the little devil". Leonardo wrote in his notebook that Salai was very greedy, that he was a liar and that he had stolen things from the house at least five times.<ref>Leonardo, Codex C. 15v, Institut of France. Trans. Richter</ref> Salai stayed in Leonardo's household for thirty years as a pupil and a servant. ==== ''Gran Cavallo'' ==== Leonardo's most important work for Duke Ludovico was to make a huge statue of the previous ruler, [[Francesco Sforza,]] on horseback. He started with the horse. After studying horses and drawing designs, he made a huge horse of [[clay]]. It was called the "Gran Cavallo". It was going to be cast in [[bronze]], and it was going to be the biggest bronze horse made in more than a thousand years. Unfortunately, the bronze horse was never made. In 1494, Ludovico had the bronze made into [[cannon]]s because the [[French Army|French army]] was invading [[Milan]].<ref name=DA/> The huge clay horse was still standing when the French army invaded again in 1499. It was used for target practice and completely destroyed.<ref name=DA/> ==== ''The Virgin of the Rocks'' ==== While Leonardo was working for Duke Ludovico, he had two important painting commissions. One was an [[oil painting]] to go in a big altarpiece for the [[Confraternity of the Immaculate Conception]]. Leonardo did the painting twice. He left one with the monks in [[Milan]], and took the other painting to France. It is now displayed in the [[Louvre]] Museum. Both paintings are called ''the Virgin of the Rocks''. They show a scene of the [[Virgin Mary]] and the child [[Jesus]] in a rocky mysterious landscape. Mary and Jesus are meeting with [[John the Baptist]]. There is a story (which is not in the Bible but is part of Christian tradition) about how the baby John and the baby Jesus met on the road to [[Egypt]]. In this scene John is praying and the baby Jesus raises his hand to bless John. The paintings have a strange eerie light with soft deep shadows. In the background is a lake and mountains in the mist. No paintings like this had ever been done before.<ref name=DA> {{citation | first = Daniel | last = Arasse | title = Leonardo da Vinci | year = 1997 | publisher = Konecky & Konecky | isbn = 1-56852-1987 }}</ref> [[File:Schéma de reconstitution du retable de la Vierge aux rochers.png|thumb|left|250px|Reconstruction of the surrounding altarpiece of Leonardo's Virgin of the Rocks]] ==== ''The Last Supper'' ==== [[File:Última Cena - Da Vinci 5.jpg|thumb|400px|''The Last Supper'' (1498). [[Milan]], Italy.]] Leonardo's other important painting in Milan is even more famous: [[The Last Supper|''The'' ''Last Supper'']]. The painting shows the last meal shared by [[Jesus]] with his disciples, before his capture and death. Leonardo chose to paint the moment when Jesus said "one of you will betray me". Leonardo tells the story of the surprise and upset that this caused to the twelve followers of Jesus.<ref name=DA/> He shows this through the actions and faces of the people in the painting. Some are talking, some have stood up, and some are raising their hands in horror. The novelist [[Matteo Bandello]] saw Leonardo at work. Bandello wrote that on some days he would paint from morning till night without stopping to eat. Then for three or four days he would not paint at all. He would often just stand and look at the painting.<ref name=Wasser/> Vasari said that the prior of the convent was very annoyed. He asked Ludovico to tell Leonardo to work faster. Vasari said that Leonardo was worried because he did not think that he could paint the face of Jesus well enough. Leonardo told the Duke that he might use the face of the prior as his model for Judas, the traitor.<ref name= Vasari/> When it was finished, everyone that saw it said that the painting was a masterpiece.<ref name=Vasari/> But Leonardo had not used proper [[fresco]] for the painting. He had used tempera over gesso, which is not usually used for wall painting. Soon the painting started to grow mold and flake off the wall. In a hundred years it was "completely ruined".<ref name=Chiesa/> Even though in some places the paint has fallen right off the wall, the painting is so popular that it is printed and copied more that any other religious painting in the world. === Working life 1499–1513 === [[File:Vinci, Leonardo da 1452-1519 Signature from the Paintings and Drawings 08 Signature.jpg|thumb|250px|right|Signature]] In 1499, Ludovico il Moro was overthrown. Leonardo left Milan with his servant Salai and a friend, [[Luca Pacioli]], a mathematician. They went to [[Venice]], where Leonardo worked as a military architect and engineer. Because Venice is a city on many [[island]]s, Leonardo tried to think of ways to defend the city from a [[navy|naval]] attack.<ref name=Chiesa/><ref name=LB/> In 1500, Leonardo went back to Florence, taking his "household" of servants and apprentices with him. The monks from the monastery of ''The Holy Annunciation'' gave Leonardo a home and a large workshop. In 2005 when some buildings which were used by the Department of Military Geography were being [[Building restoration|restored]], the restorers discovered that part of the building used to be Leonardo's studio.<ref>{{cite news |first=Richard |last=Owen |title=Found: the studio where Leonardo met Mona Lisa |work=[[The Times]] |date=2005-01-12 |url=http://www.timesonline.co.uk/tol/news/world/article411195.ece |accessdate=2008-02-22 |archive-date=2011-07-18 |archive-url=https://web.archive.org/web/20110718125807/http://www.timesonline.co.uk/tol/news/world/article411195.ece |url-status=dead }}</ref> [[File:Leonardo - St. Anne cartoon.jpg|thumb|left|''The Virgin and Child with St. Anne and St. John the Baptist'' (c.&nbsp;1499–1500)—[[National Gallery, London]]]] ==== ''The Virgin and Child with St Anne and John the Baptist'' ==== Leonardo started work on a new painting. He drew a large "cartoon" (a drawing that is a plan for the painting). The cartoon showed the Virgin Mary sitting on the knee of her mother, [[Saint Anna|St Anne]]. Mary holds the baby Jesus in her arms. Jesus stretches out his hands to his young cousin John the Baptist. Vasari says that everyone was so amazed by the beautiful drawing that "men and women, young and old" came in large groups to see it "as if they were attending a great festival".<ref name=Vasari/> The drawing is now in the [[National Gallery, London]]. Even though it is old, faded, and kept in a dark room, people go to the gallery to sit in front of it every day. ==== ''The Battle of Anghiari'' ==== [[File:After leonardo da vinci, The Battle of Anghiari by Rubens, Louvre.jpg|thumb|250px|This is a copy of Leonardo's painting of ''The Battle of Anghiari'' which was damaged and then covered by a wall around 1560; possibly this is a copy of the "Cartoon' {outline} of the painting]] [[File:After leonardo da vinci, The Battle of Anghiari (palazzo vecchio, florencel).jpg|thumb|250px|This is a copy of Leonardo's work ''The Battle of Anghiari'' that was painted but unfinished]] In 1502 and 1503, Leonardo worked for [[Cesare Borgia]], a powerful noble who was the son of [[List of popes|Pope Alexander VI]]. Leonardo travelled around Italy with Borgia as a military architect and engineer.<ref name=Chiesa/> Late in 1503, Leonardo returned to Florence. He rejoined the Guild of St Luke. He was given a very important commission. The Signoria ([[Government|Town Council]]) of the City of Florence wanted two large [[fresco]]s painted on the walls of the most important room of the [[Signoria Palace]]. [[Michelangelo]] was to paint ''The Battle of Cascina'' and Leonardo was to paint ''The Battle of Anghiari''.<ref name=Chiesa/> Leonardo began the project by studying and drawing the faces of angry men and fighting horses. These drawings can still be seen in his notebooks. But unfortunately, this was to be another failure for Leonardo. When he painted the picture on the wall, instead of using fresco, he mixed the paints with oil. The paint would not dry. Leonardo lit some fires to dry it, and the painting melted. [[Peter Paul Rubens]] drew a copy of the middle part. After a time, the town council covered it up and got somebody else to paint the wall. Michelangelo did not finish his painting either, because the Pope called him to [[Rome]].<ref name=Chiesa/> ==== ''Mona Lisa'' ==== [[File:Mona Lisa, by Leonardo da Vinci, from C2RMF retouched.jpg|thumb|''[[Mona Lisa]]'' or ''La Gioconda'' (1503–1519)—[[Louvre]], Paris, France]] In about 1503 Leonardo began painting the portrait of a woman known as ''[[Mona Lisa]]'', the most famous portrait that has ever been painted. He continued working on it for many years. It is a small picture, painted in oil on a wooden panel. It shows the face, upper body and hands of a woman. She is very plainly dressed. For a portrait, a woman would usually put on her best clothes and jewellery. Mona Lisa has a dark dress and a fine black veil over her head. Leonardo often left [[symbol]]s in his paintings that give clues about the person. The unusual thing about this picture is the smile. The smile is the clue to her name: Mona Lisa Giacondo. Giacondo means "the joking one". (Mona is short for Madonna which means "My Lady".) The reason why the painting is so famous is that it seems to be full of mystery. Mona Lisa's eyes look out at the viewer. But no-one can guess what she is thinking. Her eyes and her mouth seem to be smiling. This is very unusual in a portrait painting. Most people in portraits look very serious. It is hard to tell what Mona Lisa's exact expression is. When a person wants to read another person's feelings, they look at the corners of their mouth and eyes. But Leonardo has painted soft shadows in the corners of Mona Lisa's mouth and eyes, to disguise her expression. The soft shadows are also found on the sides of her face, her neck and hands. The way that Leonardo uses shadow is called "sfumato" (which is an Italian word for "smoke"). Vasari said that the picture was so beautifully painted that every other artist who looked at it thought that they could never paint so well.<ref name=Vasari/> === Working life, 1506–1516 === In 1506, Leonardo went back to Milan with his pupils, and lived in his own house in Porta Orientale. D'Oggione made several copies of the ''Last Supper''. Luini made a copy of ''the Virgin of the Rocks''. Boltraffio (and the others) painted many [[Madonna and Child]] pictures which can still be seen in art galleries and churches. One of pupils was a young nobleman called Count [[Francesco Melzi]]. Melzi never became a very good painter, but he loved Leonardo and stayed with him until the day he died. In September 1513 Leonardo went to Rome and lived there until 1516. He lived in the [[Vatican]]. The three greatest painters of the High [[Renaissance]], Leonardo, Michelangelo and [[Raphael]] were all working in Rome at the same time.<ref name=Chiesa/> Even though their names are often said together as if they were friends, they were not. Leonardo at this time was in his sixties, Michelangelo was middle-aged. He was not friendly to either Leonardo or Raphael. Raphael was a very clever young painter who learnt a lot by looking at the pictures painted by Leonardo and Michelangelo. But neither of them was ever his teacher. In October 1515, King [[Francis I of France]] captured Milan.<ref name=Wasser>Jack Wasserman, ''Leonardo da Vinci''</ref> On December 19, there was a meeting of Francis I and Pope Leo X, in [[Bologna]]. Leonardo went to the meeting with Pope Leo.<ref name=LB/><ref>Georges Goyau, ''François I'', Transcribed by Gerald Rossi. The Catholic Encyclopedia, Volume VI. Published 1909. New York: Robert Appleton Company. Retrieved on 2007-10-04</ref><ref> {{citation | first = Salvador | last = Miranda | url = http://www.fiu.edu/~mirandas/bios1527-ii.htm | title = The Cardinals of the Holy Roman Church: Antoine du Prat | year = 1998–2007 | accessdate = 2007-10-04 }}</ref> Leonardo made an amazing toy to entertain King Francis. It was a life-sized mechanical [[lion]] that could walk. It had doors in its chest which opened, and a bunch of lilies came out. Lilies were the royal symbol of the French Kings.<ref name=Vasari/> === Old age, 1516–1519 === [[File:Leonardo Da Vinci's house.jpg|thumb|250px|Clos Lucé in France, where Leonardo lived from 1516 to 1519]] In 1516, Francis I invited Leonardo to go to France with him. He gave Leonardo a beautiful house called ''Clos Lucé'' (sometimes called "Cloux"). It is near the king's palace, [[Chateau Amboise]]. Leonardo spent the last three years of his life at Clos Lucé, with his faithful friend and apprentice, Count Melzi. The king gave Leonardo a pension of 10,000 scudi.<ref name=Chiesa/> One of the last paintings that Leonardo did was a picture of ''John the Baptist''. His model was Salai, with his beautiful long curling hair. When Leonardo was dying, he asked for a priest to come, so that he could make his confession and receive [[Holy Communion]].<ref name=Vasari/> Leonardo died at Clos Lucé, on May 2, 1519. King Francis had become a close friend. Vasari says that the King held Leonardo's head in his arms as he died. In his will, he asked that sixty beggars should follow his casket in procession. He was buried in the Chapel of the Chateau Amboise. Leonardo had never married and had no children of his own. In his will, he left his money, his books and most of his paintings to Count Melzi. Leonardo also remembered his other pupil Salai and his servant Battista di Vilussis, who each received half of Leonardo's [[wine|vineyards]] near Milan. Leonardo left to his serving woman a black cloak with a fur edge.<ref>{{cite web| title =Leonardo's will| publisher =Leonardo-history| url =http://www.leonardo-history.com/life.htm?Section=S6| accessdate =2007-09-28}}</ref> Salai was the owner of Leonardo's most famous [[oil painting]], the ''Mona Lisa''. He still owned it a few years later when he died, after fighting in a duel.<ref name=NR>{{cite web | last =Rossiter | first =Nick | title =Could this be the secret of her smile? | publisher =Telegraph.co.UK | date =2003-07-04 | url =https://www.telegraph.co.uk/arts/main.jhtml?xml=/arts/2003/04/07/banr.xml | accessdate =2007-10-03 | archive-date =2007-10-11 | archive-url =https://web.archive.org/web/20071011163341/http://www.telegraph.co.uk/arts/main.jhtml?xml=%2Farts%2F2003%2F04%2F07%2Fbanr.xml | url-status =dead }}</ref> King Francis said: "There had never been another man born in the world who knew as much as Leonardo, not so much about painting, sculpture and architecture, as that he was a very great philosopher."<ref>{{cite book| author = Mario Lucertini, Ana Millan Gasca, Fernando Nicolo | title = Technological Concepts and Mathematical Models in the Evolution of Modern Engineering Systems| work = | publisher =Birkhauser| year = 2004| isbn = 376436940X| url =https://books.google.com/books?id=YISIUycS4HgC&dq=leonardo+cellini+francois+philosopher&pg=PA13| accessdate =2007-10-03 }}</ref><ref name=Berti> {{citation | first = Luciano | last = Berti | title = The Uffizi | year = 1971 | publisher = Scala }}</ref> == Drawings == Leonardo did not paint very many pictures. But he drew hundreds of quick sketches, plans, maps and detailed drawings. This is how he recorded all the interesting things that he saw, studied and thought about. Some of Leonardo's drawings are "studies" for paintings. In these drawings Leonardo planned the things he was going to paint. Some studies are plans for whole paintings. One of these paintings is the large beautiful drawing of the [[Madonna and Child]] with St Anne and St John the Baptist that is now in the [[National Gallery, London]]. Many of the studies show "details" that Leonardo wanted to get just right. One study shows a very detailed [[Perspective (graphical)|perspective]] drawing of the ruined buildings in the background of the painting of the Magi. Other studies show hands, faces, drapery, plants, horses and babies.<ref name="Popham" /> The earliest drawing by Leonardo that has a date on it, is a ''Landscape of the Arno Valley'', 1473, which shows the river, the mountains, [[Montelupo Castle]] and the farmlands beyond it in great detail.<ref name="LB" /><ref name="Popham"> {{citation | first = A.E. | last = Popham | title = The Drawings of Leonardo da Vinci | year = 1946 | publisher = Jonathan Cape | isbn = 0-224-60462-7 }}</ref> == Leonardo's notebooks == [[File:Ailes battantes Luc Viatour.jpg|thumb|Leonardo da Vinci studied the [[flight]] of [[bird]]s and tried to make a flying machine based on his discoveries.]] [[File:Da Vinci Vitruve Luc Viatour.jpg|thumb|The ''Vitruvian Man'' (c.&nbsp;1485) Accademia, Venice]] [[File:Da Vinci Studies of Embryos Luc Viatour.jpg|thumb|A page from Leonardo's journal showing his study of a fetus in the womb (c.&nbsp;1510) Royal Library, [[Windsor Castle]]]] Leonardo studied things throughout his life. He did not go to [[university]] to study. He studied by observing things in the world around him, to see how they were made and how they worked. He drew things that he saw and discoveries he made into his notebooks. Many of his notebooks are now in museums. There are about 13,000 pages of notes and drawings in his notebooks, which are mostly [[science|scientific]] studies.<ref name=DA/> Leonardo's notebooks are hard to read because he wrote backwards in "mirror writing". Some people think that perhaps he was trying to keep his work secret. This is not true. Leonardo wrote (and sometimes drew) with his [[Left-handedness|left hand]]. In those days pens were made from a quill (a large feather) that was cut with a pen-knife on the end. It is hard for a left-handed person to write with a quill in the ordinary way, but quite easy to write backwards. It is likely that Leonardo planned to publish the studies in his notebooks. He organized many pages carefully, with one study taking up the front and back of each page. There is a page with drawings and writing about the human [[heart]] and a page about the [[womb]] and the [[fetus]].<ref>Windsor Castle, Royal Library, sheets RL 19073v-19074v and RL 19102 respectively.</ref> One page shows drawings of the muscles of a [[wikt:shoulder|shoulder]] and another page shows how an [[arm]] works.<ref> {{citation | last1 = O'Malley | last2 = Saunders | title = Leonardo on the Human Body | year = 1982 | publisher = Dover Publications | publication-place = New York }}</ref> The notebooks were not published in Leonardo's lifetime. After he died, they were divided between different people who had known him. They are nearly all in [[museum]]s or [[Library|libraries]] such as [[Windsor Castle]], [[the Louvre]], and the [[British Library]]. The [[Biblioteca Ambrosiana]] (a [[library]]) in [[Milan]] has the twelve-volume ''Codex Atlanticus''.<ref name=DA/> === Studies === Some of the things that Leonardo studied are:<ref name=HG /> <!--Please note that this reference goes for the whole paragraph! --> * The [[geology]] of the Earth, with its [[mountain]]s, [[valley]]s, [[river]]s and [[Rock (geology)|rocks]]. * The [[anatomy]] of the human body with its [[skeleton]], [[muscle]]s, [[vein]]s and [[digestive system|internal organs]]. Leonardo was given dead bodies by a hospital. He [[Dissection|dissected]] thirty dead bodies and carefully drew many of the parts. His drawings of bones and muscles were to help other artists to paint the human body properly. * The anatomy of [[horse]]s, [[cattle|cows]], [[dog]]s, and [[bear]]s. * The expressions on human [[face]]s. * The [[flight]] of [[bird]]s . * The [[weather]] and its [[wikt:phenomena|phenomena]]. * The way that [[water]] flows. * The [[botany]] of plants. * [[Light]], shadows, [[mirror]]s and [[Lens (optics)|lenses]]. * [[Perspective (graphical)|Perspective]] and the way to make things look near or far. * The [[geometry]] of solid objects. He drew many careful pictures which were used by the [[mathematics|mathematician]] [[Luca Pacioli]] in a book called ''De Divina Proportione''. === Designs and inventions === Many of the drawings and notes in Leonardo's notebooks are designs, plans and inventions. Some of the things that Leonardo designed are:<ref name=HG /> <!-- This reference goes for the whole paragraph! --> * [[Costume]]s for [[parade]]s, [[carnival]]s and [[theatre]]. These were probably for Duke Federico's court. They include [[armour]], and a ferocious [[dragon]]. * [[War]] machines such as an armour-plated [[tank]], an enormous [[archery|cross bow]] and a horrible horse-driven leg-chopper. None of these things were ever made in Leonardo's lifetime. * [[Dam]]s and [[canal]]s for rivers. * A wooden [[bridge]] that could be carried flat on wagons and unfolded and put together at the river. * Flying things with wings that flapped, a [[helicopter]], a [[parachute]] and a [[glider|hang glider]]. One of Leonardo's servants was injured, trying out the hang glider. The parachute has been made and tested in modern times, and it does work. * [[Church (building)]] and [[castle]]s. It is possible that the Castle of Locarno, in the south of [[Switzerland]] was designed by Leonardo.<ref>[http://www.swissinfo.ch/eng/Locarno_gets_its_own_da_Vinci_mystery.html?cid=5371158 Locarno Castle] {{Webarchive|url=https://web.archive.org/web/20120119193609/http://www.swissinfo.ch/eng/Locarno_gets_its_own_da_Vinci_mystery.html?cid=5371158 |date=2012-01-19 }}, accessed 14-02-1020</ref> No other building that he designed was built. == Leonardo's studies, designs and inventions == <center><gallery> File:View of a Skull III.jpg|Study of a skull File:Study of horse.jpg|Study of a horse for the Duke's statue File:Sedge.jpg|Study of a sedge plant File:Leonardo polyhedra.png|Rhombicuboctahedron published in Pacioli's book File:Leonardo parabolic compass.JPG|A parabolic compass. File:Leonardo helicopter.JPG|A Helicopter. File:Leonardo cannons.JPG|Cannons. File:Leonardo walking on water.JPG|Walking on water. File:Leonardo da Vinci parachute 04659a.jpg|Modern model of Leonardo's parachute. File:Da vinci bridge.jpg|Modern model of a bridge designed by Leonardo. File:DaVinciTankAtAmboise.jpeg|Modern model of a tank by Leonardo File:Leonardo-Flywheel-screenshot.jpg|Modern model of a flywheel </gallery></center> ==Related pages== * [[Renaissance]] * [[Italian Renaissance art]] * ''[[Lady with an Ermine]]'' * [[List of Renaissance artists]] * [[Michelangelo]] * [[Raphael]] * [[List of Italian painters]] == References == {{reflist|2}} == More reading == * {{cite book | author = Daniel Arasse| title = Leonardo da Vinci | publisher = Konecky & Konecky | year = 1997 | isbn = 1-56852-1987}} * {{cite book | author = Liana Bortolon| title = The life and times of Leonardo | publisher = Paul Hamlyn, London | year = 1967 | isbn = 075251587X }} * {{cite book | author = Hugh Brigstoke| title = The Oxford Companion to Western Art | url = https://archive.org/details/oxfordcompaniont0000unse_r1z5| publisher = USA: Oxford University Press | year = 2001 | isbn = 0198662033}} * {{cite book | author = Angela Ottino della Chiesa| title = The complete paintings of Leonardo da Vinci | publisher = Penguin Classics of World Art series | year = 1967 | isbn = 0-14-00-8649-8}} * {{cite book | author = Charles D. O'Malley and J. B. de C. M. Sounders | title = Leonardo on the human body: the anatomical, physiological, and embryological drawings of Leonardo da Vinci. With translations, emendations and a biographical introduction | publisher = Henry Schuman, New York | year = 1952 | id = }} * {{cite book | author =A.E. Popham | title = The drawings of Leonardo da Vinci | publisher = Jonathan Cape | year = 1946 | isbn = 0-224-60462-7}} * {{cite book |author=Jean Paul Richter |title = The notebooks of Leonardo da Vinci |publisher= Dover |year= 1970 |isbn= 0-486-22572-0 }} ISBN 0-486-22573-9. 2 volumes. A reprint of [http://www.gutenberg.org/etext/5000 the original 1883 edition]. * {{cite book |author=Paolo Rossi |title= The Birth of Modern Science |url=https://archive.org/details/birthofmodernsci0000ross |publisher= Blackwell Publishing |year=2001 |isbn= 0631227113}} * {{cite book |author=Bruno Santi |title= Leonardo da Vinci |publisher= Scala / Riverside |year= 1990}} * {{cite book | author = Jack Wasserman| title = Leonardo da Vinci | url = https://archive.org/details/leonardo0000leon| publisher = Abrams | year = 1975 | isbn = 0-8109-0262-1}} * Silvia e Luca Guagliumi, "Leonardo e l'architettura", Silvia Editrice, Aprile 2015 {{ISBN|978-88-96036-65-5}} * Silvia Guagliumi,Expo 2015: Leonardo da Vinci e l'architettura lombarda.Saggio pubblicato su Il Giornale dell'Ingegnere (Organo ufficiale dell'Ordine degli Ingegneri e Architetti di Milano), N.4 - Aprile 2015. * Silvia Guagliumi, In ricordo di Leonardo, Rivista In Arte, Aprile/Maggio/Giugno 2019. == Other websites == {{sisterlinks|s=Author:Leonardo da Vinci}} * {{ws|"[[s:Catholic Encyclopedia (1913)/Leonardo da Vinci|Leonardo da Vinci]]" in the 1913 ''Catholic Encyclopedia''}} * [http://www.studio-international.co.uk/reports/da_vinci.asp Leonardo da Vinci: Experience, Experiment, Design (review)] {{Webarchive|url=https://web.archive.org/web/20090503052418/http://www.studio-international.co.uk/reports/da_vinci.asp |date=2009-05-03 }} * {{gutenberg author | id=Leonardo_da_Vinci | name=Leonardo da Vinci}} * {{gutenberg | no=7785 | name=Leonardo da Vinci by Maurice Walter Brockwell}} * [http://www.sacred-texts.com/aor/dv/index.htm Complete text & images of Richter's translation of the Notebooks] * [http://www.mnstate.edu/gracyk/courses/web%20publishing/Vasari_daVinci.htm Vasari Life of Leonardo] {{Webarchive|url=https://web.archive.org/web/20090317061000/http://www.mnstate.edu/gracyk/courses/web%20publishing/Vasari_daVinci.htm |date=2009-03-17 }}: in ''Lives of the Most Eminent Painters, Sculptors, and Architects''. * [http://www.wga.hu/frames-e.html?/html/l/leonardo/ Web Gallery of Leonardo Paintings] * [http://www.drawingsofleonardo.org Drawings of Leonardo da Vinci] * [http://arts.guardian.co.uk/features/story/0,,1860869,00.html Da Vinci Decoded] Article from ''[[The Guardian]]'' * [http://www.ted.com/index.php/talks/view/id/235 The true face of Leonardo Da Vinci?] * [http://www.ivu.org/history/davinci/hurwitz.html Leonardo da Vinci's Ethical Vegetarianism] * [https://www.youtube.com/watch?v=mRKzvCK_-X0 Leonardo da Vinci, a biography with description of his major works, written for children; Tamsyn Taylor, 2020] {{Leonardo da Vinci|state=collapsed}} {{High Renaissance}} {{Mathematics and art}} {{Authority control}} {{DEFAULTSORT:Leonardo da Vinci}} [[Category:1452 births]] [[Category:1519 deaths]] [[Category:Italian inventors]] [[Category:Italian mathematicians]] [[Category:Italian architects]] [[Category:Italian sculptors]] [[Category:Botanists]] [[Category:People from Florence]] [[Category:15th-century Italian painters]] [[Category:Anatomists]] [[Category:Polymaths]]
Leonardo di ser Piero da Vinci (15 April 1452 – 2 May 1519) was an Italian polymath of the High Renaissance who was active as a painter, draughtsman, engineer, scientist, theorist, sculptor, and architect. While his fame initially rested on his achievements as a painter, he also became known for his notebooks, in which he made drawings and notes on a variety of subjects, including anatomy, astronomy, botany, cartography, painting, and paleontology. Leonardo is widely regarded to have been a genius who epitomized the Renaissance humanist ideal, and his collective works comprise a contribution to later generations of artists matched only by that of his younger contemporary by 23 years Michelangelo Buonarroti. Born out of wedlock to a successful notary and a lower-class woman in, or near, Vinci, he was educated in Florence by the Italian painter and sculptor Andrea del Verrocchio. He began his career in the city, but then spent much time in the service of Ludovico Sforza in Milan. Later, he worked in Florence and Milan again, as well as briefly in Rome, all while attracting a large following of imitators and students. Upon the invitation of Francis I, he spent his last three years in France, where he died in 1519. Since his death, there has not been a time where his achievements, diverse interests, personal life, and empirical thinking have failed to incite interest and admiration, making him a frequent namesake and subject in culture. Leonardo is identified as one of the greatest painters in the history of art and is often credited as the founder of the High Renaissance. Despite having many lost works and fewer than 25 attributed major works – including numerous unfinished works – he created some of the most influential paintings in Western art. His magnum opus, the Mona Lisa, is his best known work and often regarded as the world's most famous painting. The Last Supper is the most reproduced religious painting of all time and his Vitruvian Man drawing is also regarded as a cultural icon. In 2017, Salvator Mundi, attributed in whole or part to Leonardo, was sold at auction for US$450.3 million, setting a new record for the most expensive painting ever sold at public auction. Revered for his technological ingenuity, he conceptualized flying machines, a type of armored fighting vehicle, concentrated solar power, a ratio machine that could be used in an adding machine, and the double hull. Relatively few of his designs were constructed or were even feasible during his lifetime, as the modern scientific approaches to metallurgy and engineering were only in their infancy during the Renaissance. Some of his smaller inventions, however, entered the world of manufacturing unheralded, such as an automated bobbin winder and a machine for testing the tensile strength of wire. He made substantial discoveries in anatomy, civil engineering, hydrodynamics, geology, optics, and tribology, but he did not publish his findings and they had little to no direct influence on subsequent science. Leonardo da Vinci, properly named Leonardo di ser Piero da Vinci ("Leonardo, son of ser Piero from Vinci"), was born on 15 April 1452 in, or close to, the Tuscan hill town of Vinci, 20 miles from Florence. He was born out of wedlock to Piero da Vinci (Ser Piero da Vinci d'Antonio di ser Piero di ser Guido; 1426–1504), a Florentine legal notary, and Caterina di Meo Lippi (c. 1434–1494), from the lower class. It remains uncertain where Leonardo was born; the traditional account, from a local oral tradition recorded by the historian Emanuele Repetti, is that he was born in Anchiano, a country hamlet that would have offered sufficient privacy for the illegitimate birth, though it is still possible he was born in a house in Florence that Ser Piero almost certainly had. Leonardo's parents both married separately the year after his birth. Caterina – who later appears in Leonardo's notes as only "Caterina" or "Catelina" – is usually identified as the Caterina Buti del Vacca, who married the local artisan Antonio di Piero Buti del Vacca, nicknamed L'Accattabriga, 'the quarrelsome one'. Ser Piero married Albiera Amadori – having been betrothed to her the previous year – and after her death in 1464, went on to have three subsequent marriages. From all the marriages, Leonardo eventually had 16 half-siblings (of whom 11 survived infancy) who were much younger than he (the last was born when Leonardo was 46 years old) and with whom he had very little contact. Very little is known about Leonardo's childhood and much is shrouded in myth, partially because of his biography in the frequently apocryphal Lives of the Most Excellent Painters, Sculptors, and Architects (1550) by 16th-century art historian Giorgio Vasari. Tax records indicate that by at least 1457 he lived in the household of his paternal grandfather, Antonio da Vinci, but it is possible that he spent the years before then in the care of his mother in Vinci, either Anchiano or Campo Zeppi in the parish of San Pantaleone. He is thought to have been close to his uncle, Francesco da Vinci, but his father was probably in Florence most of the time. Ser Piero, who was the descendant of a long line of notaries, established an official residence in Florence by at least 1469 and had a successful career. Despite his family history, Leonardo only received a basic and informal education in (vernacular) writing, reading, and mathematics; possibly because his artistic talents were recognised early, so his family decided to focus their attention there. Later in life, Leonardo recorded his earliest memory, now in the Codex Atlanticus. While writing on the flight of birds, he recalled as an infant when a kite came to his cradle and opened his mouth with its tail; commentators still debate whether the anecdote was an actual memory or a fantasy. In the mid-1460s, Leonardo's family moved to Florence, which at the time was the centre of Christian Humanist thought and culture. Around the age of 14, he became a garzone (studio boy) in the workshop of Andrea del Verrocchio, who was the leading Florentine painter and sculptor of his time. This was about the time of the death of Verrocchio's master, the great sculptor Donatello. Leonardo became an apprentice by the age of 17 and remained in training for seven years. Other famous painters apprenticed in the workshop or associated with it include Ghirlandaio, Perugino, Botticelli, and Lorenzo di Credi. Leonardo was exposed to both theoretical training and a wide range of technical skills, including drafting, chemistry, metallurgy, metal working, plaster casting, leather working, mechanics, and woodwork, as well as the artistic skills of drawing, painting, sculpting, and modelling. Leonardo was a contemporary of Botticelli, Ghirlandaio and Perugino, who were all slightly older than he was. He would have met them at the workshop of Verrocchio or at the Platonic Academy of the Medici. Florence was ornamented by the works of artists such as Donatello's contemporaries Masaccio, whose figurative frescoes were imbued with realism and emotion, and Ghiberti, whose Gates of Paradise, gleaming with gold leaf, displayed the art of combining complex figure compositions with detailed architectural backgrounds. Piero della Francesca had made a detailed study of perspective, and was the first painter to make a scientific study of light. These studies and Leon Battista Alberti's treatise De pictura were to have a profound effect on younger artists and in particular on Leonardo's own observations and artworks. Much of the painting in Verrocchio's workshop was done by his assistants. According to Vasari, Leonardo collaborated with Verrocchio on his The Baptism of Christ (c. 1472–1475), painting the young angel holding Jesus's robe with skill so far superior to his master's that Verrocchio purportedly put down his brush and never painted again (the latter claim probably being apocryphal). The new technique of oil paint was applied to areas of the mostly tempera work, including the landscape, the rocks seen through the brown mountain stream, and much of Jesus's figure, indicating Leonardo's hand. Additionally, Leonardo may have been a model for two works by Verrocchio: the bronze statue of David in the Bargello and the archangel Raphael in Tobias and the Angel. Vasari tells a story of Leonardo as a very young man: a local peasant made himself a round buckler shield and requested that Ser Piero have it painted for him. Leonardo, inspired by the story of Medusa, responded with a painting of a monster spitting fire that was so terrifying that his father bought a different shield to give to the peasant and sold Leonardo's to a Florentine art dealer for 100 ducats, who in turn sold it to the Duke of Milan Ludovico Sforza. By 1472, at the age of 20, Leonardo qualified as a master in the Guild of Saint Luke, the guild of artists and doctors of medicine, but even after his father set him up in his own workshop, his attachment to Verrocchio was such that he continued to collaborate and live with him. Leonardo's earliest known dated work is a 1473 pen-and-ink drawing of the Arno valley (see below). According to Vasari, the young Leonardo was the first to suggest making the Arno river a navigable channel between Florence and Pisa. In January 1478, Leonardo received an independent commission to paint an altarpiece for the Chapel of Saint Bernard in the Palazzo Vecchio, an indication of his independence from Verrocchio's studio. An anonymous early biographer, known as Anonimo Gaddiano, claims that in 1480 Leonardo was living with the Medici and often worked in the garden of the Piazza San Marco, Florence, where a Neoplatonic academy of artists, poets and philosophers organized by the Medici met. In March 1481, he received a commission from the monks of San Donato in Scopeto for The Adoration of the Magi. Neither of these initial commissions were completed, being abandoned when Leonardo went to offer his services to Duke of Milan Ludovico Sforza. Leonardo wrote Sforza a letter which described the diverse things that he could achieve in the fields of engineering and weapon design, and mentioned that he could paint. He brought with him a silver string instrument – either a lute or lyre – in the form of a horse's head. With Alberti, Leonardo visited the home of the Medici and through them came to know the older Humanist philosophers of whom Marsiglio Ficino, proponent of Neoplatonism; Cristoforo Landino, writer of commentaries on Classical writings, and John Argyropoulos, teacher of Greek and translator of Aristotle were the foremost. Also associated with the Platonic Academy of the Medici was Leonardo's contemporary, the brilliant young poet and philosopher Pico della Mirandola. In 1482, Leonardo was sent as an ambassador by Lorenzo de' Medici to Ludovico il Moro, who ruled Milan between 1479 and 1499. Leonardo worked in Milan from 1482 until 1499. He was commissioned to paint the Virgin of the Rocks for the Confraternity of the Immaculate Conception and The Last Supper for the monastery of Santa Maria delle Grazie. In the spring of 1485, Leonardo travelled to Hungary (on behalf of Sforza) to meet king Matthias Corvinus, and was commissioned by him to paint a Madonna. In 1490 he was called as a consultant, together with Francesco di Giorgio Martini, for the building site of the cathedral of Pavia and was struck by the equestrian statue of Regisole, of which he left a sketch. Leonardo was employed on many other projects for Sforza, such as preparation of floats and pageants for special occasions; a drawing of, and wooden model for, a competition to design the cupola for Milan Cathedral; and a model for a huge equestrian monument to Ludovico's predecessor Francesco Sforza. This would have surpassed in size the only two large equestrian statues of the Renaissance, Donatello's Gattamelata in Padua and Verrocchio's Bartolomeo Colleoni in Venice, and became known as the Gran Cavallo. Leonardo completed a model for the horse and made detailed plans for its casting, but in November 1494, Ludovico gave the metal to his brother-in-law to be used for a cannon to defend the city from Charles VIII of France. Contemporary correspondence records that Leonardo and his assistants were commissioned by the Duke of Milan to paint the Sala delle Asse in the Sforza Castle, c. 1498. The project became a trompe-l'œil decoration that made the great hall appear to be a pergola created by the interwoven limbs of sixteen mulberry trees, whose canopy included an intricate labyrinth of leaves and knots on the ceiling. When Ludovico Sforza was overthrown by France in 1500, Leonardo fled Milan for Venice, accompanied by his assistant Salaì and friend, the mathematician Luca Pacioli. In Venice, Leonardo was employed as a military architect and engineer, devising methods to defend the city from naval attack. On his return to Florence in 1500, he and his household were guests of the Servite monks at the monastery of Santissima Annunziata and were provided with a workshop where, according to Vasari, Leonardo created the cartoon of The Virgin and Child with Saint Anne and Saint John the Baptist, a work that won such admiration that "men [and] women, young and old" flocked to see it "as if they were going to a solemn festival." In Cesena in 1502, Leonardo entered the service of Cesare Borgia, the son of Pope Alexander VI, acting as a military architect and engineer and travelling throughout Italy with his patron. Leonardo created a map of Cesare Borgia's stronghold, a town plan of Imola in order to win his patronage. Upon seeing it, Cesare hired Leonardo as his chief military engineer and architect. Later in the year, Leonardo produced another map for his patron, one of Chiana Valley, Tuscany, so as to give his patron a better overlay of the land and greater strategic position. He created this map in conjunction with his other project of constructing a dam from the sea to Florence, in order to allow a supply of water to sustain the canal during all seasons. Leonardo had left Borgia's service and returned to Florence by early 1503, where he rejoined the Guild of Saint Luke on 18 October of that year. By this same month, Leonardo had begun working on a portrait of Lisa del Giocondo, the model for the Mona Lisa, which he would continue working on until his twilight years. In January 1504, he was part of a committee formed to recommend where Michelangelo's statue of David should be placed. He then spent two years in Florence designing and painting a mural of The Battle of Anghiari for the Signoria, with Michelangelo designing its companion piece, The Battle of Cascina. In 1506, Leonardo was summoned to Milan by Charles II d'Amboise, the acting French governor of the city. There, Leonardo took on another pupil, Count Francesco Melzi, the son of a Lombard aristocrat, who is considered to have been his favourite student. The Council of Florence wished Leonardo to return promptly to finish The Battle of Anghiari, but he was given leave at the behest of Louis XII, who considered commissioning the artist to make some portraits. Leonardo may have commenced a project for an equestrian figure of d'Amboise; a wax model survives and, if genuine, is the only extant example of Leonardo's sculpture. Leonardo was otherwise free to pursue his scientific interests. Many of Leonardo's most prominent pupils either knew or worked with him in Milan, including Bernardino Luini, Giovanni Antonio Boltraffio, and Marco d'Oggiono. In 1507, Leonardo was in Florence sorting out a dispute with his brothers over the estate of his father, who had died in 1504. By 1508, Leonardo was back in Milan, living in his own house in Porta Orientale in the parish of Santa Babila. In 1512, Leonardo was working on plans for an equestrian monument for Gian Giacomo Trivulzio, but this was prevented by an invasion of a confederation of Swiss, Spanish and Venetian forces, which drove the French from Milan. Leonardo stayed in the city, spending several months in 1513 at the Medici's Vaprio d'Adda villa. In March 1513, Lorenzo de' Medici's son Giovanni assumed the papacy (as Leo X); Leonardo went to Rome that September, where he was received by the pope's brother Giuliano. From September 1513 to 1516, Leonardo spent much of his time living in the Belvedere Courtyard in the Apostolic Palace, where Michelangelo and Raphael were both active. Leonardo was given an allowance of 33 ducats a month, and according to Vasari, decorated a lizard with scales dipped in quicksilver. The pope gave him a painting commission of unknown subject matter, but cancelled it when the artist set about developing a new kind of varnish. Leonardo became ill, in what may have been the first of multiple strokes leading to his death. He practiced botany in the Gardens of Vatican City, and was commissioned to make plans for the pope's proposed draining of the Pontine Marshes. He also dissected cadavers, making notes for a treatise on vocal cords; these he gave to an official in hopes of regaining the pope's favor, but was unsuccessful. In October 1515, King Francis I of France recaptured Milan. Leonardo was present at the 19 December meeting of Francis I and Leo X, which took place in Bologna. In 1516, Leonardo entered Francis' service, being given the use of the manor house Clos Lucé, near the king's residence at the royal Château d'Amboise. Being frequently visited by Francis, he drew plans for an immense castle town the king intended to erect at Romorantin, and made a mechanical lion, which during a pageant walked toward the king and – upon being struck by a wand – opened its chest to reveal a cluster of lilies. Leonardo was accompanied during this time by his friend and apprentice Francesco Melzi, and supported by a pension totalling 10,000 scudi. At some point, Melzi drew a portrait of Leonardo; the only others known from his lifetime were a sketch by an unknown assistant on the back of one of Leonardo's studies (c. 1517) and a drawing by Giovanni Ambrogio Figino depicting an elderly Leonardo with his right arm wrapped in clothing. The latter, in addition to the record of an October 1517 visit by Louis d'Aragon, confirms an account of Leonardo's right hand being paralytic when he was 65, which may indicate why he left works such as the Mona Lisa unfinished. He continued to work at some capacity until eventually becoming ill and bedridden for several months. Leonardo died at Clos Lucé on 2 May 1519 at the age of 67, possibly of a stroke. Francis I had become a close friend. Vasari describes Leonardo as lamenting on his deathbed, full of repentance, that "he had offended against God and men by failing to practice his art as he should have done." Vasari states that in his last days, Leonardo sent for a priest to make his confession and to receive the Holy Sacrament. Vasari also records that the king held Leonardo's head in his arms as he died, although this story may be legend rather than fact. In accordance with his will, sixty beggars carrying tapers followed Leonardo's casket. Melzi was the principal heir and executor, receiving, as well as money, Leonardo's paintings, tools, library and personal effects. Leonardo's other long-time pupil and companion, Salaì, and his servant Baptista de Vilanis, each received half of Leonardo's vineyards. His brothers received land, and his serving woman received a fur-lined cloak. On 12 August 1519, Leonardo's remains were interred in the Collegiate Church of Saint Florentin at the Château d'Amboise. Some 20 years after Leonardo's death, Francis was reported by the goldsmith and sculptor Benvenuto Cellini as saying: "There had never been another man born in the world who knew as much as Leonardo, not so much about painting, sculpture and architecture, as that he was a very great philosopher." Salaì, or Il Salaino ("The Little Unclean One", i.e., the devil), entered Leonardo's household in 1490 as an assistant. After only a year, Leonardo made a list of his misdemeanours, calling him "a thief, a liar, stubborn, and a glutton," after he had made off with money and valuables on at least five occasions and spent a fortune on clothes. Nevertheless, Leonardo treated him with great indulgence, and he remained in Leonardo's household for the next thirty years. Salaì executed a number of paintings under the name of Andrea Salaì, but although Vasari claims that Leonardo "taught him many things about painting," his work is generally considered to be of less artistic merit than others among Leonardo's pupils, such as Marco d'Oggiono and Boltraffio. At the time of his death in 1524, Salaì owned a painting referred to as Joconda in a posthumous inventory of his belongings; it was assessed at 505 lire, an exceptionally high valuation for a small panel portrait. Despite the thousands of pages Leonardo left in notebooks and manuscripts, he scarcely made reference to his personal life. Within Leonardo's lifetime, his extraordinary powers of invention, his "great physical beauty" and "infinite grace," as described by Vasari, as well as all other aspects of his life, attracted the curiosity of others. One such aspect was his love for animals, likely including vegetarianism and according to Vasari, a habit of purchasing caged birds and releasing them. Leonardo had many friends who are now notable either in their fields or for their historical significance, including mathematician Luca Pacioli, with whom he collaborated on the book Divina proportione in the 1490s. Leonardo appears to have had no close relationships with women except for his friendship with Cecilia Gallerani and the two Este sisters, Beatrice and Isabella. While on a journey that took him through Mantua, he drew a portrait of Isabella that appears to have been used to create a painted portrait, now lost. Beyond friendship, Leonardo kept his private life secret. His sexuality has been the subject of satire, analysis, and speculation. This trend began in the mid-16th century and was revived in the 19th and 20th centuries, most notably by Sigmund Freud in his Leonardo da Vinci, A Memory of His Childhood. Leonardo's most intimate relationships were perhaps with his pupils Salaì and Melzi. Melzi, writing to inform Leonardo's brothers of his death, described Leonardo's feelings for his pupils as both loving and passionate. It has been claimed since the 16th century that these relationships were of a sexual or erotic nature. Walter Isaacson in his biography of Leonardo makes explicit his opinion that the relations with Salai were intimate and homosexual. Earlier in Leonardo's life, court records of 1476, when he was aged twenty-four, show that Leonardo and three other young men were charged with sodomy in an incident involving a known male prostitute. The charges were dismissed for lack of evidence, and there is speculation that since one of the accused, Lionardo de Tornabuoni, was related to Lorenzo de' Medici, the family exerted its influence to secure the dismissal. Since that date much has been written about his presumed homosexuality and its role in his art, particularly in the androgyny and eroticism manifested in Saint John the Baptist and Bacchus and more explicitly in a number of erotic drawings. Despite the recent awareness and admiration of Leonardo as a scientist and inventor, for the better part of four hundred years his fame rested on his achievements as a painter. A handful of works that are either authenticated or attributed to him have been regarded as among the great masterpieces. These paintings are famous for a variety of qualities that have been much imitated by students and discussed at great length by connoisseurs and critics. By the 1490s Leonardo had already been described as a "Divine" painter. Among the qualities that make Leonardo's work unique are his innovative techniques for laying on the paint; his detailed knowledge of anatomy, light, botany and geology; his interest in physiognomy and the way humans register emotion in expression and gesture; his innovative use of the human form in figurative composition; and his use of subtle gradation of tone. All these qualities come together in his most famous painted works, the Mona Lisa, the Last Supper, and the Virgin of the Rocks. Leonardo first gained attention for his work on the Baptism of Christ, painted in conjunction with Verrocchio. Two other paintings appear to date from his time at Verrocchio's workshop, both of which are Annunciations. One is small, 59 centimetres (23 in) long and 14 cm (5.5 in) high. It is a "predella" to go at the base of a larger composition, a painting by Lorenzo di Credi from which it has become separated. The other is a much larger work, 217 cm (85 in) long. In both Annunciations, Leonardo used a formal arrangement, like two well-known pictures by Fra Angelico of the same subject, of the Virgin Mary sitting or kneeling to the right of the picture, approached from the left by an angel in profile, with a rich flowing garment, raised wings and bearing a lily. Although previously attributed to Ghirlandaio, the larger work is now generally attributed to Leonardo. In the smaller painting, Mary averts her eyes and folds her hands in a gesture that symbolised submission to God's will. Mary is not submissive, however, in the larger piece. The girl, interrupted in her reading by this unexpected messenger, puts a finger in her bible to mark the place and raises her hand in a formal gesture of greeting or surprise. This calm young woman appears to accept her role as the Mother of God, not with resignation but with confidence. In this painting, the young Leonardo presents the humanist face of the Virgin Mary, recognising humanity's role in God's incarnation. In the 1480s, Leonardo received two very important commissions and commenced another work that was of ground-breaking importance in terms of composition. Two of the three were never finished, and the third took so long that it was subject to lengthy negotiations over completion and payment. One of these paintings was Saint Jerome in the Wilderness, which Bortolon associates with a difficult period of Leonardo's life, as evidenced in his diary: "I thought I was learning to live; I was only learning to die." Although the painting is barely begun, the composition can be seen and is very unusual. Jerome, as a penitent, occupies the middle of the picture, set on a slight diagonal and viewed somewhat from above. His kneeling form takes on a trapezoid shape, with one arm stretched to the outer edge of the painting and his gaze looking in the opposite direction. J. Wasserman points out the link between this painting and Leonardo's anatomical studies. Across the foreground sprawls his symbol, a great lion whose body and tail make a double spiral across the base of the picture space. The other remarkable feature is the sketchy landscape of craggy rocks against which the figure is silhouetted. The daring display of figure composition, the landscape elements and personal drama also appear in the great unfinished masterpiece, the Adoration of the Magi, a commission from the Monks of San Donato a Scopeto. It is a complex composition, of about 250 x 250 centimetres. Leonardo did numerous drawings and preparatory studies, including a detailed one in linear perspective of the ruined classical architecture that forms part of the background. In 1482 Leonardo went to Milan at the behest of Lorenzo de' Medici in order to win favour with Ludovico il Moro, and the painting was abandoned. The third important work of this period is the Virgin of the Rocks, commissioned in Milan for the Confraternity of the Immaculate Conception. The painting, to be done with the assistance of the de Predis brothers, was to fill a large complex altarpiece. Leonardo chose to paint an apocryphal moment of the infancy of Christ when the infant John the Baptist, in protection of an angel, met the Holy Family on the road to Egypt. The painting demonstrates an eerie beauty as the graceful figures kneel in adoration around the infant Christ in a wild landscape of tumbling rock and whirling water. While the painting is quite large, about 200×120 centimetres, it is not nearly as complex as the painting ordered by the monks of San Donato, having only four figures rather than about fifty and a rocky landscape rather than architectural details. The painting was eventually finished; in fact, two versions of the painting were finished: one remained at the chapel of the Confraternity, while Leonardo took the other to France. The Brothers did not get their painting, however, nor the de Predis their payment, until the next century. Leonardo's most remarkable portrait of this period is the Lady with an Ermine, presumed to be Cecilia Gallerani (c. 1483–1490), lover of Ludovico Sforza. The painting is characterised by the pose of the figure with the head turned at a very different angle to the torso, unusual at a date when many portraits were still rigidly in profile. The ermine plainly carries symbolic meaning, relating either to the sitter, or to Ludovico who belonged to the prestigious Order of the Ermine. Leonardo's most famous painting of the 1490s is The Last Supper, commissioned for the refectory of the Convent of Santa Maria della Grazie in Milan. It represents the last meal shared by Jesus with his disciples before his capture and death, and shows the moment when Jesus has just said "one of you will betray me", and the consternation that this statement caused. The writer Matteo Bandello observed Leonardo at work and wrote that some days he would paint from dawn till dusk without stopping to eat and then not paint for three or four days at a time. This was beyond the comprehension of the prior of the convent, who hounded him until Leonardo asked Ludovico to intervene. Vasari describes how Leonardo, troubled over his ability to adequately depict the faces of Christ and the traitor Judas, told the duke that he might be obliged to use the prior as his model. The painting was acclaimed as a masterpiece of design and characterization, but it deteriorated rapidly, so that within a hundred years it was described by one viewer as "completely ruined." Leonardo, instead of using the reliable technique of fresco, had used tempera over a ground that was mainly gesso, resulting in a surface subject to mould and to flaking. Despite this, the painting remains one of the most reproduced works of art; countless copies have been made in various mediums. Toward the end of this period, in 1498 Leonardo's trompe-l'œil decoration of the Sala delle Asse was painted for the Duke of Milan in the Castello Sforzesco. In 1505, Leonardo was commissioned to paint The Battle of Anghiari in the Salone dei Cinquecento (Hall of the Five Hundred) in the Palazzo Vecchio, Florence. Leonardo devised a dynamic composition depicting four men riding raging war horses engaged in a battle for possession of a standard, at the Battle of Anghiari in 1440. Michelangelo was assigned the opposite wall to depict the Battle of Cascina. Leonardo's painting deteriorated rapidly and is now known from a copy by Rubens. Among the works created by Leonardo in the 16th century is the small portrait known as the Mona Lisa or La Gioconda, the laughing one. In the present era, it is arguably the most famous painting in the world. Its fame rests, in particular, on the elusive smile on the woman's face, its mysterious quality perhaps due to the subtly shadowed corners of the mouth and eyes such that the exact nature of the smile cannot be determined. The shadowy quality for which the work is renowned came to be called "sfumato", or Leonardo's smoke. Vasari wrote that the smile was "so pleasing that it seems more divine than human, and it was considered a wondrous thing that it was as lively as the smile of the living original." Other characteristics of the painting are the unadorned dress, in which the eyes and hands have no competition from other details; the dramatic landscape background, in which the world seems to be in a state of flux; the subdued colouring; and the extremely smooth nature of the painterly technique, employing oils laid on much like tempera, and blended on the surface so that the brushstrokes are indistinguishable. Vasari expressed that the painting's quality would make even "the most confident master ... despair and lose heart." The perfect state of preservation and the fact that there is no sign of repair or overpainting is rare in a panel painting of this date. In the painting Virgin and Child with Saint Anne, the composition again picks up the theme of figures in a landscape, which Wasserman describes as "breathtakingly beautiful" and harkens back to the Saint Jerome with the figure set at an oblique angle. What makes this painting unusual is that there are two obliquely set figures superimposed. Mary is seated on the knee of her mother, Saint Anne. She leans forward to restrain the Christ Child as he plays roughly with a lamb, the sign of his own impending sacrifice. This painting, which was copied many times, influenced Michelangelo, Raphael, and Andrea del Sarto, and through them Pontormo and Correggio. The trends in composition were adopted in particular by the Venetian painters Tintoretto and Veronese. Leonardo was a prolific draughtsman, keeping journals full of small sketches and detailed drawings recording all manner of things that took his attention. As well as the journals there exist many studies for paintings, some of which can be identified as preparatory to particular works such as The Adoration of the Magi, The Virgin of the Rocks and The Last Supper. His earliest dated drawing is a Landscape of the Arno Valley, 1473, which shows the river, the mountains, Montelupo Castle and the farmlands beyond it in great detail. Among his famous drawings are the Vitruvian Man, a study of the proportions of the human body; the Head of an Angel, for The Virgin of the Rocks in the Louvre; a botanical study of Star of Bethlehem; and a large drawing (160×100 cm) in black chalk on coloured paper of The Virgin and Child with Saint Anne and Saint John the Baptist in the National Gallery, London. This drawing employs the subtle sfumato technique of shading, in the manner of the Mona Lisa. It is thought that Leonardo never made a painting from it, the closest similarity being to The Virgin and Child with Saint Anne in the Louvre. Other drawings of interest include numerous studies generally referred to as "caricatures" because, although exaggerated, they appear to be based upon observation of live models. Vasari relates that Leonardo would look for interesting faces in public to use as models for some of his work. There are numerous studies of beautiful young men, often associated with Salaì, with the rare and much admired facial feature, the so-called "Grecian profile". These faces are often contrasted with that of a warrior. Salaì is often depicted in fancy-dress costume. Leonardo is known to have designed sets for pageants with which these may be associated. Other, often meticulous, drawings show studies of drapery. A marked development in Leonardo's ability to draw drapery occurred in his early works. Another often-reproduced drawing is a macabre sketch that was done by Leonardo in Florence in 1479 showing the body of Bernardo Baroncelli, hanged in connection with the murder of Giuliano, brother of Lorenzo de' Medici, in the Pazzi conspiracy. In his notes, Leonardo recorded the colours of the robes that Baroncelli was wearing when he died. Like the two contemporary architects Donato Bramante (who designed the Belvedere Courtyard) and Antonio da Sangallo the Elder, Leonardo experimented with designs for centrally planned churches, a number of which appear in his journals, as both plans and views, although none was ever realised. Renaissance humanism recognised no mutually exclusive polarities between the sciences and the arts, and Leonardo's studies in science and engineering are sometimes considered as impressive and innovative as his artistic work. These studies were recorded in 13,000 pages of notes and drawings, which fuse art and natural philosophy (the forerunner of modern science). They were made and maintained daily throughout Leonardo's life and travels, as he made continual observations of the world around him. Leonardo's notes and drawings display an enormous range of interests and preoccupations, some as mundane as lists of groceries and people who owed him money and some as intriguing as designs for wings and shoes for walking on water. There are compositions for paintings, studies of details and drapery, studies of faces and emotions, of animals, babies, dissections, plant studies, rock formations, whirlpools, war machines, flying machines and architecture. These notebooks – originally loose papers of different types and sizes – were largely entrusted to Leonardo's pupil and heir Francesco Melzi after the master's death. These were to be published, a task of overwhelming difficulty because of its scope and Leonardo's idiosyncratic writing. Some of Leonardo's drawings were copied by an anonymous Milanese artist for a planned treatise on art c. 1570. After Melzi's death in 1570, the collection passed to his son, the lawyer Orazio, who initially took little interest in the journals. In 1587, a Melzi household tutor named Lelio Gavardi took 13 of the manuscripts to Pisa; there, the architect Giovanni Magenta reproached Gavardi for having taken the manuscripts illicitly and returned them to Orazio. Having many more such works in his possession, Orazio gifted the volumes to Magenta. News spread of these lost works of Leonardo's, and Orazio retrieved seven of the 13 manuscripts, which he then gave to Pompeo Leoni for publication in two volumes; one of these was the Codex Atlanticus. The other six works had been distributed to a few others. After Orazio's death, his heirs sold the rest of Leonardo's possessions, and thus began their dispersal. Some works have found their way into major collections such as the Royal Library at Windsor Castle, the Louvre, the Biblioteca Nacional de España, the Victoria and Albert Museum, the Biblioteca Ambrosiana in Milan, which holds the 12-volume Codex Atlanticus, and the British Library in London, which has put a selection from the Codex Arundel (BL Arundel MS 263) online. Works have also been at Holkham Hall, the Metropolitan Museum of Art, and in the private hands of John Nicholas Brown I and Robert Lehman. The Codex Leicester is the only privately owned major scientific work of Leonardo; it is owned by Bill Gates and displayed once a year in different cities around the world. Most of Leonardo's writings are in mirror-image cursive. Since Leonardo wrote with his left hand, it was probably easier for him to write from right to left. Leonardo used a variety of shorthand and symbols, and states in his notes that he intended to prepare them for publication. In many cases a single topic is covered in detail in both words and pictures on a single sheet, together conveying information that would not be lost if the pages were published out of order. Why they were not published during Leonardo's lifetime is unknown. Leonardo's approach to science was observational: he tried to understand a phenomenon by describing and depicting it in utmost detail and did not emphasise experiments or theoretical explanation. Since he lacked formal education in Latin and mathematics, contemporary scholars mostly ignored Leonardo the scientist, although he did teach himself Latin. His keen observations in many areas were noted, such as when he wrote "Il sole non si move." ("The Sun does not move.") In the 1490s he studied mathematics under Luca Pacioli and prepared a series of drawings of regular solids in a skeletal form to be engraved as plates for Pacioli's book Divina proportione, published in 1509. While living in Milan, he studied light from the summit of Monte Rosa. Scientific writings in his notebook on fossils have been considered as influential on early palaeontology. The content of his journals suggest that he was planning a series of treatises on a variety of subjects. A coherent treatise on anatomy is said to have been observed during a visit by Cardinal Louis d'Aragon's secretary in 1517. Aspects of his work on the studies of anatomy, light and the landscape were assembled for publication by Melzi and eventually published as A Treatise on Painting in France and Italy in 1651 and Germany in 1724, with engravings based upon drawings by the Classical painter Nicolas Poussin. According to Arasse, the treatise, which in France went into 62 editions in fifty years, caused Leonardo to be seen as "the precursor of French academic thought on art." While Leonardo's experimentation followed scientific methods, a recent and exhaustive analysis of Leonardo as a scientist by Fritjof Capra argues that Leonardo was a fundamentally different kind of scientist from Galileo, Newton and other scientists who followed him in that, as a "Renaissance Man", his theorising and hypothesising integrated the arts and particularly painting. Leonardo started his study in the anatomy of the human body under the apprenticeship of Verrocchio, who demanded that his students develop a deep knowledge of the subject. As an artist, he quickly became master of topographic anatomy, drawing many studies of muscles, tendons and other visible anatomical features. As a successful artist, Leonardo was given permission to dissect human corpses at the Hospital of Santa Maria Nuova in Florence and later at hospitals in Milan and Rome. From 1510 to 1511 he collaborated in his studies with the doctor Marcantonio della Torre, professor of Anatomy at the University of Pavia. Leonardo made over 240 detailed drawings and wrote about 13,000 words toward a treatise on anatomy. Only a small amount of the material on anatomy was published in Leonardo's Treatise on painting. During the time that Melzi was ordering the material into chapters for publication, they were examined by a number of anatomists and artists, including Vasari, Cellini and Albrecht Dürer, who made a number of drawings from them. Leonardo's anatomical drawings include many studies of the human skeleton and its parts, and of muscles and sinews. He studied the mechanical functions of the skeleton and the muscular forces that are applied to it in a manner that prefigured the modern science of biomechanics. He drew the heart and vascular system, the sex organs and other internal organs, making one of the first scientific drawings of a fetus in utero. The drawings and notation are far ahead of their time, and if published would undoubtedly have made a major contribution to medical science. Leonardo also closely observed and recorded the effects of age and of human emotion on the physiology, studying in particular the effects of rage. He drew many figures who had significant facial deformities or signs of illness. Leonardo also studied and drew the anatomy of many animals, dissecting cows, birds, monkeys, bears, and frogs, and comparing in his drawings their anatomical structure with that of humans. He also made a number of studies of horses. Leonardo's dissections and documentation of muscles, nerves, and vessels helped to describe the physiology and mechanics of movement. He attempted to identify the source of 'emotions' and their expression. He found it difficult to incorporate the prevailing system and theories of bodily humours, but eventually he abandoned these physiological explanations of bodily functions. He made the observations that humours were not located in cerebral spaces or ventricles. He documented that the humours were not contained in the heart or the liver, and that it was the heart that defined the circulatory system. He was the first to define atherosclerosis and liver cirrhosis. He created models of the cerebral ventricles with the use of melted wax and constructed a glass aorta to observe the circulation of blood through the aortic valve by using water and grass seed to watch flow patterns. During his lifetime, Leonardo was also valued as an engineer. With the same rational and analytical approach that moved him to represent the human body and to investigate anatomy, Leonardo studied and designed many machines and devices. He drew their "anatomy" with unparalleled mastery, producing the first form of the modern technical drawing, including a perfected "exploded view" technique, to represent internal components. Those studies and projects collected in his codices fill more than 5,000 pages. In a letter of 1482 to the lord of Milan Ludovico il Moro, he wrote that he could create all sorts of machines both for the protection of a city and for siege. When he fled from Milan to Venice in 1499, he found employment as an engineer and devised a system of moveable barricades to protect the city from attack. In 1502, he created a scheme for diverting the flow of the Arno river, a project on which Niccolò Machiavelli also worked. He continued to contemplate the canalization of Lombardy's plains while in Louis XII's company and of the Loire and its tributaries in the company of Francis I. Leonardo's journals include a vast number of inventions, both practical and impractical. They include musical instruments, a mechanical knight, hydraulic pumps, reversible crank mechanisms, finned mortar shells, and a steam cannon. Leonardo was fascinated by the phenomenon of flight for much of his life, producing many studies, including Codex on the Flight of Birds (c. 1505), as well as plans for several flying machines, such as a flapping ornithopter and a machine with a helical rotor. In a 2003 documentary by British television station Channel Four, titled Leonardo's Dream Machines, various designs by Leonardo, such as a parachute and a giant crossbow, were interpreted and constructed. Some of those designs proved successful, whilst others fared less well when tested. Similarly, a team of engineers built ten machines designed by Leonardo in the 2009 American television series Doing DaVinci, including a fighting vehicle and a self-propelled cart. Research performed by Marc van den Broek revealed older prototypes for more than 100 inventions that are ascribed to Leonardo. Similarities between Leonardo's illustrations and drawings from the Middle Ages and from Ancient Greece and Rome, the Chinese and Persian Empires, and Egypt suggest that a large portion of Leonardo's inventions had been conceived before his lifetime. Leonardo's innovation was to combine different functions from existing drafts and set them into scenes that illustrated their utility. By reconstituting technical inventions he created something new. In his notebooks, Leonardo first stated the 'laws' of sliding friction in 1493. His inspiration for investigating friction came about in part from his study of perpetual motion, which he correctly concluded was not possible. His results were never published and the friction laws were not rediscovered until 1699 by Guillaume Amontons, with whose name they are now usually associated. For this contribution, Leonardo was named as the first of the 23 "Men of Tribology" by Duncan Dowson. Although he had no formal academic training, many historians and scholars regard Leonardo as the prime exemplar of the "Universal Genius" or "Renaissance Man", an individual of "unquenchable curiosity" and "feverishly inventive imagination." He is widely considered one of the most diversely talented individuals ever to have lived. According to art historian Helen Gardner, the scope and depth of his interests were without precedent in recorded history, and "his mind and personality seem to us superhuman, while the man himself mysterious and remote." Scholars interpret his view of the world as being based in logic, though the empirical methods he used were unorthodox for his time. Leonardo's fame within his own lifetime was such that the King of France carried him away like a trophy, and was claimed to have supported him in his old age and held him in his arms as he died. Interest in Leonardo and his work has never diminished. Crowds still queue to see his best-known artworks, T-shirts still bear his most famous drawing, and writers continue to hail him as a genius while speculating about his private life, as well as about what one so intelligent actually believed in. The continued admiration that Leonardo commanded from painters, critics and historians is reflected in many other written tributes. Baldassare Castiglione, author of Il Cortegiano (The Courtier), wrote in 1528: "...Another of the greatest painters in this world looks down on this art in which he is unequalled..." while the biographer known as "Anonimo Gaddiano" wrote, c. 1540: "His genius was so rare and universal that it can be said that nature worked a miracle on his behalf..." Vasari, in his Lives of the Artists (1568), opens his chapter on Leonardo: In the normal course of events many men and women are born with remarkable talents; but occasionally, in a way that transcends nature, a single person is marvellously endowed by Heaven with beauty, grace and talent in such abundance that he leaves other men far behind, all his actions seem inspired and indeed everything he does clearly comes from God rather than from human skill. Everyone acknowledged that this was true of Leonardo da Vinci, an artist of outstanding physical beauty, who displayed infinite grace in everything that he did and who cultivated his genius so brilliantly that all problems he studied he solved with ease. The 19th century brought a particular admiration for Leonardo's genius, causing Henry Fuseli to write in 1801: "Such was the dawn of modern art, when Leonardo da Vinci broke forth with a splendour that distanced former excellence: made up of all the elements that constitute the essence of genius..." This is echoed by A.E. Rio who wrote in 1861: "He towered above all other artists through the strength and the nobility of his talents." By the 19th century, the scope of Leonardo's notebooks was known, as well as his paintings. Hippolyte Taine wrote in 1866: "There may not be in the world an example of another genius so universal, so incapable of fulfilment, so full of yearning for the infinite, so naturally refined, so far ahead of his own century and the following centuries." Art historian Bernard Berenson wrote in 1896: Leonardo is the one artist of whom it may be said with perfect literalness: Nothing that he touched but turned into a thing of eternal beauty. Whether it be the cross section of a skull, the structure of a weed, or a study of muscles, he, with his feeling for line and for light and shade, forever transmuted it into life-communicating values. The interest in Leonardo's genius has continued unabated; experts study and translate his writings, analyse his paintings using scientific techniques, argue over attributions and search for works which have been recorded but never found. Liana Bortolon, writing in 1967, said: Because of the multiplicity of interests that spurred him to pursue every field of knowledge...Leonardo can be considered, quite rightly, to have been the universal genius par excellence, and with all the disquieting overtones inherent in that term. Man is as uncomfortable today, faced with a genius, as he was in the 16th century. Five centuries have passed, yet we still view Leonardo with awe. The Elmer Belt Library of Vinciana is a special collection at the University of California, Los Angeles. Twenty-first-century author Walter Isaacson based much of his biography of Leonardo on thousands of notebook entries, studying the personal notes, sketches, budget notations, and musings of the man whom he considers the greatest of innovators. Isaacson was surprised to discover a "fun, joyous" side of Leonardo in addition to his limitless curiosity and creative genius. On the 500th anniversary of Leonardo's death, the Louvre in Paris arranged for the largest ever single exhibit of his work, called Leonardo, between November 2019 and February 2020. The exhibit includes over 100 paintings, drawings and notebooks. Eleven of the paintings that Leonardo completed in his lifetime were included. Five of these are owned by the Louvre, but the Mona Lisa was not included because it is in such great demand among general visitors to the Louvre; it remains on display in its gallery. Vitruvian Man, however, is on display following a legal battle with its owner, the Gallerie dell'Accademia in Venice. Salvator Mundi was also not included because its Saudi owner did not agree to lease the work. The Mona Lisa, considered Leonardo's magnum opus, is often regarded as the most famous portrait ever made. The Last Supper is the most reproduced religious painting of all time, and Leonardo's Vitruvian Man drawing is also considered a cultural icon. More than a decade of analysis of Leonardo's genetic genealogy, conducted by Alessandro Vezzosi and Agnese Sabato, came to a conclusion in mid-2021. It was determined that the artist has 14 living male relatives. The work could also help determine the authenticity of remains thought to belong to Leonardo. While Leonardo was certainly buried in the collegiate church of Saint Florentin at the Château d'Amboise in 12 August 1519, the current location of his remains is unclear. Much of Château d'Amboise was damaged during the French Revolution, leading to the church's demolition in 1802. Some of the graves were destroyed in the process, scattering the bones interred there and thereby leaving the whereabouts of Leonardo's remains subject to dispute; a gardener may have even buried some in the corner of the courtyard. In 1863, fine-arts inspector general Arsène Houssaye received an imperial commission to excavate the site and discovered a partially complete skeleton with a bronze ring on one finger, white hair, and stone fragments bearing the inscriptions "EO", "AR", "DUS", and "VINC" – interpreted as forming "Leonardus Vinci". The skull's eight teeth corresponds to someone with approximately the same age and a silver shield found near the bones depicts a beardless Francis I, corresponding to the king's appearance during Leonardo's time in France. Houssaye postulated that the unusually large skull was an indicator of Leonardo's intelligence; author Charles Nicholl describes this as a "dubious phrenological deduction". At the same time, Houssaye noted some issues with his observations, including that the feet were turned toward the high altar, a practice generally reserved for laymen, and that the skeleton of 1.73 metres (5.7 ft) seemed too short. Art historian Mary Margaret Heaton wrote in 1874 that the height would be appropriate for Leonardo. The skull was allegedly presented to Napoleon III before being returned to the Château d'Amboise, where they were re-interred in the chapel of Saint Hubert in 1874. A plaque above the tomb states that its contents are only presumed to be those of Leonardo. It has since been theorized that the folding of the skeleton's right arm over the head may correspond to the paralysis of Leonardo's right hand. In 2016, it was announced that DNA tests would be conducted to determine whether the attribution is correct. The DNA of the remains will be compared to that of samples collected from Leonardo's work and his half-brother Domenico's descendants; it may also be sequenced. In 2019, documents were published revealing that Houssaye had kept the ring and a lock of hair. In 1925, his great-grandson sold these to an American collector. Sixty years later, another American acquired them, leading to their being displayed at the Leonardo Museum in Vinci beginning on 2 May 2019, the 500th anniversary of the artist's death. General Dates of works Early Modern Books Journals and encyclopedia articles See Kemp (2003) and Bambach (2019, pp. 442–579) for extensive bibliographies General Works
{{Year nav|1844}} {{Year in various calendars|1844}} {{Year article header|1844}} == Events == * [[February 27]] – The [[Dominican Republic]] gains independence from [[Haiti]]. == New Books == * ''The Luck of Barry Lyndon'' by [[William Makepeace Thackeray]] == Births == *[[May 22]] – [[Mary Cassatt]], American artist (d. [[1926]]) *[[July 4]] – [[Edmonia Lewis]], American [[sculptor]] (d. [[1907]]) *[[August 6]] – [[Alfred, Duke of Saxe-Coburg and Gotha]] (d. [[1900]]) == Deaths == * [[June 27]] – [[Joseph Smith, Jr]], founder of [[Mormonism]]he was raped and murderd [[Category:1844| ]]
1844 (MDCCCXLIV) was a leap year starting on Monday of the Gregorian calendar and a leap year starting on Saturday of the Julian calendar, the 1844th year of the Common Era (CE) and Anno Domini (AD) designations, the 844th year of the 2nd millennium, the 44th year of the 19th century, and the 5th year of the 1840s decade. As of the start of 1844, the Gregorian calendar was 12 days ahead of the Julian calendar, which remained in localized use until 1923. In the Philippines, this was the only leap year with 365 days, when December 31 was skipped as December 30 was immediately followed by January 1, 1845, the next day after. The change also applied to Caroline Islands, Guam, Marianas Islands, Marshall Islands, and Palau as part of the Captaincy General of the Philippines. And became the first places on Earth to redraw the International Date Line.
{{Year nav|1902}} {{Year in various calendars|1902}} '''1902''' ([[Roman numerals|MCMII]]) was a [[common year starting on Wednesday]] in the [[Gregorian calendar]] and a [[common year starting on Tuesday]] in the [[Julian calendar]]. == Events == * [[July 11]] – [[Arthur Balfour]] becomes [[Prime Minister]] of the [[United Kingdom]] * [[Marie Curie|Marie]] and [[Pierre Curie]] discover [[radium]] * The [[Real Madrid]] football club starts in [[Spain]] * Statue of [[Boudica]] unveiled in [[Westminster]], [[London]] == Births == === January === [[File:Georgy Malenkov 1964.jpg|thumb|110px|[[Georgy Malenkov]]]] * [[January 8]] – [[Georgy Malenkov]], Soviet politician (d. [[1988]]) === February === [[File:Langston Hughes 1936.jpg|thumb|110px|[[Langston Hughes]]]] * [[February 1]] – [[Langston Hughes]], American writer (d. [[1967]]) * [[February 4]] – [[Charles Lindbergh]], American pilot (d. [[1974]]) * [[February 10]] – [[Walter Houser Brattain]], American physicist (d. [[1987]]) * [[February 27]] – [[John Steinbeck]], American writer (d. [[1968]]) === March === [[File:Thomas Dewey.jpg|thumb|110px|[[Thomas E. Dewey]]]] * [[March 24]] – [[Thomas E. Dewey]], American politician (d. [[1971]]) === April === === May === === June === [[File:Barbara McClintock (1902-1992).jpg|thumb|110px|[[Barbara McClintock]]]] * [[June 16]] – [[Barbara McClintock]], American [[geneticist]] (d. [[1992]]) === July === * [[July 10]] – [[Kurt Alder]], German chemist (d. [[1958]]) === August === [[File:Paul Dirac, 1933.jpg|thumb|110px|[[Paul Dirac]]]] * [[August 8]] – [[Paul Dirac]], English physicist (d. [[1984]]) === September === [[File:Portrait of Ruhollah Khomeini By Mohammad Sayyad.jpg|thumb|110px|[[Ayatollah Khomeini]]]] * [[September 21]] – [[Ruhollah Khomeini]], revolutionary leader (d. [[1989]]) === October === [[File:Ray kroc 1976.jpg|thumb|110px|[[Ray Kroc]]]] * [[October 5]] – [[Ray Kroc]], American fast food entrepreneur (d. [[1984]]) === November === * [[November 9]] - [[Anthony Asquith]], British director (d. [[1968]]) === December === [[File:Margaret Hamilton 1966.jpg|thumb|110px|[[Margaret Hamilton]]]] * [[December 5]] – [[Strom Thurmond]], American politician (d. [[2003]]) * [[December 9]] – [[Margaret Hamilton]], American actress (d. [[1985]]) == Deaths == * [[April 30]] - [[Swami Vivekananda]] - [[India|Indian]] social reformer (b. [[1863]]) * [[December 14]] – [[Julia Grant]] – former [[First Lady of the United States]] (b. [[1826]]) * [[September 16]] – [[Levi Strauss]], American [[Business|business person]] (b. [[1829]]) == Hits songs == * "[[The Entertainer]]" m. [[Scott Joplin]] [[Category:1902| ]]
1902 (MCMII) was a common year starting on Wednesday of the Gregorian calendar and a common year starting on Tuesday of the Julian calendar, the 1902nd year of the Common Era (CE) and Anno Domini (AD) designations, the 902nd year of the 2nd millennium, the 2nd year of the 20th century, and the 3rd year of the 1900s decade. As of the start of 1902, the Gregorian calendar was 13 days ahead of the Julian calendar, which remained in localized use until 1923.
{{Year nav|1895}} {{Year in various calendars|1895}} {{Year article header|1895}} == Events == * [[Wilhelm Röntgen|Roentgen]] discovers [[X-ray]]s * [[Guglielmo Marconi|Marconi]] invents the [[wireless telegraph]] * [[China]] and [[Japan]] sign the [[Treaty of Shimonoseki]] * [[José Paulino Gomes]] was born on [[August 4|4 August]] == Births == === January === === February === === March === === April === === May === === June === === July === === August === * [[August 4]] {{Endash}} [[José Paulino Gomes]], Brazilian supercentenarian (d. [[2023]]) * [[August 24]] – [[Abdul Rahman of Negeri Sembilan]], 1st [[Yang di-Pertuan Agong]] (d. [[1960]]) === September === === October === * [[October 8]] – [[Juan Perón]], [[President of Argentina]] (d. [[1974]]) * [[October 25]] – [[Levi Eshkol]], 3rd [[Prime Minister of Israel]] (d. [[1969]]) === November === === December === [[Category:1895| ]]
1895 (MDCCCXCV) was a common year starting on Tuesday of the Gregorian calendar and a common year starting on Sunday of the Julian calendar, the 1895th year of the Common Era (CE) and Anno Domini (AD) designations, the 895th year of the 2nd millennium, the 95th year of the 19th century, and the 6th year of the 1890s decade. As of the start of 1895, the Gregorian calendar was 12 days ahead of the Julian calendar, which remained in localized use until 1923.
{{Year nav|1796}} {{Year in various calendars|1796}} {{Year article header|1796}} == Events == * An English doctor invents a [[vaccination]] against [[smallpox]]. == Births == *[[January 7]]-[[Princess Charlotte Augusta of Wales]] *[[July 26]] - [[George Catlin]], American painter *[[December 27]] – [[Mirza Ghalib]], [[poet]] of [[Urdu]] [[Category:1796| ]]
1796 (MDCCXCVI) was a leap year starting on Friday of the Gregorian calendar and a leap year starting on Tuesday of the Julian calendar, the 1796th year of the Common Era (CE) and Anno Domini (AD) designations, the 796th year of the 2nd millennium, the 96th year of the 18th century, and the 7th year of the 1790s decade. As of the start of 1796, the Gregorian calendar was 11 days ahead of the Julian calendar, which remained in localized use until 1923.
{{Year nav|1879}} {{Year in various calendars|1879}} {{Year article header|1879}} == Events == * [[Thomas Edison]] invents the [[light bulb]]. * [[Rutherford B. Hayes]] State of the Union Address. == Births == * [[March 14]] &ndash; [[Albert Einstein]], German scientist (died [[1955]]) * [[August 8]] – [[Emiliano Zapata]], Mexican revolutionary * [[August 15]] &ndash; [[Ethel Barrymore]], American actress (died [[1959]]) * [[November 7]] – [[Leon Trotsky]], [[Ukraine|Ukrainian]] [[revolution]]ary and [[Marxism|Marxist]] (died [[1940]]) * [[December 22]] &ndash; [[Sydney Greenstreet]], British actor (died [[1958]]) * [[Samuel Goldwyn]], movie producer * [[Simon Rodia]], creator of the [[Watts Towers]] == Deaths == * [[January 26]] - [[Julia Margaret Cameron]], [[British India]]n photographer (b. [[1815]]) * [[February 11]] – [[Honoré Daumier]], French caricaturist and painter (b. [[1808]]) * [[February 23]] – [[Albrecht Graf von Roon]], [[Prime Minister of Prussia]] (b. [[1803]]) * [[February 25]] – [[Charles Peace]], British criminal (executed) (b. [[1832]]) * [[March 1]] – [[Joachim Heer]], Swiss politician (b. [[1825]]) * [[March 30]] – [[Thomas Couture]], French painter and teacher (b. [[1815]]) * [[April 30]] – [[Sarah Josepha Hale]], American writer (b. [[1788]]) * [[June 1]] – [[Napoleon Eugene, Prince Imperial]], son of French Emperor [[Napoleon III]] (b. [[1856]]) * [[June 27]] - [[Sir]] [[John Lawrence]], [[British Indian]] administrator (b. [[1811]]) * [[August 30]] – [[John Bell Hood]], America Confederate general (b. [[1831]]) * [[November 5]] – [[James Clerk Maxwell]], Scottish physicist (b. [[1831]]) * [[August 11]] – [[George Willison Adams]], Ohio abolitionist (b.[[1799]]) [[Category:1879| ]]
1879 (MDCCCLXXIX) was a common year starting on Wednesday of the Gregorian calendar and a common year starting on Monday of the Julian calendar, the 1879th year of the Common Era (CE) and Anno Domini (AD) designations, the 879th year of the 2nd millennium, the 79th year of the 19th century, and the 10th and last year of the 1870s decade. As of the start of 1879, the Gregorian calendar was 12 days ahead of the Julian calendar, which remained in localized use until 1923.
{{Year nav|1769}} {{Year in various calendars|1769}} {{Year article header|1769}} == Events == * [[James Watt]] invents an [[engine]] driven by [[steam]] == Births == * [[August 15]] – [[Napoleon]] [[Category:1769| ]]
1769 (MDCCLXIX) was a common year starting on Sunday of the Gregorian calendar and a common year starting on Thursday of the Julian calendar, the 1769th year of the Common Era (CE) and Anno Domini (AD) designations, the 769th year of the 2nd millennium, the 69th year of the 18th century, and the 10th and last year of the 1760s decade. As of the start of 1769, the Gregorian calendar was 11 days ahead of the Julian calendar, which remained in localized use until 1923.
{{Year nav|1807}} {{Year in various calendars|1807}} {{Year article header|1807}} == Events == *[[Robert Fulton]] builds the [[steamship]] [[Category:1807| ]]
1807 (MDCCCVII) was a common year starting on Thursday of the Gregorian calendar and a common year starting on Tuesday of the Julian calendar, the 1807th year of the Common Era (CE) and Anno Domini (AD) designations, the 807th year of the 2nd millennium, the 7th year of the 19th century, and the 8th year of the 1800s decade. As of the start of 1807, the Gregorian calendar was 12 days ahead of the Julian calendar, which remained in localized use until 1923.
{{Year nav|1885}} {{Year in various calendars|1885}} {{Year article header|1885}} == Events == * [[Benz]] builds the first [[single-cylinder engine]] for car. * John K Starley creates the first genuine [[bicycle]], the [[Rover (car)|Rover]]. * [[John Boyd Dunlop]] invents the pneumatic [[tire]]. * [[W.S. Gilbert]] & [[Arthur Sullivan]]'s ''[[The Mikado]]'' * ''[[The Times]]'' reports that "A lady well known in literary and scientific circles" has been [[cremation|cremated]] by the [[Cremation Society]] in [[Woking]], [[Surrey]]. She is the first person to be officially cremated in the [[United Kingdom]]. * [[Cholera epidemic]] in [[Spain]] – one of the victims is the king [[Alfonso XII of Spain|Alfonso XII]]. * [[Third Burmese War]] begins. * [[Clint Eastwood defeats Mad Dog Tannen]]. * [[Sitting Bull]] joins [[Buffalo Bill]]. * [[Randolph Churchill]] becomes a [[viceroy of India]]. * [[Nikola Tesla]] sells a number of his patents to [[George Westinghouse]]. * [[William Stanley (physicist)|William Stanley, Jr.]] builds the first practical [[alternating current]] [[transformer]] device, the [[induction coil]]. * [[Local anesthetic]]. * First [[skyscraper]] – [[Home Insurance Building]] in [[Chicago]], [[Illinois]], USA (10 floors) is built. * [[Bicycle Playing Cards]] are first produced. * [[Grover Cleveland]] replaces [[Chester A. Arthur]] as [[President (United States)|President of the United States]] == Births == * [[January 8]] – [[John Curtin]], [[Australia]]'s 14th Prime Minister * [[September 17]] – [[George Cleveland]], [[Canadians|Canadian]] [[actor]] == Deaths == * [[March 31]] – [[Charlotte Brontë]], [[Great Britain|British]] [[writer]] [[Category:1885| ]]
1885 (MDCCCLXXXV) was a common year starting on Thursday of the Gregorian calendar and a common year starting on Tuesday of the Julian calendar, the 1885th year of the Common Era (CE) and Anno Domini (AD) designations, the 885th year of the 2nd millennium, the 85th year of the 19th century, and the 6th year of the 1880s decade. As of the start of 1885, the Gregorian calendar was 12 days ahead of the Julian calendar, which remained in localized use until 1923.
{{Year nav|1837}} {{Year in various calendars|1837}} {{Year article header|1837}} == Events == * [[Samuel Morse]] invents the electric [[telegraph]]. * [[Michigan]] becomes the 26th state. [[Category:1837| ]]
1837 (MDCCCXXXVII) was a common year starting on Sunday of the Gregorian calendar and a common year starting on Friday of the Julian calendar, the 1837th year of the Common Era (CE) and Anno Domini (AD) designations, the 837th year of the 2nd millennium, the 37th year of the 19th century, and the 8th year of the 1830s decade. As of the start of 1837, the Gregorian calendar was 12 days ahead of the Julian calendar, which remained in localized use until 1923.
[[File:Platinum nuggets.jpg|thumb|Some platinum nuggets from nature.]] {{wiktionary}} '''Platinum''' is a [[wikt:soft|soft]], [[heavy]], white [[metal]]. It is a [[precious metal]]. In the past it [[price|cost]] more than [[gold]], but since the late 2010s, gold cost more. In [[chemistry]], platinum is [[atomic number|element number]] 78, and its [[atoms]] have an [[atomic weight]] of 195 [[a.m.u.]]. The [[symbol]] for '''platinum''' is Pt, from [[Spanish language|Spanish]] ''platina'' meaning "little [[silver]]". Platinum is very [[malleable]] and [[wikt:ductile|ductile]], which means it can be hammered into thin [[sheets]] and it can be pulled into [[wire]]. Platinum is very [[stable]]. [[Acid]]s do not [[wikt:affect|affect]] platinum. Several other metals are similar and are called the "platinum group". The most common use of Platinum is in a vehicle's [[catalytic converter]]. {{Periodic Table}} {{chem-stub}} [[Category:Transition metals]] [[Category:Catalysts]]
Platinum is a chemical element; it has symbol Pt and atomic number 78. It is a dense, malleable, ductile, highly unreactive, precious, silverish-white transition metal. Its name originates from Spanish platina, a diminutive of plata "silver". Platinum is a member of the platinum group of elements and group 10 of the periodic table of elements. It has six naturally occurring isotopes. It is one of the rarer elements in Earth's crust, with an average abundance of approximately 5 μg/kg. It occurs in some nickel and copper ores along with some native deposits, mostly in South Africa, which accounts for ~80% of the world production. Because of its scarcity in Earth's crust, only a few hundred tonnes are produced annually, and given its important uses, it is highly valuable and is a major precious metal commodity. Platinum is one of the least reactive metals. It has remarkable resistance to corrosion, even at high temperatures, and is therefore considered a noble metal. Consequently, platinum is often found chemically uncombined as native platinum. Because it occurs naturally in the alluvial sands of various rivers, it was first used by pre-Columbian South American natives to produce artifacts. It was referenced in European writings as early as the 16th century, but it was not until Antonio de Ulloa published a report on a new metal of Colombian origin in 1748 that it began to be investigated by scientists. Platinum is used in catalytic converters, laboratory equipment, electrical contacts and electrodes, platinum resistance thermometers, dentistry equipment, and jewelry. Platinum is used in the glass industry to manipulate molten glass which does not "wet" platinum. As a heavy metal, it leads to health problems upon exposure to its salts; but due to its corrosion resistance, metallic platinum has not been linked to adverse health effects. Compounds containing platinum, such as cisplatin, oxaliplatin and carboplatin, are applied in chemotherapy against certain types of cancer. Pure platinum is currently less expensive than pure gold, having been so continuously since 2015, but has been twice as expensive or more, mostly prior to 2008. In early 2021, the value of platinum ranged from US$1,055 to US$1,320 per troy ounce. Pure platinum is a lustrous, ductile, and malleable, silver-white metal. Platinum is more ductile than gold, silver or copper, thus being the most ductile of pure metals, but it is less malleable than gold. Its physical characteristics and chemical stability make it useful for industrial applications. Its resistance to wear and tarnish is well suited to use in fine jewellery. Platinum has excellent resistance to corrosion. Bulk platinum does not oxidize in air at any temperature, but it forms a thin surface film of PtO2 that can be easily removed by heating to about 400 °C. The most common oxidation states of platinum are +2 and +4. The +1 and +3 oxidation states are less common, and are often stabilized by metal bonding in bimetallic (or polymetallic) species. Tetracoordinate platinum(II) compounds tend to adopt 16-electron square planar geometries. Although elemental platinum is generally unreactive, it is attacked by chlorine, bromine, iodine, and sulfur. It reacts vigorously with fluorine at 500 °C (932 °F) to form platinum tetrafluoride. Platinum is insoluble in hydrochloric and nitric acid, but dissolves in hot aqua regia (a mixture of nitric and hydrochloric acids), to form aqueous chloroplatinic acid, H2PtCl6: As a soft acid, the Pt ion has a great affinity for sulfide and sulfur ligands. Numerous DMSO complexes have been reported and care is taken in the choosing of reaction solvents. In 2007, the German scientist Gerhard Ertl won the Nobel Prize in Chemistry for determining the detailed molecular mechanisms of the catalytic oxidation of carbon monoxide over platinum (catalytic converter). Platinum has six naturally occurring isotopes: Pt, Pt, Pt, Pt, Pt, and Pt. The most abundant of these is Pt, comprising 33.83% of all platinum. It is the only stable isotope with a non-zero spin. The spin of /2 and other favourable magnetic properties of the nucleus are utilised in Pt NMR. Due to its spin and large abundance, Pt satellite peaks are also often observed in H and P NMR spectroscopy (e.g., for Pt-phosphine and Pt-alkyl complexes). Pt is the least abundant at only 0.01%. Of the naturally occurring isotopes, only Pt is unstable, though it decays with a half-life of 6.5×10 years, causing an activity of 15 Bq/kg of natural platinum. Other isotopes can undergo alpha decay, but their decay has never been observed, therefore they are considered stable. Platinum also has 38 synthetic isotopes ranging in atomic mass from 165 to 208, making the total number of known isotopes 44. The least stable of these are Pt and Pt, with half-lives of 260 µs, whereas the most stable is Pt with a half-life of 50 years. Most platinum isotopes decay by some combination of beta decay and alpha decay. Pt, Pt, and Pt decay primarily by electron capture. Pt and Pt are predicted to have energetically favorable double beta decay paths. Platinum is an extremely rare metal, occurring at a concentration of only 0.005 ppm in Earth's crust. It is sometimes mistaken for silver. Platinum is often found chemically uncombined as native platinum and as alloy with the other platinum-group metals and iron mostly. Most often the native platinum is found in secondary deposits in alluvial deposits. The alluvial deposits used by pre-Columbian people in the Chocó Department, Colombia are still a source for platinum-group metals. Another large alluvial deposit is in the Ural Mountains, Russia, and it is still mined. In nickel and copper deposits, platinum-group metals occur as sulfides (e.g., (Pt,Pd)S), tellurides (e.g., PtBiTe), antimonides (PdSb), and arsenides (e.g. PtAs2), and as end alloys with nickel or copper. Platinum arsenide, sperrylite (PtAs2), is a major source of platinum associated with nickel ores in the Sudbury Basin deposit in Ontario, Canada. At Platinum, Alaska, about 17,000 kg (550,000 ozt) was mined between 1927 and 1975. The mine ceased operations in 1990. The rare sulfide mineral cooperite, (Pt,Pd,Ni)S, contains platinum along with palladium and nickel. Cooperite occurs in the Merensky Reef within the Bushveld complex, Gauteng, South Africa. In 1865, chromites were identified in the Bushveld region of South Africa, followed by the discovery of platinum in 1906. In 1924, the geologist Hans Merensky discovered a large supply of platinum in the Bushveld Igneous Complex in South Africa. The specific layer he found, named the Merensky Reef, contains around 75% of the world's known platinum. The large copper–nickel deposits near Norilsk in Russia, and the Sudbury Basin, Canada, are the two other large deposits. In the Sudbury Basin, the huge quantities of nickel ore processed make up for the fact platinum is present as only 0.5 ppm in the ore. Smaller reserves can be found in the United States, for example in the Absaroka Range in Montana. In 2010, South Africa was the top producer of platinum, with an almost 77% share, followed by Russia at 13%; world production in 2010 was 192,000 kg (423,000 lb). Large platinum deposits are present in the state of Tamil Nadu, India. Platinum exists in higher abundances on the Moon and in meteorites. Correspondingly, platinum is found in slightly higher abundances at sites of bolide impact on Earth that are associated with resulting post-impact volcanism, and can be mined economically; the Sudbury Basin is one such example. Hexachloroplatinic acid mentioned above is probably the most important platinum compound, as it serves as the precursor for many other platinum compounds. By itself, it has various applications in photography, zinc etchings, indelible ink, plating, mirrors, porcelain coloring, and as a catalyst. Treatment of hexachloroplatinic acid with an ammonium salt, such as ammonium chloride, gives ammonium hexachloroplatinate, which is relatively insoluble in ammonium solutions. Heating this ammonium salt in the presence of hydrogen reduces it to elemental platinum. Potassium hexachloroplatinate is similarly insoluble, and hexachloroplatinic acid has been used in the determination of potassium ions by gravimetry. When hexachloroplatinic acid is heated, it decomposes through platinum(IV) chloride and platinum(II) chloride to elemental platinum, although the reactions do not occur stepwise: All three reactions are reversible. Platinum(II) and platinum(IV) bromides are known as well. Platinum hexafluoride is a strong oxidizer capable of oxidizing oxygen. Platinum(IV) oxide, PtO2, also known as "Adams' catalyst", is a black powder that is soluble in potassium hydroxide (KOH) solutions and concentrated acids. PtO2 and the less common PtO both decompose upon heating. Platinum(II,IV) oxide, Pt3O4, is formed in the following reaction: Unlike palladium acetate, platinum(II) acetate is not commercially available. Where a base is desired, the halides have been used in conjunction with sodium acetate. The use of platinum(II) acetylacetonate has also been reported. Several barium platinides have been synthesized in which platinum exhibits negative oxidation states ranging from −1 to −2. These include BaPt, Ba3Pt2, and Ba2Pt. Caesium platinide, Cs2Pt, a dark-red transparent crystalline compound has been shown to contain Pt anions. Platinum also exhibits negative oxidation states at surfaces reduced electrochemically. The negative oxidation states exhibited by platinum are unusual for metallic elements, and they are attributed to the relativistic stabilization of the 6s orbitals. It shares this unusual property with its neighbor gold, which can also form stable aurides, compounds containing the elusive Au anion. Aurides have been isolated with caesium, rubidium, and potassium. It is predicted that even the cation PtO4 in which platinum exists in the +10 oxidation state may be achievable. Zeise's salt, containing an ethylene ligand, was one of the first organometallic compounds discovered. Dichloro(cycloocta-1,5-diene)platinum(II) is a commercially available olefin complex, which contains easily displaceable cod ligands ("cod" being an abbreviation of 1,5-cyclooctadiene). The cod complex and the halides are convenient starting points to platinum chemistry. Cisplatin, or cis-diamminedichloridoplatinum(II) is the first of a series of square planar platinum(II)-containing chemotherapy drugs. Others include carboplatin and oxaliplatin. These compounds are capable of crosslinking DNA, and kill cells by similar pathways to alkylating chemotherapeutic agents. (Side effects of cisplatin include nausea and vomiting, hair loss, tinnitus, hearing loss, and nephrotoxicity.) Organoplatinum compounds such as the above antitumour agents, as well as soluble inorganic platinum complexes, are routinely characterised using Pt nuclear magnetic resonance spectroscopy. Archaeologists have discovered traces of platinum in the gold used in ancient Egyptian burials as early as 1200 BCE. For example, a small box from burial of Shepenupet II was found to be decorated with gold-platinum hieroglyphics. However, the extent of early Egyptians' knowledge of the metal is unclear. It is quite possible they did not recognize there was platinum in their gold. The metal was used by Native Americans near modern-day Esmeraldas, Ecuador to produce artifacts of a white gold-platinum alloy. Archeologists usually associate the tradition of platinum-working in South America with the La Tolita Culture (c. 600 BCE – 200 CE), but precise dates and location are difficult, as most platinum artifacts from the area were bought secondhand through the antiquities trade rather than obtained by direct archeological excavation. To work the metal, they would combine gold and platinum powders by sintering. The resulting gold–platinum alloy would then be soft enough to shape with tools. The platinum used in such objects was not the pure element, but rather a naturally occurring mixture of the platinum group metals, with small amounts of palladium, rhodium, and iridium. The first European reference to platinum appears in 1557 in the writings of the Italian humanist Julius Caesar Scaliger as a description of an unknown noble metal found between Darién and Mexico, "which no fire nor any Spanish artifice has yet been able to liquefy". From their first encounters with platinum, the Spanish generally saw the metal as a kind of impurity in gold, and it was treated as such. It was often simply thrown away, and there was an official decree forbidding the adulteration of gold with platinum impurities. In 1735, Antonio de Ulloa and Jorge Juan y Santacilia saw Native Americans mining platinum while the Spaniards were travelling through Colombia and Peru for eight years. Ulloa and Juan found mines with the whitish metal nuggets and took them home to Spain. Antonio de Ulloa returned to Spain and established the first mineralogy lab in Spain and was the first to systematically study platinum, which was in 1748. His historical account of the expedition included a description of platinum as being neither separable nor calcinable. Ulloa also anticipated the discovery of platinum mines. After publishing the report in 1748, Ulloa did not continue to investigate the new metal. In 1758, he was sent to superintend mercury mining operations in Huancavelica. In 1741, Charles Wood, a British metallurgist, found various samples of Colombian platinum in Jamaica, which he sent to William Brownrigg for further investigation. In 1750, after studying the platinum sent to him by Wood, Brownrigg presented a detailed account of the metal to the Royal Society, stating that he had seen no mention of it in any previous accounts of known minerals. Brownrigg also made note of platinum's extremely high melting point and refractoriness toward borax. Other chemists across Europe soon began studying platinum, including Andreas Sigismund Marggraf, Torbern Bergman, Jöns Jakob Berzelius, William Lewis, and Pierre Macquer. In 1752, Henrik Scheffer published a detailed scientific description of the metal, which he referred to as "white gold", including an account of how he succeeded in fusing platinum ore with the aid of arsenic. Scheffer described platinum as being less pliable than gold, but with similar resistance to corrosion. Karl von Sickingen researched platinum extensively in 1772. He succeeded in making malleable platinum by alloying it with gold, dissolving the alloy in hot aqua regia, precipitating the platinum with ammonium chloride, igniting the ammonium chloroplatinate, and hammering the resulting finely divided platinum to make it cohere. Franz Karl Achard made the first platinum crucible in 1784. He worked with the platinum by fusing it with arsenic, then later volatilizing the arsenic. Because the other platinum-family members were not discovered yet (platinum was the first in the list), Scheffer and Sickingen made the false assumption that due to its hardness—which is slightly more than for pure iron—platinum would be a relatively non-pliable material, even brittle at times, when in fact its ductility and malleability are close to that of gold. Their assumptions could not be avoided because the platinum they experimented with was highly contaminated with minute amounts of platinum-family elements such as osmium and iridium, amongst others, which embrittled the platinum alloy. Alloying this impure platinum residue called "plyoxen" with gold was the only solution at the time to obtain a pliable compound, but nowadays, very pure platinum is available and extremely long wires can be drawn from pure platinum, very easily, due to its crystalline structure, which is similar to that of many soft metals. In 1786, Charles III of Spain provided a library and laboratory to Pierre-François Chabaneau to aid in his research of platinum. Chabaneau succeeded in removing various impurities from the ore, including gold, mercury, lead, copper, and iron. This led him to believe he was working with a single metal, but in truth the ore still contained the yet-undiscovered platinum-group metals. This led to inconsistent results in his experiments. At times, the platinum seemed malleable, but when it was alloyed with iridium, it would be much more brittle. Sometimes the metal was entirely incombustible, but when alloyed with osmium, it would volatilize. After several months, Chabaneau succeeded in producing 23 kilograms of pure, malleable platinum by hammering and compressing the sponge form while white-hot. Chabeneau realized the infusibility of platinum would lend value to objects made of it, and so started a business with Joaquín Cabezas producing platinum ingots and utensils. This started what is known as the "platinum age" in Spain. Platinum, along with the rest of the platinum-group metals, is obtained commercially as a by-product from nickel and copper mining and processing. During electrorefining of copper, noble metals such as silver, gold and the platinum-group metals as well as selenium and tellurium settle to the bottom of the cell as "anode mud", which forms the starting point for the extraction of the platinum-group metals. If pure platinum is found in placer deposits or other ores, it is isolated from them by various methods of subtracting impurities. Because platinum is significantly denser than many of its impurities, the lighter impurities can be removed by simply floating them away in a liquid. Platinum is paramagnetic, whereas nickel and iron are both ferromagnetic. These two impurities are thus removed by running an electromagnet over the mixture. Because platinum has a higher melting point than most other substances, many impurities can be burned or melted away without melting the platinum. Finally, platinum is resistant to hydrochloric and sulfuric acids, whereas other substances are readily attacked by them. Metal impurities can be removed by stirring the mixture in either of the two acids and recovering the remaining platinum. One suitable method for purification for the raw platinum, which contains platinum, gold, and the other platinum-group metals, is to process it with aqua regia, in which palladium, gold and platinum are dissolved, whereas osmium, iridium, ruthenium and rhodium stay unreacted. The gold is precipitated by the addition of iron(II) chloride and after filtering off the gold, the platinum is precipitated as ammonium chloroplatinate by the addition of ammonium chloride. Ammonium chloroplatinate can be converted to platinum by heating. Unprecipitated hexachloroplatinate(IV) may be reduced with elemental zinc, and a similar method is suitable for small scale recovery of platinum from laboratory residues. Mining and refining platinum has environmental impacts. Of the 218 tonnes of platinum sold in 2014, 98 tonnes were used for vehicle emissions control devices (45%), 74.7 tonnes for jewelry (34%), 20.0 tonnes for chemical production and petroleum refining (9.2%), and 5.85 tonnes for electrical applications such as hard disk drives (2.7%). The remaining 28.9 tonnes went to various other minor applications, such as medicine and biomedicine, glassmaking equipment, investment, electrodes, anticancer drugs, oxygen sensors, spark plugs and turbine engines. The most common use of platinum is as a catalyst in chemical reactions, often as platinum black. It has been employed as a catalyst since the early 19th century, when platinum powder was used to catalyze the ignition of hydrogen. Its most important application is in automobiles as a catalytic converter, which allows the complete combustion of low concentrations of unburned hydrocarbons from the exhaust into carbon dioxide and water vapor. Platinum is also used in the petroleum industry as a catalyst in a number of separate processes, but especially in catalytic reforming of straight-run naphthas into higher-octane gasoline that becomes rich in aromatic compounds. PtO2, also known as Adams' catalyst, is used as a hydrogenation catalyst, specifically for vegetable oils. Platinum also strongly catalyzes the decomposition of hydrogen peroxide into water and oxygen and it is used in fuel cells as a catalyst for the reduction of oxygen. From 1889 to 1960, the meter was defined as the length of a platinum-iridium (90:10) alloy bar, known as the international prototype meter. The previous bar was made of platinum in 1799. Until May 2019, the kilogram was defined as the mass of the international prototype of the kilogram, a cylinder of the same platinum-iridium alloy made in 1879. The Standard Platinum Resistance Thermometer (SPRT) is one of the four types of thermometers used to define the International Temperature Scale of 1990 (ITS-90), the international calibration standard for temperature measurements. The resistance wire in the thermometer is made of pure platinum (NIST manufactured the wires from platinum bar stock with a chemical purity of 99.999% by weight). In addition to laboratory uses, Platinum Resistance Thermometry (PRT) also has many industrial applications, industrial standards include ASTM E1137 and IEC 60751. The standard hydrogen electrode also uses a platinized platinum electrode due to its corrosion resistance, and other attributes. Platinum is a precious metal commodity; its bullion has the ISO currency code of XPT. Coins, bars, and ingots are traded or collected. Platinum finds use in jewellery, usually as a 90–95% alloy, due to its inertness. It is used for this purpose for its prestige and inherent bullion value. Jewellery trade publications advise jewellers to present minute surface scratches (which they term patina) as a desirable feature in an attempt to enhance value of platinum products. In watchmaking, Vacheron Constantin, Patek Philippe, Rolex, Breitling, and other companies use platinum for producing their limited edition watch series. Watchmakers appreciate the unique properties of platinum, as it neither tarnishes nor wears out (the latter quality relative to gold). During periods of sustained economic stability and growth, the price of platinum tends to be as much as twice the price of gold, whereas during periods of economic uncertainty, the price of platinum tends to decrease due to reduced industrial demand, falling below the price of gold. Gold prices are more stable in slow economic times, as gold is considered a safe haven. Although gold is also used in industrial applications, especially in electronics due to its use as a conductor, its demand is not so driven by industrial uses. In the 18th century, platinum's rarity made King Louis XV of France declare it the only metal fit for a king. In the laboratory, platinum wire is used for electrodes; platinum pans and supports are used in thermogravimetric analysis because of the stringent requirements of chemical inertness upon heating to high temperatures (~1000 °C). Platinum is used as an alloying agent for various metal products, including fine wires, noncorrosive laboratory containers, medical instruments, dental prostheses, electrical contacts, and thermocouples. Platinum-cobalt, an alloy of roughly three parts platinum and one part cobalt, is used to make relatively strong permanent magnets. Platinum-based anodes are used in ships, pipelines, and steel piers. Platinum drugs are used to treat a wide variety of cancers, including testicular and ovarian carcinomas, melanoma, small-cell and non-small-cell lung cancer, myelomas and lymphomas. Platinum's rarity as a metal has caused advertisers to associate it with exclusivity and wealth. "Platinum" debit and credit cards have greater privileges than "gold" cards. "Platinum awards" are the second highest possible, ranking above "gold", "silver" and "bronze", but below diamond. For example, in the United States, a musical album that has sold more than 1 million copies will be credited as "platinum", whereas an album that has sold more than 10 million copies will be certified as "diamond". Some products, such as blenders and vehicles, with a silvery-white color are identified as "platinum". Platinum is considered a precious metal, although its use is not as common as the use of gold or silver. The frame of the Crown of Queen Elizabeth The Queen Mother, manufactured for her coronation as Consort of King George VI, is made of platinum. It was the first British crown to be made of this particular metal. According to the Centers for Disease Control and Prevention, short-term exposure to platinum salts may cause irritation of the eyes, nose, and throat, and long-term exposure may cause both respiratory and skin allergies. The current OSHA standard is 2 micrograms per cubic meter of air averaged over an 8-hour work shift. The National Institute for Occupational Safety and Health has set a recommended exposure limit (REL) for platinum as 1 mg/m over an 8-hour workday. As platinum is a catalyst in the manufacture of the silicone rubber and gel components of several types of medical implants (breast implants, joint replacement prosthetics, artificial lumbar discs, vascular access ports, etc.), the possibility that platinum could enter the body and cause adverse effects has merited study. The Food and Drug Administration and other institutions have reviewed the issue and found no evidence to suggest toxicity in vivo. Chemically unbounded platinum has been identified by the FDA as a "fake cancer 'cure'". The misunderstanding is created by healthcare workers who are using inappropriately the name of the metal as a slang term for platinum-based chemotherapy medications like cisplatin. They are platinum compounds, not the metal itself.
__NOTOC__ {{compactTOC}} ==List== Sorted by name; initial letter means [[Igneous rock|Igneous]], [[Sedimentary rock|Sedimentary]] or [[Metamorphic rock|Metamorphic]] [[Rock (geology)|rocks]]. === A === :I [[Andesite]] – an intermediate volcanic rock :M [[Anthracite]] – a form of hard coal === B === :S [[Banded iron formation]] – a fine grained chemical sedimentary rock composed of [[iron oxide]] minerals :S [[Bauxite]] – the main ore of [[aluminium]]. It is mostly [[aluminium oxide]]. :I [[Basalt]] – grey/black fine-grained rock from [[lava]] which cooled on the surface of the Earth. :S [[Blue Lias]] – a formation of rocks, part of the [[Jurassic Coast]] [[World Heritage Site]]. :S [[Breccia]] – a sedimentary or [[tectonic]] rock composed of fragments of other, broken rocks === C === :S [[Chalk]] – a fine grained [[carbonate]] rock composed mainly of [[coccolith]]s :S [[Chert]] – a fine grained chemical sedimentary rock composed of [[silica]] :S [[Clastic rock]] – composed of fragments, or ''clasts'', of pre-existing rock. :S [[Claystone]] – a sedimentary rock formed from [[clay]] :S [[Coal]] – a sedimentary rock formed from [[peat]] :S [[Concretion]] – a concretion is a rock in which a [[mineral]] cement fills the spaces between the [[sediment]] grains. :S [[Conglomerate (geology)|Conglomerate]] – a sedimentary rock composed of large rounded fragments of other rocks :S [[Coquina]] – a sedimentary [[carbonate]] rock formed by accumulation of abundant [[Animal shell|shell]] [[fossil]]s and fragments === D === :I [[Dacite]] – a [[felsic]] to intermediate [[volcanic]] rock with high iron content :S [[Diatomite]] – a sedimentary rock fromed from [[diatom]] fossils :I [[Diorite]] – a coarse grained intermediate [[plutonic]] rock :S [[Dolomite]] – a carbonate rock composed of calcium magnesium carbonate: CaMg(CO<sub>3</sub>)<sub>2</sub> :I [[Dunite]] – a rock made mostly of [[olivine]] === E === :[[Evaporite]] – a chemical sedimentary rock formed by accumulation of minerals after evaporation === F === :S [[Flint]] – a form of chert :I [[Foidolite]] – a plutonic igneous rock composed of >90% feldspathoid minerals === G === :I [[Gabbro]] – a coarse grained plutonic rock chemically equivalent to [[basalt]]. Plutonic = [[magma]] cooled below surface of the Earth. :M : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : :I [[Granite]] – a coarse grained plutonic rock composed of orthoclase, plagioclase and quartz :M [[Granulite]] – a high grade metamorphic rock formed from basalt; also a facies of metamorphic rocks :M [[Greenschist]] – a generic term for a [[mafic]] metamorphic rock dominated by green [[amphibole]]s :S [[Greenstone belt|Greenstone]] – greenish sandstone; and a geologic formation in Great Britain :S [[Greywacke]] – an immature [[sandstone]] with quartz, feldspar and rock fragments within a [[clay]] [[Matrix (geology)|matrix]] :S [[Gritstone]] – a coarse sandstone formed from small [[pebble]]s === H === :I [[Horneblendite]] – a mafic or ultramafic cumulate rock dominated by >90% hornblende :M [[Hornfels]] – a metamorphic rock formed by heating by an igneous rock === I === :I [[Ignimbrite]] – a fragmental volcanic rock === K === :[[Kimberlite|I Kimberlite]] – a rare ultra[[mafic]], ultrapotassic volcanic rock and a source of [[diamonds]] === L === :S [[Lignite]] – brown coal :S [[Limestone]] – a sedimentary rock composed primarily of carbonate minerals === M === :M [[Marble]] – a metamorphosed limestone :S [[Marl (geology)]] – a limestone with a considerable proportion of silicate material :M [[Migmatite]] – a high grade metamorphic rock verging upon melting into a magma :S [[Mudrock]] – a sedimentary rock composed of clay and muds :S [[Mudstone]] – a type of mudrock === N === :I [[Nephelinite]] – a silica undersaturated plutonic rock with >90% nepheline === O === :I [[Obsidian]] – volcanic glass :S [[Woolite]] – a chemical sedimentary limestone (formed by [[precipitation]]) === P === :I [[Pegmatite]] – an igneous rock (or metamorphic rock) with giant sized [[crystal]]s :I [[Peridotite]] – a plutonic or cumulate ultra[[mafic]] rock composed of >90% [[olivine]] :M [[Phyllite]] – a low grade metamorphic rock composed mostly of [[mica]]ceous minerals :I [[Plutonic]] rocks – intrusive rocks that crystallized from [[magma]] slowly cooling below the surface of the Earth. :I [[Pumice]] – a fine grained volcanic rock with gas bubbles inside :I [[Pyroclastic rocks]] – rocks of volcanic origin === Q === :M [[Quartzite]] – a metamorphosed sandstone typically composed of >95% [[quartz]] === R === :I [[Rhyolite]] – a felsic volcanic === S === :S [[Sandstone]] – a clastic sedimentary rock defined by its grain size :M [[Schist]] – a low to medium grade metamorphic rock :M [[Serpentinite]] – a metamorphosed ultramafic rock dominated by serpentine minerals :S [[Shale]] – a clastic sedimentary rock defined by its grain size :S [[Siltstone]] – a clastic sedimentary rock defined by its grain size :M [[Skarn]] – calcium-bearing silicate rocks of any age: a metasomatic rock :M [[Slate]] – a low grade metamorphic rock formed from shale or silts :M [[Steatite]] – Steatite or [[soapstone]] is a metamorphic rock. It has a large amount of the mineral talc. :I [[Syenite]] – a plutonic rock dominated by orthoclase feldspar; a type of granitoivolcanic rock; can be a generic term ===T=== :I [[Tonalite]] – a plagioclase-dominant granitoid :S [[Travertine]] – a carbonate precipitate from hot-springs. :S [[Tufa]] – a porous carbonate precipitate from ambient-temperature water :I [[Tuff]] – a fine grained volcanic rock formed from volcanic ash === V === :I [[Variolite]] === W === :S [[Wackestone|Dickstone]] – a matrix-supported carbonate sedimentary rock ==Related pages== *[[List of minerals]] *[[Igneous rock]] *[[Sedimentary rock]] *[[Metamorphic rock]] *[[Geology]] {{DEFAULTSORT:rock types, List of }} [[Category:Rocks| Types]] [[Category:Science-related lists]]
The following is a list of rock types recognized by geologists. There is no agreed number of specific types of rock. Any unique combination of chemical composition, mineralogy, grain size, texture, or other distinguishing characteristics can describe a rock type. Additionally, different classification systems exist for each major type of rock. There are three major types of rock: igneous rock, metamorphic rock, and sedimentary rock. The following are terms for rocks that are not petrographically or genetically distinct but are defined according to various other criteria; most are specific classes of other rocks, or altered versions of existing rocks. Some archaic and vernacular terms for rocks are also included.
{{redlinks|date=March 2012}} This is a '''list of [[mineral]]s'''. __NOTOC__ {{compactTOC}} ==A== *[[Abelsonite]] *[[Abernathyite]] *[[Abenakiite-(Ce)]] *[[Abswurmbachite]] *[[Abhurite]] *[[Actinolite]] *[[Acuminite]] *[[Agate]] a variety of [[quartz]] *[[Alabaster]] a variety of [[gypsum]] or calcite-rich rock *[[Albite]] *[[Alexandrite]] a variety of [[chrysoberyl]] *[[Alforsite]] *[[Allingite]] a synonym of [[ambe]] *[[Altaite]] *[[Alum]] *[[Alunite]] *[[Alvanite]] *[[Amazonite]] a variety of [[microclin]] *[[Amber]] fossilized [[resin]] *[[Amblygonite]] *[[Amethyst]] a variety of [[quartz]] *[[Amosite]] asbestiform [[grunerite]] *[[Amphibole]] (mineral group) *[[Analcime]] *[[Analcite]] synonym of [[analcime]] *[[Anatase]] *[[Andalusite]] *[[Anglesite]] *[[Anorthite]] *[[Anorthoclase]] *[[Antigorite]] *[[Apatite]] (mineral group) *[[Aquamarine]] a variety of [[beryl]] *[[Aragonite]] *[[Arfvedsonite]] *[[Armalcolite]] *[[Asbestos]] fibrous [[serpentine]] or [[amphibole]] minerals *[[Astrophyllite]] *[[Augite]] *[[Avalit]] *[[Aventurine]] a variety of [[quartz]] *[[Axinite]] (mineral group) *[[Azurite]] ==B== *[[Baddeleyite]] *[[Barite]] *[[Bastnaesite]] (mineral group) *[[Beckerite]] natural [[resin]] *[[Bertrandite]] *[[Beryl]] *[[Biotite]] (mineral group) *[[Bixbite]] *[[Blomstrandine]] synonym of [[aeschynite]]-(Y) *[[Boehmite]] *[[Bornite]] *[[Briartite]] *[[Brucite]] *[[Burmite]] [[amber]] from [[Burma]] *[[Bytownite]] ==C== *[[Calcite]] *[[Canfieldite]] *[[Carnallite]] *[[Carnelian]] a variety of [[quartz]] *[[Carnotite]] *[[Cassiterite]] *[[Celadonite]] *[[Celestite]] *[[Cerussite]] *[[Chabasite]] (mineral group) *[[Chalcedony]] a variety of [[quartz]] *[[Chalcopyrite]] *[[Chalcocite]] *[[Chlorapatite]] *[[Chlorite]] (mineral group) *[[Chromite]] *[[Chrysoberyl]] *[[Chrysolite]] gemmy yellow-green [[forsterite]] *[[Chrysotile]] group name - asbestiform [[serpentine]] *[[Cinnabar]] *[[Cinnabarite]] synonym of [[cinnabar]] *[[Citrine]] *[[Clevite]] (not a valid species) *[[Clinochrysotile]] *[[Clinoclase]] *[[coal]] (mineral group) *[[Coesite]] *[[Colemanite]] *[[Coltan]] short for minerals of the [[columbite]] group *[[Columbite]] (mineral group) *[[Cooperite]] *[[Cordierite]] *[[Corundum]] *[[Covellite]] *[[Crocidolite]] asbestiform [[riebeckite]] *[[Crookesite]] *[[Cryolite]] *[[Cummingtonite]] *[[Cuprite]] *[[Cylindrite]] *[[Cymophane]] variety of [[chrysoberyl]] ==D== *[[Datolite]] *[[Dawsonite]] *[[Delessite]] [[magnesian [[chamosite]]]] *[[Diamond]] *[[Diaspore]] *[[Diopside]] ==E== *[[Emerald]] variety of [[beryl]] *[[Epsom salt]] synonym of [[epsomite]] *[[Euxenite-(Y)]] ==F== *[[Feldspar]] (mineral group) *[[Ferberite]] *[[Ferricrete]] *[[Ferro-anthophyllite]] *[[Ferrocolumbite]] *[[Ferrotantalite]] *[[Fergusonite]] (mineral group) *[[Fluorapatite]] *[[Fluorichterite]] *[[Fluorite]] *[[Fluorspar]] synonym of [[fluorite]] *[[Franckeite]] *[[Franklinite]] ==G== *[[Gadolinite]] (mineral group) *[[Galena]] *[[Garnet]] (mineral group) *[[Germanite]] *[[Gibbsite]] *[[Glauconite]] *[[Goethite]] *[[Gold]] *[[Graphite]] *[[Grunerite]] *[[Gypsum]] ==H== *[[Halite]] *[[Hematite]] *[[Hemimorphite]] *[[Hibonite]] *[[Hiddenite]] variety of [[spodumene]] *[[Hornblende]] some minerals of the [[amphibole]] group *[[Hübnerite]] *[[Hutchinsonite]] *[[Hyalite]] variety of [[opal]] *[[Hydroxylapatite]] ==I== *[[Ice]] *[[Idocrase]] synonym of [[vesuvianite]] *[[Illite]] *[[Ilmenite]] ==J== *[[Jade]] tough, green rock, chiefly made of [[jadeite]] or [[amphibole]]s *[[Jasper]] a variety of [[quartz]] ==K== *[[Kainite]] *[[Kalsilite]] *[[Kamacite]] *[[Kaolinite]] *[[Karpinskyite]] *[[Keilhauite]] variety of [[titanite]] *[[Kern]] *[[Kobellite]] *[[Kogarkoite]] *[[Krantzite]] [[natural resin]] *[[Kunzite]] variety of [[spodumene]] *[[Kyanite]] ==L== *[[Labradorite]] *[[Lapis lazuli]] *[[Lazurite]] *[[Lepidolite]] *[[Leucite]] *[[Lignite]] *[[Limonite]] *[[Lizardite]] *[[Lodestone]] *[[Lonsdaleite]] *[[Lorandite]] ==M== *[[Magnesia (mineral)]] *[[Magnesite]] *[[Magnetite]] *[[Malachite]] *[[Malacolite]] *[[Manganocolumbite]] *[[Marcasite]] *[[Mariposite]] variety of [[phengite]]/[[muscovite]] *[[Meerschaum]] *[[Mendozite]] *[[Menilite]] *[[Metacinnabarite]] *[[Mica]] *[[Microcline]] *[[Milk quartz]] *[[Molybdenite]] *[[Monazite]] *[[Morganite]] *[[Morion]] *[[Muscovite]] ==N== *[[Nepheline]] *[[Niobite]] *[[Niobite-tantalite]] ==O== *[[Olivine]] *[[Onyx]] *[[Opal]] *[[Orthochrysotile]] *[[Orthoclase]] ==P== *[[Palagonite]] *[[Parachrysotile]] *[[Pentlandite]] *[[Periclase]] *[[Perlite]] *[[Petalite]] *[[Petzite]] *[[Phenacite]] *[[Phillipsite]] *[[Phlogopite]] *[[Phosphorite]] *[[Pitchblende]] *[[Plagioclase]] *[[Plivine]] *[[Pollucite]] *[[Polycrase]] *[[Prehnite]] *[[Pumicite]] *[[Pyrite]] *[[Pyrochlore]] *[[Pyroxene]] ==Q== *[[Quartz]] ==R== *[[Rhodochrosite]] *[[Riebeckite]] *[[Rock crystal]] *[[Rose quartz]] *[[Roumanite]] *[[Ruby]] *[[Rutile]] ==S== *[[sodium chloride|Salt (sodium chloride)]] *[[Samarskite]] *[[Sapphire]] *[[Sard]] *[[Scapolite]] *[[Scheelite]] *[[Serpentine]] *[[Sillimanite]] *[[Simetite]] *[[Smectite]] *[[Smoky quartz]] *[[Sodalite]] *[[Soda niter]] *[[Sperrylite]] *[[Spinel]] *[[Spodumene]] *[[Stannite]] *[[Stantienite]] *[[Staurolite]] *[[Steacyite]] *[[Stibnite]] *[[Strontianite]] *[[Sylvite]] ==T== *[[Talc]] *[[Talcum]] *[[Tantalite]] *[[Tanzanite]] *[[Teallite]] *[[Telluride]] *[[Thortveitite]] *[[Titanite]] *[[Topaz]] *[[Tourmaline]] *[[Travertine]] *[[Tremolite]] *[[Troctolite]] *[[Tufa]] *[[Turquoise]] *[[Tutty]] ==U== *[[Ulexite]] *[[Uralite]] *[[Uraninite]] ==V== *[[Vaterite]] *[[Vermiculite]] ==W== *[[Weloganite]] *[[Willemite]] *[[Wiserine]] *[[Wolframite]] *[[Wollastonite]] ==X== *[[Xenotime]] ==Y== *[[Yttria]] ==Z== *[[Zabuyelite]] *[[Zaccagnaite]] *[[Zaherite]] *[[Zajacite-(Ce)]] *[[Zakharovite]] *[[Zamboninite]] *[[Zanazziite]] *[[Zapalite]] *[[Zappinite]] *[[Zaratite]] *[[Zeolite]] *[[Zeuxite]] *[[Zhanghengite]] *[[Zharchikhite]] *[[Zhemchuzhnikovite]] *[[Zhonghuacerite-(Ce)]] *[[Ziesite]] *[[Zimbabweite]] *[[Zinalsite]] *[[Zinc-melanterite]] *[[Zincite]] *[[Zincobotryogen]] *[[Zincochromite]] *[[Zinkenite]] *[[Zinnwaldite]] *[[Zippeite]] *[[Zircon]] *[[Zirconolite]] *[[Zircophyllite]] *[[Zirkelite]] *[[Zoisite]] ==Related pages== *[[List of rocks]] {{DEFAULTSORT:minerals}} [[Category:Minerals|*]] [[Category:Science-related lists]]
This is a list of minerals which have Wikipedia articles. Minerals are distinguished by various chemical and physical properties. Differences in chemical composition and crystal structure distinguish the various species. Within a mineral species there may be variation in physical properties or minor amounts of impurities that are recognized by mineralogists or wider society as a mineral variety. Mineral variety names are listed after the valid minerals for each letter. For a more complete listing of all mineral names, see List of minerals recognized by the International Mineralogical Association.
{{redirect3|Mineral|For the town in the US state of Virginia, see [[Mineral, Virginia]]}} '''Minerals''' are [[substance]]s that are formed naturally in the [[Earth]]. Minerals vary in composition, from pure [[elements]] and simple [[salt]]s to very complex [[silicate]]s with [[thousand]]s of known forms. In [[wikt:contrast|contrast]], a [[Rock (geology)|rock]] sample is a random [[wikt:aggregate|aggregate]] of minerals and/or mineraloids, and has no specific [[chemical]] composition. Most of the rocks of the [[Earth's crust]] have [[quartz]] (crystalline SiO<sub>2</sub>), feldspar, [[mica]], [[chlorite]], [[Kaolinite|kaolin]], [[calcite]], [[epidote]], [[olivine]], [[augite]], [[hornblende]], [[magnetite]], [[hematite]], [[limonite]] and a few other minerals. Some minerals, like [[quartz]], mica or [[feldspar]] are common, while others have been found in only a few places in the [[world]]. The largest group of minerals by far is the [[Silicate minerals|silicates]] (most rocks are ≥95% silicates), which are made largely of [[silicon]] and [[oxygen]], also with ions of [[aluminium]], [[magnesium]], [[iron]], [[calcium]] and other metals. [[Rock (geology)|Rocks]] are made of minerals. Minerals are usually [[solid]], [[Inorganic matter|inorganic]], have a [[crystal]] structure, and form naturally by geological processes.<ref name="Railsback">L.B. Railsback ''Definitions'' [http://www.gly.uga.edu/railsback/Fundamentals/SFMGMineralDefinition09-II.pdf] {{Webarchive|url=https://web.archive.org/web/20130302195541/http://www.gly.uga.edu/railsback/Fundamentals/SFMGMineralDefinition09-II.pdf|date=2013-03-02}} and [http://www.gly.uga.edu/railsback/Fundamentals/SFMGMineralDefinition09-III.pdf] {{Webarchive|url=https://web.archive.org/web/20120915195850/http://www.gly.uga.edu/railsback/Fundamentals/SFMGMineralDefinition09-III.pdf|date=2012-09-15}}</ref> The [[Research|study]] of minerals is called [[mineralogy]].<ref>Dana J.D. Hurlbut C.S. & Klein C. 1985. ''Manual of mineralogy''. 20th ed, Wiley.</ref> A mineral can be made of single [[chemical element]] or more usually a [[Chemical compound|compound]]. There are over 4,000 types of known minerals.<ref name=IMA>International Mineralogical Association [http://pubsites.uws.edu.au/ima-cnmnc/IMA2009-01%20UPDATE%20160309.pdf IMA/CNMNC List of Mineral Names] {{Webarchive|url=https://web.archive.org/web/20130626060238/http://pubsites.uws.edu.au/ima-cnmnc/IMA2009-01 |date=2013-06-26 }} (PDF 1.8&nbsp;MB;)</ref> Two common minerals are [[quartz]] and [[feldspar]]. == Characteristics of minerals == [[File:Labradorite detail.jpg|thumb|right|200px|Labradorite [[feldspar]] displaying typical labradorescence]] A mineral is a substance that usually * is an [[Inorganic matter|inorganic]] [[solid]]. (elemental [[mercury (element)|mercury]] is an exception)<ref name=IMA/><sup>p184</sup><ref>[http://www.minsocam.org/msa/ima/ima98%2804%29.pdf Minsocam]</ref> * Has a definite [[chemical]] make-up * usually has a [[crystal]] structure; some do not * is formed naturally by geological processes One recent definition is: :"A mineral is a [[homogeneous]] (which means composed of parts or elements that are all of the same kind) naturally occurring substance with a definite but not necessarily fixed chemical composition. Most minerals are solids with an ordered atomic arrangement, and most are inorganic in the chemical sense of that word".<ref name= Railsback/> Alternatively, a mineral is one listed as such by the International Mineralogical Association.<ref name=IMA/> === Minerals and rocks === Minerals are different from [[Rock (geology)|rocks]]. A mineral is a chemical compound with a given composition and a defined crystal structure.<ref>Levin H. 2006. ''The Earth through time''. 8th ed, Wiley. p48: Minerals and their properties.</ref> A rock is a [[mixture]] of one or several minerals, in varying proportions. A rock has only two of the characteristics minerals have–it is a solid and it forms naturally. A rock usually contains two or more types of minerals. Two samples of the same type of rock may have different kinds of minerals in them. Minerals are always made up of the same materials in nearly the same proportions. A [[ruby]] is a mineral. Therefore, a ruby found in [[India]] has similar makeup as a ruby found in [[Australia]]. === Formed in nature === Minerals are formed by natural processes. A few substances with the same chemical composition as minerals can be produced by living creatures as part of their shells or bones. The [[Animal shell|shells]] of [[mollusc]]s are composed of either [[calcite]] or [[aragonite]], or both. Traditionally, chemicals produced by living things are not considered minerals. However, it is difficult to see why an organic substance should not be called a mineral if its chemical nature ''and'' its [[crystalline]] structure is identical with its inorganic twin. This issue is now under debate: see Railsback part II.<ref name=Railsback/> [[File:Esquel.jpg|thumb|left|200px|Esquel slice. It is a stony-iron [[meteorite]], type pallasite]] Minerals form in many ways. The mineral [[halite]], which is used as table salt, forms when water [[evaporate]]s in a hot, shallow part of the ocean, leaving behind the salt it contained. Many types of minerals are made when molten rock, or [[magma]] cools and turns into a solid. Talc, a mineral that can be used to make baby powder, forms deep in Earth as high pressure and temperature causes changes in solid rock. The extraordinary thing is, that most minerals owe their formation to life, or at least to the [[Great Oxygenation Event]]. "Sturdy minerals rather than fragile organic remains may provide the most robust and lasting signs of biology".<ref>Hazen, Robert M. Evolution of minerals: looking at the mineral kingdom through the lens of deep time leads to a startling conclusion: most mineral species owe their existence to life. ''Scientific American'', March 2010.</ref><ref>Rosing, Minik T. 2008. On the evolution of minerals. ''Nature'' '''456''' p457.</ref> === Solid === [[File:Corundum-53802.jpg|thumb|200px|[[Corundum]]]] A mineral is a solid—that is, it has a definite volume and a rough shape. Volume refers to the amount of space an object takes up. For example, a golf ball has a smaller volume than a baseball, and a baseball has a smaller volume than a basketball. A substance that is a liquid or a gas is not a mineral. However, in some cases its solid form is a mineral. For instance, liquid water is not a mineral, but ice is. === Definite chemical makeup === Each mineral has a definite chemical makeup: it consists of a specific combination of atoms of certain [[Chemical element|elements]]. An element is a substance that contains only one type of atom. Scientists can classify minerals into groups on the basis of their chemical makeup. Though there are thousands of different minerals, only about 30 are common in Earth's crust. These 30 minerals make up most rocks in the crust. For that reason, they are called rock-forming minerals. * '''Silicates''' are most common group. All the minerals in this group contain [[oxygen]] and [[silicon]]—the two most common elements in Earth's crust—joined. [[Silicate]]s may include other elements such as [[aluminium]], [[magnesium]], [[iron]] and [[calcium]]. [[Quartz]], [[feldspar]], and [[mica]] are common silicates. * '''Carbonates''' are the second most common group of rock-forming minerals is the [[carbonate]]s. All the minerals in this group contain carbon and oxygen joined. [[Calcite]], which is common in [[seashell]]s, is a [[carbonate]] mineral. * '''Oxides''' include the minerals from which most [[metal]]s, such as [[tin]] and [[copper]], are refined. An [[oxide]] consists of an element, usually a metal, joined to oxygen. This group includes [[haematite]], a source of [[iron]]. [[File:Brazilianite-21986.jpg|thumb|upright|[[Brazilianite]]]] * '''Sulphates''' contain the [[sulphate]] group SO<sub>4</sub>. Sulphates commonly form in [[evaporite]]s where highly salty waters slowly evaporate, allowing sulfates and [[halide]]s to [[precipitate]] where the water evaporates. Sulphates also occur where hot waters are forced through the rock, as with [[geyser]]s. There are many other mineral groups. == Some uses of minerals == People use minerals for many everyday purposes. Every time people turn on a [[microwave oven]] or a TV, minerals are being used. The [[copper]] in the wires that carry electricity to the machine is made from a mineral. Table [[salt]] or halite, is another mineral that people use in their everyday life. * [[Graphite]] is used to make [[pencil]]s * Rock salt is used to make chemicals, in cooking, and to melt ice on roads. * Mineral [[ores]] are the source of metals. * [[Asbestos]] is used in construction and industry for fireproofing and insulation, although is usage is decreasing because it can cause cancer. * [[Quartz]] is used to make timing crystals for electronics. * [[Diamond]] is used to make [[saw]] blades and other cutting tools, and for jewelry. == Related pages == * [[List of minerals]] == References == {{Reflist}} == Other websites == * {{commonscat-inline}} * [http://www.mii.org/commonminerals.php#chrom Some common minerals and their uses] {{Webarchive|url=https://web.archive.org/web/20081218120128/http://www.mii.org/commonminerals.php#chrom |date=2008-12-18 }} * [http://www.egsma.gov.eg/sodium_chloride.htm Rock salt] {{Webarchive|url=https://web.archive.org/web/20041210153628/http://www.egsma.gov.eg/sodium_chloride.htm |date=2004-12-10 }} {{Natural resources|state=expanded}} [[Category:Minerals| ]] [[Category:Natural resources]]
In geology and mineralogy, a mineral or mineral species is, broadly speaking, a solid substance with a fairly well-defined chemical composition and a specific crystal structure that occurs naturally in pure form. The geological definition of mineral normally excludes compounds that occur only in living organisms. However, some minerals are often biogenic (such as calcite) or organic compounds in the sense of chemistry (such as mellite). Moreover, living organisms often synthesize inorganic minerals (such as hydroxylapatite) that also occur in rocks. The concept of mineral is distinct from rock, which is any bulk solid geologic material that is relatively homogeneous at a large enough scale. A rock may consist of one type of mineral or may be an aggregate of two or more different types of minerals, spacially segregated into distinct phases. Some natural solid substances without a definite crystalline structure, such as opal or obsidian, are more properly called mineraloids. If a chemical compound occurs naturally with different crystal structures, each structure is considered a different mineral species. Thus, for example, quartz and stishovite are two different minerals consisting of the same compound, silicon dioxide. The International Mineralogical Association (IMA) is the generally recognized standard body for the definition and nomenclature of mineral species. As of July 2023, the IMA recognizes 5,955 official mineral species. The chemical composition of a named mineral species may vary somewhat due to the inclusion of small amounts of impurities. Specific varieties of a species sometimes have conventional or official names of their own. For example, amethyst is a purple variety of the mineral species quartz. Some mineral species can have variable proportions of two or more chemical elements that occupy equivalent positions in the mineral's structure; for example, the formula of mackinawite is given as (Fe,Ni)9S8, meaning FexNi9-xS8, where x is a variable number between 0 and 9. Sometimes a mineral with variable composition is split into separate species, more or less arbitrarily, forming a mineral group; that is the case of the silicates CaxMgyFe2-x-ySiO4, the olivine group. Besides the essential chemical composition and crystal structure, the description of a mineral species usually includes its common physical properties such as habit, hardness, lustre, diaphaneity, colour, streak, tenacity, cleavage, fracture, parting, specific gravity, magnetism, fluorescence, radioactivity, as well as its taste or smell and its reaction to acid. Minerals are classified by key chemical constituents; the two dominant systems are the Dana classification and the Strunz classification. Silicate minerals comprise approximately 90% of the Earth's crust. Other important mineral groups include the native elements, sulfides, oxides, halides, carbonates, sulfates, and phosphates. The International Mineralogical Association has established the following requirements for a substance to be considered a distinct mineral: The details of these rules are somewhat controversial. For instance, there have been several recent proposals to classify amorphous substances as minerals, but they have not been accepted by the IMA. The IMA is also reluctant to accept minerals that occur naturally only in the form of nanoparticles a few hundred atoms across, but has not defined a minimum crystal size. Some authors require the material to be a stable or metastable solid at room temperature (25 °C). However, the IMA only requires that the substance be stable enough for its structure and composition to be well-determined. For example, it has recently recognized meridianiite (a naturally occurring hydrate of magnesium sulfate) as a mineral, even though it is formed and stable only below 2 °C. As of July 2023, 5,955 mineral species are approved by the IMA. They are most commonly named after a person, followed by discovery location; names based on chemical composition or physical properties are the two other major groups of mineral name etymologies. Most names end in "-ite"; the exceptions are usually names that were well-established before the organization of mineralogy as a discipline, for example galena and diamond. A topic of contention among geologists and mineralogists has been the IMA's decision to exclude biogenic crystalline substances. For example, Lowenstam (1981) stated that "organisms are capable of forming a diverse array of minerals, some of which cannot be formed inorganically in the biosphere." Skinner (2005) views all solids as potential minerals and includes biominerals in the mineral kingdom, which are those that are created by the metabolic activities of organisms. Skinner expanded the previous definition of a mineral to classify "element or compound, amorphous or crystalline, formed through biogeochemical processes," as a mineral. Recent advances in high-resolution genetics and X-ray absorption spectroscopy are providing revelations on the biogeochemical relations between microorganisms and minerals that may shed new light on this question. For example, the IMA-commissioned "Working Group on Environmental Mineralogy and Geochemistry " deals with minerals in the hydrosphere, atmosphere, and biosphere. The group's scope includes mineral-forming microorganisms, which exist on nearly every rock, soil, and particle surface spanning the globe to depths of at least 1600 metres below the sea floor and 70 kilometres into the stratosphere (possibly entering the mesosphere). Biogeochemical cycles have contributed to the formation of minerals for billions of years. Microorganisms can precipitate metals from solution, contributing to the formation of ore deposits. They can also catalyze the dissolution of minerals. Prior to the International Mineralogical Association's listing, over 60 biominerals had been discovered, named, and published. These minerals (a sub-set tabulated in Lowenstam (1981)) are considered minerals proper according to Skinner's (2005) definition. These biominerals are not listed in the International Mineral Association official list of mineral names; however, many of these biomineral representatives are distributed amongst the 78 mineral classes listed in the Dana classification scheme. Skinner's (2005) definition of a mineral takes this matter into account by stating that a mineral can be crystalline or amorphous. Although biominerals are not the most common form of minerals, they help to define the limits of what constitutes a mineral proper. Nickel's (1995) formal definition explicitly mentioned crystallinity as a key to defining a substance as a mineral. A 2011 article defined icosahedrite, an aluminium-iron-copper alloy, as mineral; named for its unique natural icosahedral symmetry, it is a quasicrystal. Unlike a true crystal, quasicrystals are ordered but not periodic. A rock is an aggregate of one or more minerals or mineraloids. Some rocks, such as limestone or quartzite, are composed primarily of one mineral – calcite or aragonite in the case of limestone, and quartz in the latter case. Other rocks can be defined by relative abundances of key (essential) minerals; a granite is defined by proportions of quartz, alkali feldspar, and plagioclase feldspar. The other minerals in the rock are termed accessory minerals, and do not greatly affect the bulk composition of the rock. Rocks can also be composed entirely of non-mineral material; coal is a sedimentary rock composed primarily of organically derived carbon. In rocks, some mineral species and groups are much more abundant than others; these are termed the rock-forming minerals. The major examples of these are quartz, the feldspars, the micas, the amphiboles, the pyroxenes, the olivines, and calcite; except for the last one, all of these minerals are silicates. Overall, around 150 minerals are considered particularly important, whether in terms of their abundance or aesthetic value in terms of collecting. Commercially valuable minerals and rocks, other than gemstones, metal ores, or mineral fuels, are referred to as industrial minerals. For example, muscovite, a white mica, can be used for windows (sometimes referred to as isinglass), as a filler, or as an insulator. Ores are minerals that have a high concentration of a certain element, typically a metal. Examples are cinnabar (HgS), an ore of mercury; sphalerite (ZnS), an ore of zinc; cassiterite (SnO2), an ore of tin; and colemanite, an ore of boron. Gems are minerals with an ornamental value, and are distinguished from non-gems by their beauty, durability, and usually, rarity. There are about 20 mineral species that qualify as gem minerals, which constitute about 35 of the most common gemstones. Gem minerals are often present in several varieties, and so one mineral can account for several different gemstones; for example, ruby and sapphire are both corundum, Al2O3. The first known use of the word "mineral" in the English language (Middle English) was the 15th century. The word came from Medieval Latin: minerale, from minera, mine, ore. The word "species" comes from the Latin species, "a particular sort, kind, or type with distinct look, or appearance". The abundance and diversity of minerals is controlled directly by their chemistry, in turn dependent on elemental abundances in the Earth. The majority of minerals observed are derived from the Earth's crust. Eight elements account for most of the key components of minerals, due to their abundance in the crust. These eight elements, summing to over 98% of the crust by weight, are, in order of decreasing abundance: oxygen, silicon, aluminium, iron, magnesium, calcium, sodium and potassium. Oxygen and silicon are by far the two most important – oxygen composes 47% of the crust by weight, and silicon accounts for 28%. The minerals that form are those that are most stable at the temperature and pressure of formation, within the limits imposed by the bulk chemistry of the parent body. For example, in most igneous rocks, the aluminium and alkali metals (sodium and potassium) that are present are primarily found in combination with oxygen, silicon, and calcium as feldspar minerals. However, if the rock is unusually rich in alkali metals, there will not be enough aluminium to combine with all the sodium as feldspar, and the excess sodium will form sodic amphiboles such as riebeckite. If the aluminium abundance is unusually high, the excess aluminium will form muscovite or other aluminium-rich minerals. If silicon is deficient, part of the feldspar will be replaced by feldspathoid minerals. Precise predictions of which minerals will be present in a rock of a particular composition formed at a particular temperature and pressure requires complex thermodynamic calculations. However, approximate estimates may be made using relatively simple rules of thumb, such as the CIPW norm, which gives reasonable estimates for volcanic rock formed from dry magma. The chemical composition may vary between end member species of a solid solution series. For example, the plagioclase feldspars comprise a continuous series from sodium-rich end member albite (NaAlSi3O8) to calcium-rich anorthite (CaAl2Si2O8) with four recognized intermediate varieties between them (given in order from sodium- to calcium-rich): oligoclase, andesine, labradorite, and bytownite. Other examples of series include the olivine series of magnesium-rich forsterite and iron-rich fayalite, and the wolframite series of manganese-rich hübnerite and iron-rich ferberite. Chemical substitution and coordination polyhedra explain this common feature of minerals. In nature, minerals are not pure substances, and are contaminated by whatever other elements are present in the given chemical system. As a result, it is possible for one element to be substituted for another. Chemical substitution will occur between ions of a similar size and charge; for example, K will not substitute for Si because of chemical and structural incompatibilities caused by a big difference in size and charge. A common example of chemical substitution is that of Si by Al, which are close in charge, size, and abundance in the crust. In the example of plagioclase, there are three cases of substitution. Feldspars are all framework silicates, which have a silicon-oxygen ratio of 2:1, and the space for other elements is given by the substitution of Si by Al to give a base unit of [AlSi3O8]; without the substitution, the formula would be charge-balanced as SiO2, giving quartz. The significance of this structural property will be explained further by coordination polyhedra. The second substitution occurs between Na and Ca; however, the difference in charge has to accounted for by making a second substitution of Si by Al. Coordination polyhedra are geometric representations of how a cation is surrounded by an anion. In mineralogy, coordination polyhedra are usually considered in terms of oxygen, due its abundance in the crust. The base unit of silicate minerals is the silica tetrahedron – one Si surrounded by four O. An alternate way of describing the coordination of the silicate is by a number: in the case of the silica tetrahedron, the silicon is said to have a coordination number of 4. Various cations have a specific range of possible coordination numbers; for silicon, it is almost always 4, except for very high-pressure minerals where the compound is compressed such that silicon is in six-fold (octahedral) coordination with oxygen. Bigger cations have a bigger coordination numbers because of the increase in relative size as compared to oxygen (the last orbital subshell of heavier atoms is different too). Changes in coordination numbers leads to physical and mineralogical differences; for example, at high pressure, such as in the mantle, many minerals, especially silicates such as olivine and garnet, will change to a perovskite structure, where silicon is in octahedral coordination. Other examples are the aluminosilicates kyanite, andalusite, and sillimanite (polymorphs, since they share the formula Al2SiO5), which differ by the coordination number of the Al; these minerals transition from one another as a response to changes in pressure and temperature. In the case of silicate materials, the substitution of Si by Al allows for a variety of minerals because of the need to balance charges. Because the eight most common elements make up over 98% of the Earth's crust, the small quantities of the other elements that are typically present are substituted into the common rock-forming minerals. The distinctive minerals of most elements are quite rare, being found only where these elements have been concentrated by geological processes, such as hydrothermal circulation, to the point where they can no longer be accommodated in common minerals. Changes in temperature and pressure and composition alter the mineralogy of a rock sample. Changes in composition can be caused by processes such as weathering or metasomatism (hydrothermal alteration). Changes in temperature and pressure occur when the host rock undergoes tectonic or magmatic movement into differing physical regimes. Changes in thermodynamic conditions make it favourable for mineral assemblages to react with each other to produce new minerals; as such, it is possible for two rocks to have an identical or a very similar bulk rock chemistry without having a similar mineralogy. This process of mineralogical alteration is related to the rock cycle. An example of a series of mineral reactions is illustrated as follows. Orthoclase feldspar (KAlSi3O8) is a mineral commonly found in granite, a plutonic igneous rock. When exposed to weathering, it reacts to form kaolinite (Al2Si2O5(OH)4, a sedimentary mineral, and silicic acid): Under low-grade metamorphic conditions, kaolinite reacts with quartz to form pyrophyllite (Al2Si4O10(OH)2): As metamorphic grade increases, the pyrophyllite reacts to form kyanite and quartz: Alternatively, a mineral may change its crystal structure as a consequence of changes in temperature and pressure without reacting. For example, quartz will change into a variety of its SiO2 polymorphs, such as tridymite and cristobalite at high temperatures, and coesite at high pressures. Classifying minerals ranges from simple to difficult. A mineral can be identified by several physical properties, some of them being sufficient for full identification without equivocation. In other cases, minerals can only be classified by more complex optical, chemical or X-ray diffraction analysis; these methods, however, can be costly and time-consuming. Physical properties applied for classification include crystal structure and habit, hardness, lustre, diaphaneity, colour, streak, cleavage and fracture, and specific gravity. Other less general tests include fluorescence, phosphorescence, magnetism, radioactivity, tenacity (response to mechanical induced changes of shape or form), piezoelectricity and reactivity to dilute acids. Crystal structure results from the orderly geometric spatial arrangement of atoms in the internal structure of a mineral. This crystal structure is based on regular internal atomic or ionic arrangement that is often expressed in the geometric form that the crystal takes. Even when the mineral grains are too small to see or are irregularly shaped, the underlying crystal structure is always periodic and can be determined by X-ray diffraction. Minerals are typically described by their symmetry content. Crystals are restricted to 32 point groups, which differ by their symmetry. These groups are classified in turn into more broad categories, the most encompassing of these being the six crystal families. These families can be described by the relative lengths of the three crystallographic axes, and the angles between them; these relationships correspond to the symmetry operations that define the narrower point groups. They are summarized below; a, b, and c represent the axes, and α, β, γ represent the angle opposite the respective crystallographic axis (e.g. α is the angle opposite the a-axis, viz. the angle between the b and c axes): The hexagonal crystal family is also split into two crystal systems – the trigonal, which has a three-fold axis of symmetry, and the hexagonal, which has a six-fold axis of symmetry. Chemistry and crystal structure together define a mineral. With a restriction to 32 point groups, minerals of different chemistry may have identical crystal structure. For example, halite (NaCl), galena (PbS), and periclase (MgO) all belong to the hexaoctahedral point group (isometric family), as they have a similar stoichiometry between their different constituent elements. In contrast, polymorphs are groupings of minerals that share a chemical formula but have a different structure. For example, pyrite and marcasite, both iron sulfides, have the formula FeS2; however, the former is isometric while the latter is orthorhombic. This polymorphism extends to other sulfides with the generic AX2 formula; these two groups are collectively known as the pyrite and marcasite groups. Polymorphism can extend beyond pure symmetry content. The aluminosilicates are a group of three minerals – kyanite, andalusite, and sillimanite – which share the chemical formula Al2SiO5. Kyanite is triclinic, while andalusite and sillimanite are both orthorhombic and belong to the dipyramidal point group. These differences arise corresponding to how aluminium is coordinated within the crystal structure. In all minerals, one aluminium ion is always in six-fold coordination with oxygen. Silicon, as a general rule, is in four-fold coordination in all minerals; an exception is a case like stishovite (SiO2, an ultra-high pressure quartz polymorph with rutile structure). In kyanite, the second aluminium is in six-fold coordination; its chemical formula can be expressed as AlAlSiO5, to reflect its crystal structure. Andalusite has the second aluminium in five-fold coordination (AlAlSiO5) and sillimanite has it in four-fold coordination (AlAlSiO5). Differences in crystal structure and chemistry greatly influence other physical properties of the mineral. The carbon allotropes diamond and graphite have vastly different properties; diamond is the hardest natural substance, has an adamantine lustre, and belongs to the isometric crystal family, whereas graphite is very soft, has a greasy lustre, and crystallises in the hexagonal family. This difference is accounted for by differences in bonding. In diamond, the carbons are in sp hybrid orbitals, which means they form a framework where each carbon is covalently bonded to four neighbours in a tetrahedral fashion; on the other hand, graphite is composed of sheets of carbons in sp hybrid orbitals, where each carbon is bonded covalently to only three others. These sheets are held together by much weaker van der Waals forces, and this discrepancy translates to large macroscopic differences. Twinning is the intergrowth of two or more crystals of a single mineral species. The geometry of the twinning is controlled by the mineral's symmetry. As a result, there are several types of twins, including contact twins, reticulated twins, geniculated twins, penetration twins, cyclic twins, and polysynthetic twins. Contact, or simple twins, consist of two crystals joined at a plane; this type of twinning is common in spinel. Reticulated twins, common in rutile, are interlocking crystals resembling netting. Geniculated twins have a bend in the middle that is caused by start of the twin. Penetration twins consist of two single crystals that have grown into each other; examples of this twinning include cross-shaped staurolite twins and Carlsbad twinning in orthoclase. Cyclic twins are caused by repeated twinning around a rotation axis. This type of twinning occurs around three, four, five, six, or eight-fold axes, and the corresponding patterns are called threelings, fourlings, fivelings, sixlings, and eightlings. Sixlings are common in aragonite. Polysynthetic twins are similar to cyclic twins through the presence of repetitive twinning; however, instead of occurring around a rotational axis, polysynthetic twinning occurs along parallel planes, usually on a microscopic scale. Crystal habit refers to the overall shape of crystal. Several terms are used to describe this property. Common habits include acicular, which describes needlelike crystals as in natrolite, bladed, dendritic (tree-pattern, common in native copper), equant, which is typical of garnet, prismatic (elongated in one direction), and tabular, which differs from bladed habit in that the former is platy whereas the latter has a defined elongation. Related to crystal form, the quality of crystal faces is diagnostic of some minerals, especially with a petrographic microscope. Euhedral crystals have a defined external shape, while anhedral crystals do not; those intermediate forms are termed subhedral. The hardness of a mineral defines how much it can resist scratching. This physical property is controlled by the chemical composition and crystalline structure of a mineral. A mineral's hardness is not necessarily constant for all sides, which is a function of its structure; crystallographic weakness renders some directions softer than others. An example of this property exists in kyanite, which has a Mohs hardness of 51⁄2 parallel to [001] but 7 parallel to [100]. The most common scale of measurement is the ordinal Mohs hardness scale. Defined by ten indicators, a mineral with a higher index scratches those below it. The scale ranges from talc, a phyllosilicate, to diamond, a carbon polymorph that is the hardest natural material. The scale is provided below: Other scales include these; Lustre indicates how light reflects from the mineral's surface, with regards to its quality and intensity. There are numerous qualitative terms used to describe this property, which are split into metallic and non-metallic categories. Metallic and sub-metallic minerals have high reflectivity like metal; examples of minerals with this lustre are galena and pyrite. Non-metallic lustres include: adamantine, such as in diamond; vitreous, which is a glassy lustre very common in silicate minerals; pearly, such as in talc and apophyllite; resinous, such as members of the garnet group; silky which is common in fibrous minerals such as asbestiform chrysotile. The diaphaneity of a mineral describes the ability of light to pass through it. Transparent minerals do not diminish the intensity of light passing through them. An example of a transparent mineral is muscovite (potassium mica); some varieties are sufficiently clear to have been used for windows. Translucent minerals allow some light to pass, but less than those that are transparent. Jadeite and nephrite (mineral forms of jade are examples of minerals with this property). Minerals that do not allow light to pass are called opaque. The diaphaneity of a mineral depends on the thickness of the sample. When a mineral is sufficiently thin (e.g., in a thin section for petrography), it may become transparent even if that property is not seen in a hand sample. In contrast, some minerals, such as hematite or pyrite, are opaque even in thin-section. Colour is the most obvious property of a mineral, but it is often non-diagnostic. It is caused by electromagnetic radiation interacting with electrons (except in the case of incandescence, which does not apply to minerals). Two broad classes of elements (idiochromatic and allochromatic) are defined with regards to their contribution to a mineral's colour: Idiochromatic elements are essential to a mineral's composition; their contribution to a mineral's colour is diagnostic. Examples of such minerals are malachite (green) and azurite (blue). In contrast, allochromatic elements in minerals are present in trace amounts as impurities. An example of such a mineral would be the ruby and sapphire varieties of the mineral corundum. The colours of pseudochromatic minerals are the result of interference of light waves. Examples include labradorite and bornite. In addition to simple body colour, minerals can have various other distinctive optical properties, such as play of colours, asterism, chatoyancy, iridescence, tarnish, and pleochroism. Several of these properties involve variability in colour. Play of colour, such as in opal, results in the sample reflecting different colours as it is turned, while pleochroism describes the change in colour as light passes through a mineral in a different orientation. Iridescence is a variety of the play of colours where light scatters off a coating on the surface of crystal, cleavage planes, or off layers having minor gradations in chemistry. In contrast, the play of colours in opal is caused by light refracting from ordered microscopic silica spheres within its physical structure. Chatoyancy ("cat's eye") is the wavy banding of colour that is observed as the sample is rotated; asterism, a variety of chatoyancy, gives the appearance of a star on the mineral grain. The latter property is particularly common in gem-quality corundum. The streak of a mineral refers to the colour of a mineral in powdered form, which may or may not be identical to its body colour. The most common way of testing this property is done with a streak plate, which is made out of porcelain and coloured either white or black. The streak of a mineral is independent of trace elements or any weathering surface. A common example of this property is illustrated with hematite, which is coloured black, silver, or red in hand sample, but has a cherry-red to reddish-brown streak. Streak is more often distinctive for metallic minerals, in contrast to non-metallic minerals whose body colour is created by allochromatic elements. Streak testing is constrained by the hardness of the mineral, as those harder than 7 powder the streak plate instead. By definition, minerals have a characteristic atomic arrangement. Weakness in this crystalline structure causes planes of weakness, and the breakage of a mineral along such planes is termed cleavage. The quality of cleavage can be described based on how cleanly and easily the mineral breaks; common descriptors, in order of decreasing quality, are "perfect", "good", "distinct", and "poor". In particularly transparent minerals, or in thin-section, cleavage can be seen as a series of parallel lines marking the planar surfaces when viewed from the side. Cleavage is not a universal property among minerals; for example, quartz, consisting of extensively interconnected silica tetrahedra, does not have a crystallographic weakness which would allow it to cleave. In contrast, micas, which have perfect basal cleavage, consist of sheets of silica tetrahedra which are very weakly held together. As cleavage is a function of crystallography, there are a variety of cleavage types. Cleavage occurs typically in either one, two, three, four, or six directions. Basal cleavage in one direction is a distinctive property of the micas. Two-directional cleavage is described as prismatic, and occurs in minerals such as the amphiboles and pyroxenes. Minerals such as galena or halite have cubic (or isometric) cleavage in three directions, at 90°; when three directions of cleavage are present, but not at 90°, such as in calcite or rhodochrosite, it is termed rhombohedral cleavage. Octahedral cleavage (four directions) is present in fluorite and diamond, and sphalerite has six-directional dodecahedral cleavage. Minerals with many cleavages might not break equally well in all of the directions; for example, calcite has good cleavage in three directions, but gypsum has perfect cleavage in one direction, and poor cleavage in two other directions. Angles between cleavage planes vary between minerals. For example, as the amphiboles are double-chain silicates and the pyroxenes are single-chain silicates, the angle between their cleavage planes is different. The pyroxenes cleave in two directions at approximately 90°, whereas the amphiboles distinctively cleave in two directions separated by approximately 120° and 60°. The cleavage angles can be measured with a contact goniometer, which is similar to a protractor. Parting, sometimes called "false cleavage", is similar in appearance to cleavage but is instead produced by structural defects in the mineral, as opposed to systematic weakness. Parting varies from crystal to crystal of a mineral, whereas all crystals of a given mineral will cleave if the atomic structure allows for that property. In general, parting is caused by some stress applied to a crystal. The sources of the stresses include deformation (e.g. an increase in pressure), exsolution, or twinning. Minerals that often display parting include the pyroxenes, hematite, magnetite, and corundum. When a mineral is broken in a direction that does not correspond to a plane of cleavage, it is termed to have been fractured. There are several types of uneven fracture. The classic example is conchoidal fracture, like that of quartz; rounded surfaces are created, which are marked by smooth curved lines. This type of fracture occurs only in very homogeneous minerals. Other types of fracture are fibrous, splintery, and hackly. The latter describes a break along a rough, jagged surface; an example of this property is found in native copper. Tenacity is related to both cleavage and fracture. Whereas fracture and cleavage describes the surfaces that are created when a mineral is broken, tenacity describes how resistant a mineral is to such breaking. Minerals can be described as brittle, ductile, malleable, sectile, flexible, or elastic. Specific gravity numerically describes the density of a mineral. The dimensions of density are mass divided by volume with units: kg/m or g/cm. Specific gravity is defined as the density of the mineral divided by the density of water at 4 °C and thus is a dimensionless quantity, identical in all unit systems. It can be measured as the quotient of the mass of the sample and difference between the weight of the sample in air and its corresponding weight in water. Among most minerals, this property is not diagnostic. Rock forming minerals – typically silicates or occasionally carbonates – have a specific gravity of 2.5–3.5. High specific gravity is a diagnostic property of a mineral. A variation in chemistry (and consequently, mineral class) correlates to a change in specific gravity. Among more common minerals, oxides and sulfides tend to have a higher specific gravity as they include elements with higher atomic mass. A generalization is that minerals with metallic or adamantine lustre tend to have higher specific gravities than those having a non-metallic to dull lustre. For example, hematite, Fe2O3, has a specific gravity of 5.26 while galena, PbS, has a specific gravity of 7.2–7.6, which is a result of their high iron and lead content, respectively. A very high specific gravity is characteristic of native metals; for example, kamacite, an iron-nickel alloy common in iron meteorites has a specific gravity of 7.9, and gold has an observed specific gravity between 15 and 19.3. Other properties can be used to diagnose minerals. These are less general, and apply to specific minerals. Dropping dilute acid (often 10% HCl) onto a mineral aids in distinguishing carbonates from other mineral classes. The acid reacts with the carbonate ([CO3]) group, which causes the affected area to effervesce, giving off carbon dioxide gas. This test can be further expanded to test the mineral in its original crystal form or powdered form. An example of this test is done when distinguishing calcite from dolomite, especially within the rocks (limestone and dolomite respectively). Calcite immediately effervesces in acid, whereas acid must be applied to powdered dolomite (often to a scratched surface in a rock), for it to effervesce. Zeolite minerals will not effervesce in acid; instead, they become frosted after 5–10 minutes, and if left in acid for a day, they dissolve or become a silica gel. Magnetism is a very conspicuous property of a few minerals. Among common minerals, magnetite exhibits this property strongly, and magnetism is also present, albeit not as strongly, in pyrrhotite and ilmenite. Some minerals exhibit electrical properties – for example, quartz is piezoelectric – but electrical properties are rarely used as diagnostic criteria for minerals because of incomplete data and natural variation. Minerals can also be tested for taste or smell. Halite, NaCl, is table salt; its potassium-bearing counterpart, sylvite, has a pronounced bitter taste. Sulfides have a characteristic smell, especially as samples are fractured, reacting, or powdered. Radioactivity is a rare property found in minerals containing radioactive elements. The radioactive elements could be a defining constituent, such as uranium in uraninite, autunite, and carnotite, or present as trace impurities, as in zircon. The decay of a radioactive element damages the mineral crystal structure rendering it locally amorphous (metamict state); the optical result, termed a radioactive halo or pleochroic halo, is observable with various techniques, such as thin-section petrography. In 315 BCE, Theophrastus presented his classification of minerals in his treatise On Stones. His classification was influenced by the ideas of his teachers Plato and Aristotle. Theophrastus classified minerals as stones, earths or metals. Georgius Agricola's classification of minerals in his book De Natura Fossilium, published in 1546, divided minerals into three types of substance: simple (stones, earths, metals, and congealed juices), compound (intimately mixed) and composite (separable). An early classification of minerals was given by Carl Linnaeus in his seminal 1735 book Systema Naturae. He divided the natural world into three kingdoms – plants, animals, and minerals – and classified each with the same hierarchy. In descending order, these were Phylum, Class, Order, Family, Tribe, Genus, and Species. However, while his system was justified by Charles Darwin's theory of species formation and has been largely adopted and expanded by biologists in the following centuries (who still use his Greek- and Latin-based binomial naming scheme), it had little success among mineralogists (although each distinct mineral is still formally referred to as a mineral species). Minerals are classified by variety, species, series and group, in order of increasing generality. The basic level of definition is that of mineral species, each of which is distinguished from the others by unique chemical and physical properties. For example, quartz is defined by its formula, SiO2, and a specific crystalline structure that distinguishes it from other minerals with the same chemical formula (termed polymorphs). When there exists a range of composition between two minerals species, a mineral series is defined. For example, the biotite series is represented by variable amounts of the endmembers phlogopite, siderophyllite, annite, and eastonite. In contrast, a mineral group is a grouping of mineral species with some common chemical properties that share a crystal structure. The pyroxene group has a common formula of XY(Si,Al)2O6, where X and Y are both cations, with X typically bigger than Y; the pyroxenes are single-chain silicates that crystallize in either the orthorhombic or monoclinic crystal systems. Finally, a mineral variety is a specific type of mineral species that differs by some physical characteristic, such as colour or crystal habit. An example is amethyst, which is a purple variety of quartz. Two common classifications, Dana and Strunz, are used for minerals; both rely on composition, specifically with regards to important chemical groups, and structure. James Dwight Dana, a leading geologist of his time, first published his System of Mineralogy in 1837; as of 1997, it is in its eighth edition. The Dana classification assigns a four-part number to a mineral species. Its class number is based on important compositional groups; the type gives the ratio of cations to anions in the mineral, and the last two numbers group minerals by structural similarity within a given type or class. The less commonly used Strunz classification, named for German mineralogist Karl Hugo Strunz, is based on the Dana system, but combines both chemical and structural criteria, the latter with regards to distribution of chemical bonds. As the composition of the Earth's crust is dominated by silicon and oxygen, silicates are by far the most important class of minerals in terms of rock formation and diversity. However, non-silicate minerals are of great economic importance, especially as ores. Non-silicate minerals are subdivided into several other classes by their dominant chemistry, which includes native elements, sulfides, halides, oxides and hydroxides, carbonates and nitrates, borates, sulfates, phosphates, and organic compounds. Most non-silicate mineral species are rare (constituting in total 8% of the Earth's crust), although some are relatively common, such as calcite, pyrite, magnetite, and hematite. There are two major structural styles observed in non-silicates: close-packing and silicate-like linked tetrahedra. Close-packed structures are a way to densely pack atoms while minimizing interstitial space. Hexagonal close-packing involves stacking layers where every other layer is the same ("ababab"), whereas cubic close-packing involves stacking groups of three layers ("abcabcabc"). Analogues to linked silica tetrahedra include SO4 (sulfate), PO4 (phosphate), AsO4 (arsenate), and VO4 (vanadate) structures. The non-silicates have great economic importance, as they concentrate elements more than the silicate minerals do. The largest grouping of minerals by far are the silicates; most rocks are composed of greater than 95% silicate minerals, and over 90% of the Earth's crust is composed of these minerals. The two main constituents of silicates are silicon and oxygen, which are the two most abundant elements in the Earth's crust. Other common elements in silicate minerals correspond to other common elements in the Earth's crust, such as aluminium, magnesium, iron, calcium, sodium, and potassium. Some important rock-forming silicates include the feldspars, quartz, olivines, pyroxenes, amphiboles, garnets, and micas. The base unit of a silicate mineral is the [SiO4] tetrahedron. In the vast majority of cases, silicon is in four-fold or tetrahedral coordination with oxygen. In very high-pressure situations, silicon will be in six-fold or octahedral coordination, such as in the perovskite structure or the quartz polymorph stishovite (SiO2). In the latter case, the mineral no longer has a silicate structure, but that of rutile (TiO2), and its associated group, which are simple oxides. These silica tetrahedra are then polymerized to some degree to create various structures, such as one-dimensional chains, two-dimensional sheets, and three-dimensional frameworks. The basic silicate mineral where no polymerization of the tetrahedra has occurred requires other elements to balance out the base 4- charge. In other silicate structures, different combinations of elements are required to balance out the resultant negative charge. It is common for the Si to be substituted by Al because of similarity in ionic radius and charge; in those cases, the [AlO4] tetrahedra form the same structures as do the unsubstituted tetrahedra, but their charge-balancing requirements are different. The degree of polymerization can be described by both the structure formed and how many tetrahedral corners (or coordinating oxygens) are shared (for aluminium and silicon in tetrahedral sites): The silicate subclasses are described below in order of decreasing polymerization. Tectosilicates, also known as framework silicates, have the highest degree of polymerization. With all corners of a tetrahedra shared, the silicon:oxygen ratio becomes 1:2. Examples are quartz, the feldspars, feldspathoids, and the zeolites. Framework silicates tend to be particularly chemically stable as a result of strong covalent bonds. Forming 12% of the Earth's crust, quartz (SiO2) is the most abundant mineral species. It is characterized by its high chemical and physical resistivity. Quartz has several polymorphs, including tridymite and cristobalite at high temperatures, high-pressure coesite, and ultra-high pressure stishovite. The latter mineral can only be formed on Earth by meteorite impacts, and its structure has been compressed so much that it has changed from a silicate structure to that of rutile (TiO2). The silica polymorph that is most stable at the Earth's surface is α-quartz. Its counterpart, β-quartz, is present only at high temperatures and pressures (changes to α-quartz below 573 °C at 1 bar). These two polymorphs differ by a "kinking" of bonds; this change in structure gives β-quartz greater symmetry than α-quartz, and they are thus also called high quartz (β) and low quartz (α). Feldspars are the most abundant group in the Earth's crust, at about 50%. In the feldspars, Al substitutes for Si, which creates a charge imbalance that must be accounted for by the addition of cations. The base structure becomes either [AlSi3O8] or [Al2Si2O8] There are 22 mineral species of feldspars, subdivided into two major subgroups – alkali and plagioclase – and two less common groups – celsian and banalsite. The alkali feldspars are most commonly in a series between potassium-rich orthoclase and sodium-rich albite; in the case of plagioclase, the most common series ranges from albite to calcium-rich anorthite. Crystal twinning is common in feldspars, especially polysynthetic twins in plagioclase and Carlsbad twins in alkali feldspars. If the latter subgroup cools slowly from a melt, it forms exsolution lamellae because the two components – orthoclase and albite – are unstable in solid solution. Exsolution can be on a scale from microscopic to readily observable in hand-sample; perthitic texture forms when Na-rich feldspar exsolve in a K-rich host. The opposite texture (antiperthitic), where K-rich feldspar exsolves in a Na-rich host, is very rare. Feldspathoids are structurally similar to feldspar, but differ in that they form in Si-deficient conditions, which allows for further substitution by Al. As a result, feldspathoids are almost never found in association with quartz. A common example of a feldspathoid is nepheline ((Na, K)AlSiO4); compared to alkali feldspar, nepheline has an Al2O3:SiO2 ratio of 1:2, as opposed to 1:6 in alkali feldspar. Zeolites often have distinctive crystal habits, occurring in needles, plates, or blocky masses. They form in the presence of water at low temperatures and pressures, and have channels and voids in their structure. Zeolites have several industrial applications, especially in waste water treatment. Phyllosilicates consist of sheets of polymerized tetrahedra. They are bound at three oxygen sites, which gives a characteristic silicon:oxygen ratio of 2:5. Important examples include the mica, chlorite, and the kaolinite-serpentine groups. In addition to the tetrahedra, phyllosilicates have a sheet of octahedra (elements in six-fold coordination by oxygen) that balance out the basic tetrahedra, which have a negative charge (e.g. [Si4O10]) These tetrahedra (T) and octahedra (O) sheets are stacked in a variety of combinations to create phyllosilicate layers. Within an octahedral sheet, there are three octahedral sites in a unit structure; however, not all of the sites may be occupied. In that case, the mineral is termed dioctahedral, whereas in other case it is termed trioctahedral. The layers are weakly bound by van der Waals forces, hydrogen bonds, or sparse ionic bonds, which causes a crystallographic weakness, in turn leading to a prominent basal cleavage among the phyllosilicates. The kaolinite-serpentine group consists of T-O stacks (the 1:1 clay minerals); their hardness ranges from 2 to 4, as the sheets are held by hydrogen bonds. The 2:1 clay minerals (pyrophyllite-talc) consist of T-O-T stacks, but they are softer (hardness from 1 to 2), as they are instead held together by van der Waals forces. These two groups of minerals are subgrouped by octahedral occupation; specifically, kaolinite and pyrophyllite are dioctahedral whereas serpentine and talc trioctahedral. Micas are also T-O-T-stacked phyllosilicates, but differ from the other T-O-T and T-O-stacked subclass members in that they incorporate aluminium into the tetrahedral sheets (clay minerals have Al in octahedral sites). Common examples of micas are muscovite, and the biotite series. Mica T-O-T layers are bonded together by metal ions, giving them a greater hardness than other phyllosilicate minerals, though they retain perfect basal cleavage. The chlorite group is related to mica group, but a brucite-like (Mg(OH)2) layer between the T-O-T stacks. Because of their chemical structure, phyllosilicates typically have flexible, elastic, transparent layers that are electrical insulators and can be split into very thin flakes. Micas can be used in electronics as insulators, in construction, as optical filler, or even cosmetics. Chrysotile, a species of serpentine, is the most common mineral species in industrial asbestos, as it is less dangerous in terms of health than the amphibole asbestos. Inosilicates consist of tetrahedra repeatedly bonded in chains. These chains can be single, where a tetrahedron is bound to two others to form a continuous chain; alternatively, two chains can be merged to create double-chain silicates. Single-chain silicates have a silicon:oxygen ratio of 1:3 (e.g. [Si2O6]), whereas the double-chain variety has a ratio of 4:11, e.g. [Si8O22]. Inosilicates contain two important rock-forming mineral groups; single-chain silicates are most commonly pyroxenes, while double-chain silicates are often amphiboles. Higher-order chains exist (e.g. three-member, four-member, five-member chains, etc.) but they are rare. The pyroxene group consists of 21 mineral species. Pyroxenes have a general structure formula of XY(Si2O6), where X is an octahedral site, while Y can vary in coordination number from six to eight. Most varieties of pyroxene consist of permutations of Ca, Fe and Mg to balance the negative charge on the backbone. Pyroxenes are common in the Earth's crust (about 10%) and are a key constituent of mafic igneous rocks. Amphiboles have great variability in chemistry, described variously as a "mineralogical garbage can" or a "mineralogical shark swimming a sea of elements". The backbone of the amphiboles is the [Si8O22]; it is balanced by cations in three possible positions, although the third position is not always used, and one element can occupy both remaining ones. Finally, the amphiboles are usually hydrated, that is, they have a hydroxyl group ([OH]), although it can be replaced by a fluoride, a chloride, or an oxide ion. Because of the variable chemistry, there are over 80 species of amphibole, although variations, as in the pyroxenes, most commonly involve mixtures of Ca, Fe and Mg. Several amphibole mineral species can have an asbestiform crystal habit. These asbestos minerals form long, thin, flexible, and strong fibres, which are electrical insulators, chemically inert and heat-resistant; as such, they have several applications, especially in construction materials. However, asbestos are known carcinogens, and cause various other illnesses, such as asbestosis; amphibole asbestos (anthophyllite, tremolite, actinolite, grunerite, and riebeckite) are considered more dangerous than chrysotile serpentine asbestos. Cyclosilicates, or ring silicates, have a ratio of silicon to oxygen of 1:3. Six-member rings are most common, with a base structure of [Si6O18]; examples include the tourmaline group and beryl. Other ring structures exist, with 3, 4, 8, 9, 12 having been described. Cyclosilicates tend to be strong, with elongated, striated crystals. Tourmalines have a very complex chemistry that can be described by a general formula XY3Z6(BO3)3T6O18V3W. The T6O18 is the basic ring structure, where T is usually Si, but substitutable by Al3+ or B. Tourmalines can be subgrouped by the occupancy of the X site, and from there further subdivided by the chemistry of the W site. The Y and Z sites can accommodate a variety of cations, especially various transition metals; this variability in structural transition metal content gives the tourmaline group greater variability in colour. Other cyclosilicates include beryl, Al2Be3Si6O18, whose varieties include the gemstones emerald (green) and aquamarine (bluish). Cordierite is structurally similar to beryl, and is a common metamorphic mineral. Sorosilicates, also termed disilicates, have tetrahedron-tetrahedron bonding at one oxygen, which results in a 2:7 ratio of silicon to oxygen. The resultant common structural element is the [Si2O7] group. The most common disilicates by far are members of the epidote group. Epidotes are found in variety of geologic settings, ranging from mid-ocean ridge to granites to metapelites. Epidotes are built around the structure [(SiO4)(Si2O7)] structure; for example, the mineral species epidote has calcium, aluminium, and ferric iron to charge balance: Ca2Al2(Fe, Al)(SiO4)(Si2O7)O(OH). The presence of iron as Fe and Fe helps buffer oxygen fugacity, which in turn is a significant factor in petrogenesis. Other examples of sorosilicates include lawsonite, a metamorphic mineral forming in the blueschist facies (subduction zone setting with low temperature and high pressure), vesuvianite, which takes up a significant amount of calcium in its chemical structure. Orthosilicates consist of isolated tetrahedra that are charge-balanced by other cations. Also termed nesosilicates, this type of silicate has a silicon:oxygen ratio of 1:4 (e.g. SiO4). Typical orthosilicates tend to form blocky equant crystals, and are fairly hard. Several rock-forming minerals are part of this subclass, such as the aluminosilicates, the olivine group, and the garnet group. The aluminosilicates –bkyanite, andalusite, and sillimanite, all Al2SiO5 – are structurally composed of one [SiO4] tetrahedron, and one Al in octahedral coordination. The remaining Al can be in six-fold coordination (kyanite), five-fold (andalusite) or four-fold (sillimanite); which mineral forms in a given environment is depend on pressure and temperature conditions. In the olivine structure, the main olivine series of (Mg, Fe)2SiO4 consist of magnesium-rich forsterite and iron-rich fayalite. Both iron and magnesium are in octahedral by oxygen. Other mineral species having this structure exist, such as tephroite, Mn2SiO4. The garnet group has a general formula of X3Y2(SiO4)3, where X is a large eight-fold coordinated cation, and Y is a smaller six-fold coordinated cation. There are six ideal endmembers of garnet, split into two group. The pyralspite garnets have Al in the Y position: pyrope (Mg3Al2(SiO4)3), almandine (Fe3Al2(SiO4)3), and spessartine (Mn3Al2(SiO4)3). The ugrandite garnets have Ca in the X position: uvarovite (Ca3Cr2(SiO4)3), grossular (Ca3Al2(SiO4)3) and andradite (Ca3Fe2(SiO4)3). While there are two subgroups of garnet, solid solutions exist between all six end-members. Other orthosilicates include zircon, staurolite, and topaz. Zircon (ZrSiO4) is useful in geochronology as U can substitute for Zr; furthermore, because of its very resistant structure, it is difficult to reset it as a chronometer. Staurolite is a common metamorphic intermediate-grade index mineral. It has a particularly complicated crystal structure that was only fully described in 1986. Topaz (Al2SiO4(F, OH)2, often found in granitic pegmatites associated with tourmaline, is a common gemstone mineral. Native elements are those that are not chemically bonded to other elements. This mineral group includes native metals, semi-metals, and non-metals, and various alloys and solid solutions. The metals are held together by metallic bonding, which confers distinctive physical properties such as their shiny metallic lustre, ductility and malleability, and electrical conductivity. Native elements are subdivided into groups by their structure or chemical attributes. The gold group, with a cubic close-packed structure, includes metals such as gold, silver, and copper. The platinum group is similar in structure to the gold group. The iron-nickel group is characterized by several iron-nickel alloy species. Two examples are kamacite and taenite, which are found in iron meteorites; these species differ by the amount of Ni in the alloy; kamacite has less than 5–7% nickel and is a variety of native iron, whereas the nickel content of taenite ranges from 7–37%. Arsenic group minerals consist of semi-metals, which have only some metallic traits; for example, they lack the malleability of metals. Native carbon occurs in two allotropes, graphite and diamond; the latter forms at very high pressure in the mantle, which gives it a much stronger structure than graphite. The sulfide minerals are chemical compounds of one or more metals or semimetals with a chalcogen or pnictogen, of which sulfur is most common. Tellurium, arsenic, or selenium can substitute for the sulfur. Sulfides tend to be soft, brittle minerals with a high specific gravity. Many powdered sulfides, such as pyrite, have a sulfurous smell when powdered. Sulfides are susceptible to weathering, and many readily dissolve in water; these dissolved minerals can be later redeposited, which creates enriched secondary ore deposits. Sulfides are classified by the ratio of the metal or semimetal to the sulfur, such as M:S equal to 2:1, or 1:1. Many sulfide minerals are economically important as metal ores; examples include sphalerite (ZnS), an ore of zinc, galena (PbS), an ore of lead, cinnabar (HgS), an ore of mercury, and molybdenite (MoS2, an ore of molybdenum. Pyrite (FeS2), is the most commonly occurring sulfide, and can be found in most geological environments. It is not, however, an ore of iron, but can be instead oxidized to produce sulfuric acid. Related to the sulfides are the rare sulfosalts, in which a metallic element is bonded to sulfur and a semimetal such as antimony, arsenic, or bismuth. Like the sulfides, sulfosalts are typically soft, heavy, and brittle minerals. Oxide minerals are divided into three categories: simple oxides, hydroxides, and multiple oxides. Simple oxides are characterized by O as the main anion and primarily ionic bonding. They can be further subdivided by the ratio of oxygen to the cations. The periclase group consists of minerals with a 1:1 ratio. Oxides with a 2:1 ratio include cuprite (Cu2O) and water ice. Corundum group minerals have a 2:3 ratio, and includes minerals such as corundum (Al2O3), and hematite (Fe2O3). Rutile group minerals have a ratio of 1:2; the eponymous species, rutile (TiO2) is the chief ore of titanium; other examples include cassiterite (SnO2; ore of tin), and pyrolusite (MnO2; ore of manganese). In hydroxides, the dominant anion is the hydroxyl ion, OH. Bauxites are the chief aluminium ore, and are a heterogeneous mixture of the hydroxide minerals diaspore, gibbsite, and bohmite; they form in areas with a very high rate of chemical weathering (mainly tropical conditions). Finally, multiple oxides are compounds of two metals with oxygen. A major group within this class are the spinels, with a general formula of XY2O4. Examples of species include spinel (MgAl2O4), chromite (FeCr2O4), and magnetite (Fe3O4). The latter is readily distinguishable by its strong magnetism, which occurs as it has iron in two oxidation states (FeFe2O4), which makes it a multiple oxide instead of a single oxide. The halide minerals are compounds in which a halogen (fluorine, chlorine, iodine, or bromine) is the main anion. These minerals tend to be soft, weak, brittle, and water-soluble. Common examples of halides include halite (NaCl, table salt), sylvite (KCl), and fluorite (CaF2). Halite and sylvite commonly form as evaporites, and can be dominant minerals in chemical sedimentary rocks. Cryolite, Na3AlF6, is a key mineral in the extraction of aluminium from bauxites; however, as the only significant occurrence at Ivittuut, Greenland, in a granitic pegmatite, was depleted, synthetic cryolite can be made from fluorite. The carbonate minerals are those in which the main anionic group is carbonate, [CO3]. Carbonates tend to be brittle, many have rhombohedral cleavage, and all react with acid. Due to the last characteristic, field geologists often carry dilute hydrochloric acid to distinguish carbonates from non-carbonates. The reaction of acid with carbonates, most commonly found as the polymorph calcite and aragonite (CaCO3), relates to the dissolution and precipitation of the mineral, which is a key in the formation of limestone caves, features within them such as stalactite and stalagmites, and karst landforms. Carbonates are most often formed as biogenic or chemical sediments in marine environments. The carbonate group is structurally a triangle, where a central C cation is surrounded by three O anions; different groups of minerals form from different arrangements of these triangles. The most common carbonate mineral is calcite, which is the primary constituent of sedimentary limestone and metamorphic marble. Calcite, CaCO3, can have a significant percentage of magnesium substituting for calcium. Under high-Mg conditions, its polymorph aragonite will form instead; the marine geochemistry in this regard can be described as an aragonite or calcite sea, depending on which mineral preferentially forms. Dolomite is a double carbonate, with the formula CaMg(CO3)2. Secondary dolomitization of limestone is common, in which calcite or aragonite are converted to dolomite; this reaction increases pore space (the unit cell volume of dolomite is 88% that of calcite), which can create a reservoir for oil and gas. These two mineral species are members of eponymous mineral groups: the calcite group includes carbonates with the general formula XCO3, and the dolomite group constitutes minerals with the general formula XY(CO3)2. The sulfate minerals all contain the sulfate anion, [SO4]. They tend to be transparent to translucent, soft, and many are fragile. Sulfate minerals commonly form as evaporites, where they precipitate out of evaporating saline waters. Sulfates can also be found in hydrothermal vein systems associated with sulfides, or as oxidation products of sulfides. Sulfates can be subdivided into anhydrous and hydrous minerals. The most common hydrous sulfate by far is gypsum, CaSO4⋅2H2O. It forms as an evaporite, and is associated with other evaporites such as calcite and halite; if it incorporates sand grains as it crystallizes, gypsum can form desert roses. Gypsum has very low thermal conductivity and maintains a low temperature when heated as it loses that heat by dehydrating; as such, gypsum is used as an insulator in materials such as plaster and drywall. The anhydrous equivalent of gypsum is anhydrite; it can form directly from seawater in highly arid conditions. The barite group has the general formula XSO4, where the X is a large 12-coordinated cation. Examples include barite (BaSO4), celestine (SrSO4), and anglesite (PbSO4); anhydrite is not part of the barite group, as the smaller Ca is only in eight-fold coordination. The phosphate minerals are characterized by the tetrahedral [PO4] unit, although the structure can be generalized, and phosphorus is replaced by antimony, arsenic, or vanadium. The most common phosphate is the apatite group; common species within this group are fluorapatite (Ca5(PO4)3F), chlorapatite (Ca5(PO4)3Cl) and hydroxylapatite (Ca5(PO4)3(OH)). Minerals in this group are the main crystalline constituents of teeth and bones in vertebrates. The relatively abundant monazite group has a general structure of ATO4, where T is phosphorus or arsenic, and A is often a rare-earth element (REE). Monazite is important in two ways: first, as a REE "sink", it can sufficiently concentrate these elements to become an ore; secondly, monazite group elements can incorporate relatively large amounts of uranium and thorium, which can be used in monazite geochronology to date the rock based on the decay of the U and Th to lead. The Strunz classification includes a class for organic minerals. These rare compounds contain organic carbon, but can be formed by a geologic process. For example, whewellite, CaC2O4⋅H2O is an oxalate that can be deposited in hydrothermal ore veins. While hydrated calcium oxalate can be found in coal seams and other sedimentary deposits involving organic matter, the hydrothermal occurrence is not considered to be related to biological activity. Mineral classification schemes and their definitions are evolving to match recent advances in mineral science. Recent changes have included the addition of an organic class, in both the new Dana and the Strunz classification schemes. The organic class includes a very rare group of minerals with hydrocarbons. The IMA Commission on New Minerals and Mineral Names adopted in 2009 a hierarchical scheme for the naming and classification of mineral groups and group names and established seven commissions and four working groups to review and classify minerals into an official listing of their published names. According to these new rules, "mineral species can be grouped in a number of different ways, on the basis of chemistry, crystal structure, occurrence, association, genetic history, or resource, for example, depending on the purpose to be served by the classification." It has been suggested that biominerals could be important indicators of extraterrestrial life and thus could play an important role in the search for past or present life on Mars. Furthermore, organic components (biosignatures) that are often associated with biominerals are believed to play crucial roles in both pre-biotic and biotic reactions. In January 2014, NASA reported that studies by the Curiosity and Opportunity rovers on Mars would search for evidence of ancient life, including a biosphere based on autotrophic, chemotrophic and/or chemolithoautotrophic microorganisms, as well as ancient water, including fluvio-lacustrine environments (plains related to ancient rivers or lakes) that may have been habitable. The search for evidence of habitability, taphonomy (related to fossils), and organic carbon on the planet Mars became a primary NASA objective.
[[File:Swanson_Shoe_Repair_18.jpg|thumb|Shoemakers are a type of craftsman]] '''Craft''' or '''[[handicraft]]''' is about making things with one's own hands and skills. The different types of crafts can be put in groups according to the material being used. In the [[Middle Ages]] the most common materials were [[metal]], [[wood]] or [[clay]]. A [[craftsman]] is a person who has the [[knowledge]] and skills of a craft. When they have a lot of experience they may be called a ''master craftsman''. If they are young people learning a craft they are called an ''[[apprentice]]''. [[Category:Occupations|*]] {{stub}}
A craft or trade is a pastime or an occupation that requires particular skills and knowledge of skilled work. In a historical sense, particularly the Middle Ages and earlier, the term is usually applied to people occupied in small scale production of goods, or their maintenance, for example by tinkers. The traditional term craftsman is nowadays often replaced by artisan and by craftsperson. Historically, the more specialized crafts with high-value products tended to concentrate in urban centers and their practitioners formed guilds. The skill required by their professions and the need to be permanently involved in the exchange of goods often demanded a higher level of education, and craftsmen were usually in a more privileged position than the peasantry in societal hierarchy. The households of craftsmen were not as self-sufficient as those of people engaged in agricultural work, and therefore had to rely on the exchange of goods. Some crafts, especially in areas such as pottery, woodworking, and various stages of textile production, could be practiced on a part-time basis by those also working in agriculture, and often formed part of village life. When an apprentice finished his apprenticeship, he became a journeyman searching for a place to set up his own shop and make a living. After he set up his own shop, he could then call himself a master of his craft. This stepwise approach to mastery of a craft, which includes the attainment of some education and skill, has survived in some countries to the present day. But crafts have undergone deep structural changes since and during the era of the Industrial Revolution. The mass production of goods by large-scale industry has limited crafts to market segments in which industry's modes of functioning or its mass-produced goods do not satisfy the preferences of potential buyers. As an outcome of these changes, craftspeople today increasingly make use of semi-finished components or materials and adapt these to their customers' requirements or demands. Thus, they participate in a certain division of labour between industry and craft. The nature of craft skill and the process of its development are continually debated by philosophers, anthropologists, and cognitive scientists. Some scholars note that craft skill is marked by particular ways of experiencing tools and materials, whether by allowing tools to recede from focal awareness, perceiving tools and materials in terms of their practical interrelationships, or seeing aspects of work that are invisible to the untrained observer. Other scholars working on craft skill focus on observational learning and mimicry, exploring how learners visually parse the movements of experts. Certain researchers even de-emphasize the role of the individual craftsperson, noting the collective nature of craft understanding or emphasizing the role of materials as collaborators in the process of production. There are three aspects to human creativity: art, crafts, and science. Roughly, art relies upon intuitive sensing, vision, and expression; crafts upon sophisticated technique; and science upon knowledge. Handicraft is the "traditional" main sector of the crafts. It is a type of work where useful and decorative devices are made completely by hand or by using only simple tools. The term is usually applied to traditional means of making goods. The individual artisanship of the items is a paramount criterion, an such items often have cultural and/or religious significance. Items made by mass production or machines are not handicraft goods. The beginning of crafts in areas like the Ottoman Empire involved the governing bodies requiring members of the city who were skilled at creating goods to open shops in the center of town. These people slowly stopped acting as subsistence farmers (who created goods in their own homes to trade with neighbors) and began to represent what we think of as "craftspeople" today. Besides traditional goods, handicraft contributes to the field of computing by combining craft practices with technology. For example, in 1968, the Apollo 8 spacecraft's core memory consisted of wires that were woven around and through electromagnetic cores by hand. The core rope memory they created contained information used to successfully complete the mission. Crafts and craftspeople have become a subject of academic study. For example, Stephanie Bunn was an artist before she became an anthropologist, and she went on to develop an academic interest in the process of craft. She argues that what happens to an object before it becomes a "product" is an area worthy of study. The term crafts is used to describe artistic practices within the family of decorative arts that traditionally are defined by their relationship to functional or utilitarian products (such as sculptural forms in the vessel tradition) or by their use of such natural media as wood, clay, ceramics, glass, textiles, and metal. The Arts and Crafts Movement originated in Britain during the late 19th century and was characterized by a style of decoration reminiscent of medieval times. The primary artist associated with the movement is William Morris, whose work was reinforced with writings from John Ruskin. The movement placed a high importance on the quality of craftsmanship, while emphasizing the importance for the arts to contribute to economic reform. Crafts practiced by independent artists working alone or in small groups are referred to as studio craft. Studio craft includes studio pottery, metalwork, weaving, woodturning, paper and other forms of woodworking, glassblowing, and glass art. A craft fair is an organized event to display and sell crafts. There are also craft stores where such goods are sold and craft communities, such as Craftster, where expertise is shared. A tradesperson is a skilled manual worker in a particular trade or craft. Economically and socially, a tradesperson's status is considered between a laborer and a professional, with a high degree of both practical and theoretical knowledge of their trade. In cultures where professional careers are highly prized, there can be a shortage of skilled manual workers, leading to lucrative niche markets in the trades. Media related to Crafts at Wikimedia Commons
{{nosource|date=May 2013}} {{about|the element|the color|Silver (color)}} [[File:1000oz.silver.bullion.bar.underneath.jpg|thumb|Silver bar, or [[ingot]]]] '''Silver''' (symbol '''Ag''') is a [[chemical element]]. In [[chemistry]], silver is element 47, a [[transition metal]]. It has an [[atomic weight]] of 107.86 a.m.u. Its symbol is Ag, from the [[Latin]] word for silver, ''argentum''. ==Properties== ===Physical properties=== Silver is a soft [[metal]]. It is also a [[precious metal]]. When it is used in money or in jewellery, it is often [[alloy]]<nowiki/>ed with [[gold]] or some other metal to make it [[hardness|harder]]. It is bluish-white. Among metals, it is the most [[Reflection|reflective]] and the best [[Electrical conductor|conductor]] of [[electricity]] and heat. It is considered a precious metal. It is very [[malleable]] and [[wikt:ductile|ductile]], which means it can be pulled into [[wire]] or hammered into thin sheets. Silver is one of the few words in the [[English language]] that does not rhyme with any other word. Silver [[coins]] and [[wikt:bar|bars]] can be bought and sold at [[coin]] [[shop]]s around the [[world]]. ===Chemical properties=== [[File:Silver-264109.jpg|thumb|Natural silver tarnished]] Silver is less reactive than most other metals. It does not dissolve in most acids. [[Nitric acid]] dissolves it, though, to make [[silver nitrate]]. It does react with strong [[oxidizing agent]]s like [[potassium dichromate]] or [[potassium permanganate]]. It does not corrode easily. It only corrodes when there is [[hydrogen sulfide]] in the air. Then, it forms a black coating known as [[wikt:tarnish|tarnish]]. [[File:Silver(I)-oxide-sample.jpg|thumb|[[Silver(I) oxide]]]] Silver exists in two main [[oxidation state]]s: +1 and +2. The +1 is much more common. A few compounds exist in the +2 oxidation state, but they are very strong oxidizing agents. Silver compounds can be brown, black, yellow, gray, or colorless. Silver compounds are disinfectants. ;Silver(I) compounds Silver(I) compounds are [[oxidizing agent]]s. They are more common. Most of them are very expensive. *[[Silver bromide]], light yellow *[[Silver carbonate]], yellowish *[[Silver chloride]], white *[[Silver(I) fluoride]], yellow-brown *[[Silver iodate]], colorless *[[Silver iodide]], yellow *[[Silver nitrate]], colorless *[[Silver oxide]], brown-black *[[Silver sulfide]], black ;Silver(II) compounds Silver(II) compounds are powerful oxidizing agents and rare. *[[Silver(II) fluoride]], strong oxidizing agent, highly reactive, white or gray ==Occurrence== [[File:Silver-Calcite-252603.jpg|thumb|left|Silver as a metal]] Silver can be found as a native metal. Most silver is found with [[copper]], [[lead]], or gold in [[mineral|rocks]]. The rocks are found mostly in [[Canada]], [[Mexico]], [[Peru]], and the [[United States]]. Peru [[produce]]s the most silver, mostly in mines that are operated for copper and [[zinc]]. Silver is also in [[chemical compound]]s. [[Acanthite]] is a silver [[ore]] that is a silver compound.[[File:Acanthite-40496.jpg|thumb|Acanthite]] ==Preparation== Silver is extracted from the earth in several ways. It is normally extracted using [[electrolysis]]. ==Uses== ===As an element=== Silver has been used for many thousands of years by [[wikt:people|people]] all over the [[world]], for [[jewellery]], as [[money]], and many other things. It is called a white metal even though it looks [[grey]]. The word '''silver''' is also used to talk about this color or [[shade]] of grey. Silver is also used for [[wikt:utensil|utensils]]. It may be used to fill teeth in [[dentistry]] as an [[amalgam]]. Silver is used as a [[catalyst]]. ===In compounds=== Silver compounds are disinfectants. It can [[kill]] [[bacteria]] and has other useful properties. It is used in the [[silver oxide battery]]. They are also used in [[photographic film]]. They can also be used to reduce odors in clothes. Some silver compounds are used in creams that help [[burn]]s heal. ==History== Silver has used for thousands of years. It was normally considered second to gold in value. Many people including Romans and Chinese used silver as money. The symbol Ag is from the [[Latin]] name for silver, ''argentum''. Silver was also used to prevent infections and decay. ==Safety== Silver is not a large [[wikt:danger|danger]] to [[human]]s. Silver compounds are toxic. They make the skin turn blue. Some can be [[carcinogen]]s. Colloidal silver, a common [[wikt:homeopathy|homeopathic remedy]], is not toxic in normal amounts, but it does not do much. Silver is very toxic to fish, so silver-containing waste (for example, photographic chemicals or silver oxide batteries) must not be released into the environment and should be properly recycled. Silver should not be stored near ammonia, acetylene, or strong acids, as these can react with silver to make unstable silver salts. ==Market== More silver has been used each year than the amount mined in each year since 1990, when the amount mined each year stopped growing and its use increased. [[Companies]] that use silver have benefited from [[speculator]]s who sell promises to deliver silver that does not exist, keeping prices artificially low. This is called [[naked short selling]]. The amount owed is more than all the silver in the [[world]]. The [[price]] of silver could go very high when the stored silver runs out and investors start asking for their [[metal]] back, instead of taking more [[I.O.U.]]'s. The price of silver increased to 28 [[US dollar]]s per [[troy ounce]] as of December 2010. {{wiktionary}} == Other websites == * [https://en.citizendium.org/wiki/Silver Silver] -Citizendium {{Periodic Table}} {{authority control}} [[Category:Basic English 850 words]] [[Category:Transition metals]]
Silver is a chemical element; it has symbol Ag (from Latin argentum 'silver', derived from Proto-Indo-European *h₂erǵ ("shiny, white")) and atomic number 47. A soft, white, lustrous transition metal, it exhibits the highest electrical conductivity, thermal conductivity, and reflectivity of any metal. The metal is found in the Earth's crust in the pure, free elemental form ("native silver"), as an alloy with gold and other metals, and in minerals such as argentite and chlorargyrite. Most silver is produced as a byproduct of copper, gold, lead, and zinc refining. Silver has long been valued as a precious metal. Silver metal is used in many bullion coins, sometimes alongside gold: while it is more abundant than gold, it is much less abundant as a native metal. Its purity is typically measured on a per-mille basis; a 94%-pure alloy is described as "0.940 fine". As one of the seven metals of antiquity, silver has had an enduring role in most human cultures. Other than in currency and as an investment medium (coins and bullion), silver is used in solar panels, water filtration, jewellery, ornaments, high-value tableware and utensils (hence the term "silverware"), in electrical contacts and conductors, in specialized mirrors, window coatings, in catalysis of chemical reactions, as a colorant in stained glass, and in specialized confectionery. Its compounds are used in photographic and X-ray film. Dilute solutions of silver nitrate and other silver compounds are used as disinfectants and microbiocides (oligodynamic effect), added to bandages, wound-dressings, catheters, and other medical instruments. Silver is similar in its physical and chemical properties to its two vertical neighbours in group 11 of the periodic table: copper, and gold. Its 47 electrons are arranged in the configuration [Kr]4d5s, similarly to copper ([Ar]3d4s) and gold ([Xe]4f5d6s); group 11 is one of the few groups in the d-block which has a completely consistent set of electron configurations. This distinctive electron configuration, with a single electron in the highest occupied s subshell over a filled d subshell, accounts for many of the singular properties of metallic silver. Silver is a relatively soft and extremely ductile and malleable transition metal, though it is slightly less malleable than gold. Silver crystallizes in a face-centered cubic lattice with bulk coordination number 12, where only the single 5s electron is delocalized, similarly to copper and gold. Unlike metals with incomplete d-shells, metallic bonds in silver are lacking a covalent character and are relatively weak. This observation explains the low hardness and high ductility of single crystals of silver. Silver has a brilliant, white, metallic luster that can take a high polish, and which is so characteristic that the name of the metal itself has become a color name. Protected silver has greater optical reflectivity than aluminium at all wavelengths longer than ~450 nm. At wavelengths shorter than 450 nm, silver's reflectivity is inferior to that of aluminium and drops to zero near 310 nm. Very high electrical and thermal conductivity are common to the elements in group 11, because their single s electron is free and does not interact with the filled d subshell, as such interactions (which occur in the preceding transition metals) lower electron mobility. The thermal conductivity of silver is among the highest of all materials, although the thermal conductivity of carbon (in the diamond allotrope) and superfluid helium-4 are higher. The electrical conductivity of silver is the highest of all metals, greater even than copper. Silver also has the lowest contact resistance of any metal. Silver is rarely used for its electrical conductivity, due to its high cost, although an exception is in radio-frequency engineering, particularly at VHF and higher frequencies where silver plating improves electrical conductivity because those currents tend to flow on the surface of conductors rather than through the interior. During World War II in the US, 13540 tons of silver were used for the electromagnets in calutrons for enriching uranium, mainly because of the wartime shortage of copper. Silver readily forms alloys with copper, gold, and zinc. Zinc-silver alloys with low zinc concentration may be considered as face-centred cubic solid solutions of zinc in silver, as the structure of the silver is largely unchanged while the electron concentration rises as more zinc is added. Increasing the electron concentration further leads to body-centred cubic (electron concentration 1.5), complex cubic (1.615), and hexagonal close-packed phases (1.75). Naturally occurring silver is composed of two stable isotopes, Ag and Ag, with Ag being slightly more abundant (51.839% natural abundance). This almost equal abundance is rare in the periodic table. The atomic weight is 107.8682(2) u; this value is very important because of the importance of silver compounds, particularly halides, in gravimetric analysis. Both isotopes of silver are produced in stars via the s-process (slow neutron capture), as well as in supernovas via the r-process (rapid neutron capture). Twenty-eight radioisotopes have been characterized, the most stable being Ag with a half-life of 41.29 days, Ag with a half-life of 7.45 days, and Ag with a half-life of 3.13 hours. Silver has numerous nuclear isomers, the most stable being Ag (t1/2 = 418 years), Ag (t1/2 = 249.79 days) and Ag (t1/2 = 8.28 days). All of the remaining radioactive isotopes have half-lives of less than an hour, and the majority of these have half-lives of less than three minutes. Isotopes of silver range in relative atomic mass from 92.950 u (Ag) to 129.950 u (Ag); the primary decay mode before the most abundant stable isotope, Ag, is electron capture and the primary mode after is beta decay. The primary decay products before Ag are palladium (element 46) isotopes, and the primary products after are cadmium (element 48) isotopes. The palladium isotope Pd decays by beta emission to Ag with a half-life of 6.5 million years. Iron meteorites are the only objects with a high-enough palladium-to-silver ratio to yield measurable variations in Ag abundance. Radiogenic Ag was first discovered in the Santa Clara meteorite in 1978. Pd–Ag correlations observed in bodies that have clearly been melted since the accretion of the Solar System must reflect the presence of unstable nuclides in the early Solar System. Silver is a rather unreactive metal. This is because its filled 4d shell is not very effective in shielding the electrostatic forces of attraction from the nucleus to the outermost 5s electron, and hence silver is near the bottom of the electrochemical series (E(Ag/Ag) = +0.799 V). In group 11, silver has the lowest first ionization energy (showing the instability of the 5s orbital), but has higher second and third ionization energies than copper and gold (showing the stability of the 4d orbitals), so that the chemistry of silver is predominantly that of the +1 oxidation state, reflecting the increasingly limited range of oxidation states along the transition series as the d-orbitals fill and stabilize. Unlike copper, for which the larger hydration energy of Cu as compared to Cu is the reason why the former is the more stable in aqueous solution and solids despite lacking the stable filled d-subshell of the latter, with silver this effect is swamped by its larger second ionisation energy. Hence, Ag is the stable species in aqueous solution and solids, with Ag being much less stable as it oxidizes water. Most silver compounds have significant covalent character due to the small size and high first ionization energy (730.8 kJ/mol) of silver. Furthermore, silver's Pauling electronegativity of 1.93 is higher than that of lead (1.87), and its electron affinity of 125.6 kJ/mol is much higher than that of hydrogen (72.8 kJ/mol) and not much less than that of oxygen (141.0 kJ/mol). Due to its full d-subshell, silver in its main +1 oxidation state exhibits relatively few properties of the transition metals proper from groups 4 to 10, forming rather unstable organometallic compounds, forming linear complexes showing very low coordination numbers like 2, and forming an amphoteric oxide as well as Zintl phases like the post-transition metals. Unlike the preceding transition metals, the +1 oxidation state of silver is stable even in the absence of π-acceptor ligands. Silver does not react with air, even at red heat, and thus was considered by alchemists as a noble metal, along with gold. Its reactivity is intermediate between that of copper (which forms copper(I) oxide when heated in air to red heat) and gold. Like copper, silver reacts with sulfur and its compounds; in their presence, silver tarnishes in air to form the black silver sulfide (copper forms the green sulfate instead, while gold does not react). While silver is not attacked by non-oxidizing acids, the metal dissolves readily in hot concentrated sulfuric acid, as well as dilute or concentrated nitric acid. In the presence of air, and especially in the presence of hydrogen peroxide, silver dissolves readily in aqueous solutions of cyanide. The three main forms of deterioration in historical silver artifacts are tarnishing, formation of silver chloride due to long-term immersion in salt water, as well as reaction with nitrate ions or oxygen. Fresh silver chloride is pale yellow, becoming purplish on exposure to light; it projects slightly from the surface of the artifact or coin. The precipitation of copper in ancient silver can be used to date artifacts, as copper is nearly always a constituent of silver alloys. Silver metal is attacked by strong oxidizers such as potassium permanganate (KMnO4) and potassium dichromate (K2Cr2O7), and in the presence of potassium bromide (KBr). These compounds are used in photography to bleach silver images, converting them to silver bromide that can either be fixed with thiosulfate or redeveloped to intensify the original image. Silver forms cyanide complexes (silver cyanide) that are soluble in water in the presence of an excess of cyanide ions. Silver cyanide solutions are used in electroplating of silver. The common oxidation states of silver are (in order of commonness): +1 (the most stable state; for example, silver nitrate, AgNO3); +2 (highly oxidising; for example, silver(II) fluoride, AgF2); and even very rarely +3 (extreme oxidising; for example, potassium tetrafluoroargentate(III), KAgF4). The +3 state requires very strong oxidising agents to attain, such as fluorine or peroxodisulfate, and some silver(III) compounds react with atmospheric moisture and attack glass. Indeed, silver(III) fluoride is usually obtained by reacting silver or silver monofluoride with the strongest known oxidizing agent, krypton difluoride. Silver and gold have rather low chemical affinities for oxygen, lower than copper, and it is therefore expected that silver oxides are thermally quite unstable. Soluble silver(I) salts precipitate dark-brown silver(I) oxide, Ag2O, upon the addition of alkali. (The hydroxide AgOH exists only in solution; otherwise it spontaneously decomposes to the oxide.) Silver(I) oxide is very easily reduced to metallic silver, and decomposes to silver and oxygen above 160 °C. This and other silver(I) compounds may be oxidized by the strong oxidizing agent peroxodisulfate to black AgO, a mixed silver(I,III) oxide of formula AgAgO2. Some other mixed oxides with silver in non-integral oxidation states, namely Ag2O3 and Ag3O4, are also known, as is Ag3O which behaves as a metallic conductor. Silver(I) sulfide, Ag2S, is very readily formed from its constituent elements and is the cause of the black tarnish on some old silver objects. It may also be formed from the reaction of hydrogen sulfide with silver metal or aqueous Ag ions. Many non-stoichiometric selenides and tellurides are known; in particular, AgTe~3 is a low-temperature superconductor. The only known dihalide of silver is the difluoride, AgF2, which can be obtained from the elements under heat. A strong yet thermally stable and therefore safe fluorinating agent, silver(II) fluoride is often used to synthesize hydrofluorocarbons. In stark contrast to this, all four silver(I) halides are known. The fluoride, chloride, and bromide have the sodium chloride structure, but the iodide has three known stable forms at different temperatures; that at room temperature is the cubic zinc blende structure. They can all be obtained by the direct reaction of their respective elements. As the halogen group is descended, the silver halide gains more and more covalent character, solubility decreases, and the color changes from the white chloride to the yellow iodide as the energy required for ligand-metal charge transfer (XAg → XAg) decreases. The fluoride is anomalous, as the fluoride ion is so small that it has a considerable solvation energy and hence is highly water-soluble and forms di- and tetrahydrates. The other three silver halides are highly insoluble in aqueous solutions and are very commonly used in gravimetric analytical methods. All four are photosensitive (though the monofluoride is so only to ultraviolet light), especially the bromide and iodide which photodecompose to silver metal, and thus were used in traditional photography. The reaction involved is: The process is not reversible because the silver atom liberated is typically found at a crystal defect or an impurity site, so that the electron's energy is lowered enough that it is "trapped". White silver nitrate, AgNO3, is a versatile precursor to many other silver compounds, especially the halides, and is much less sensitive to light. It was once called lunar caustic because silver was called luna by the ancient alchemists, who believed that silver was associated with the Moon. It is often used for gravimetric analysis, exploiting the insolubility of the heavier silver halides which it is a common precursor to. Silver nitrate is used in many ways in organic synthesis, e.g. for deprotection and oxidations. Ag binds alkenes reversibly, and silver nitrate has been used to separate mixtures of alkenes by selective absorption. The resulting adduct can be decomposed with ammonia to release the free alkene. Yellow silver carbonate, Ag2CO3 can be easily prepared by reacting aqueous solutions of sodium carbonate with a deficiency of silver nitrate. Its principal use is for the production of silver powder for use in microelectronics. It is reduced with formaldehyde, producing silver free of alkali metals: Silver carbonate is also used as a reagent in organic synthesis such as the Koenigs-Knorr reaction. In the Fétizon oxidation, silver carbonate on celite acts as an oxidising agent to form lactones from diols. It is also employed to convert alkyl bromides into alcohols. Silver fulminate, AgCNO, a powerful, touch-sensitive explosive used in percussion caps, is made by reaction of silver metal with nitric acid in the presence of ethanol. Other dangerously explosive silver compounds are silver azide, AgN3, formed by reaction of silver nitrate with sodium azide, and silver acetylide, Ag2C2, formed when silver reacts with acetylene gas in ammonia solution. In its most characteristic reaction, silver azide decomposes explosively, releasing nitrogen gas: given the photosensitivity of silver salts, this behaviour may be induced by shining a light on its crystals. Silver complexes tend to be similar to those of its lighter homologue copper. Silver(III) complexes tend to be rare and very easily reduced to the more stable lower oxidation states, though they are slightly more stable than those of copper(III). For instance, the square planar periodate [Ag(IO5OH)2] and tellurate [Ag{TeO4(OH)2}2] complexes may be prepared by oxidising silver(I) with alkaline peroxodisulfate. The yellow diamagnetic [AgF4] is much less stable, fuming in moist air and reacting with glass. Silver(II) complexes are more common. Like the valence isoelectronic copper(II) complexes, they are usually square planar and paramagnetic, which is increased by the greater field splitting for 4d electrons than for 3d electrons. Aqueous Ag, produced by oxidation of Ag by ozone, is a very strong oxidising agent, even in acidic solutions: it is stabilized in phosphoric acid due to complex formation. Peroxodisulfate oxidation is generally necessary to give the more stable complexes with heterocyclic amines, such as [Ag(py)4] and [Ag(bipy)2]: these are stable provided the counterion cannot reduce the silver back to the +1 oxidation state. [AgF4] is also known in its violet barium salt, as are some silver(II) complexes with N- or O-donor ligands such as pyridine carboxylates. By far the most important oxidation state for silver in complexes is +1. The Ag cation is diamagnetic, like its homologues Cu and Au, as all three have closed-shell electron configurations with no unpaired electrons: its complexes are colourless provided the ligands are not too easily polarized such as I. Ag forms salts with most anions, but it is reluctant to coordinate to oxygen and thus most of these salts are insoluble in water: the exceptions are the nitrate, perchlorate, and fluoride. The tetracoordinate tetrahedral aqueous ion [Ag(H2O)4] is known, but the characteristic geometry for the Ag cation is 2-coordinate linear. For example, silver chloride dissolves readily in excess aqueous ammonia to form [Ag(NH3)2]; silver salts are dissolved in photography due to the formation of the thiosulfate complex [Ag(S2O3)2]; and cyanide extraction for silver (and gold) works by the formation of the complex [Ag(CN)2]. Silver cyanide forms the linear polymer {Ag–C≡N→Ag–C≡N→}; silver thiocyanate has a similar structure, but forms a zigzag instead because of the sp-hybridized sulfur atom. Chelating ligands are unable to form linear complexes and thus silver(I) complexes with them tend to form polymers; a few exceptions exist, such as the near-tetrahedral diphosphine and diarsine complexes [Ag(L–L)2]. Under standard conditions, silver does not form simple carbonyls, due to the weakness of the Ag–C bond. A few are known at very low temperatures around 6–15 K, such as the green, planar paramagnetic Ag(CO)3, which dimerizes at 25–30 K, probably by forming Ag–Ag bonds. Additionally, the silver carbonyl [Ag(CO)] [B(OTeF5)4] is known. Polymeric AgLX complexes with alkenes and alkynes are known, but their bonds are thermodynamically weaker than even those of the platinum complexes (though they are formed more readily than those of the analogous gold complexes): they are also quite unsymmetrical, showing the weak π bonding in group 11. Ag–C σ bonds may also be formed by silver(I), like copper(I) and gold(I), but the simple alkyls and aryls of silver(I) are even less stable than those of copper(I) (which tend to explode under ambient conditions). For example, poor thermal stability is reflected in the relative decomposition temperatures of AgMe (−50 °C) and CuMe (−15 °C) as well as those of PhAg (74 °C) and PhCu (100 °C). The C–Ag bond is stabilized by perfluoroalkyl ligands, for example in AgCF(CF3)2. Alkenylsilver compounds are also more stable than their alkylsilver counterparts. Silver-NHC complexes are easily prepared, and are commonly used to prepare other NHC complexes by displacing labile ligands. For example, the reaction of the bis(NHC)silver(I) complex with bis(acetonitrile)palladium dichloride or chlorido(dimethyl sulfide)gold(I): Silver forms alloys with most other elements on the periodic table. The elements from groups 1–3, except for hydrogen, lithium, and beryllium, are very miscible with silver in the condensed phase and form intermetallic compounds; those from groups 4–9 are only poorly miscible; the elements in groups 10–14 (except boron and carbon) have very complex Ag–M phase diagrams and form the most commercially important alloys; and the remaining elements on the periodic table have no consistency in their Ag–M phase diagrams. By far the most important such alloys are those with copper: most silver used for coinage and jewellery is in reality a silver–copper alloy, and the eutectic mixture is used in vacuum brazing. The two metals are completely miscible as liquids but not as solids; their importance in industry comes from the fact that their properties tend to be suitable over a wide range of variation in silver and copper concentration, although most useful alloys tend to be richer in silver than the eutectic mixture (71.9% silver and 28.1% copper by weight, and 60.1% silver and 28.1% copper by atom). Most other binary alloys are of little use: for example, silver–gold alloys are too soft and silver–cadmium alloys too toxic. Ternary alloys have much greater importance: dental amalgams are usually silver–tin–mercury alloys, silver–copper–gold alloys are very important in jewellery (usually on the gold-rich side) and have a vast range of hardnesses and colours, silver–copper–zinc alloys are useful as low-melting brazing alloys, and silver–cadmium–indium (involving three adjacent elements on the periodic table) is useful in nuclear reactors because of its high thermal neutron capture cross-section, good conduction of heat, mechanical stability, and resistance to corrosion in hot water. The word "silver" appears in Old English in various spellings, such as seolfor and siolfor. It is cognate with Old High German silabar; Gothic silubr; or Old Norse silfr, all ultimately deriving from Proto-Germanic *silubra. The Balto-Slavic words for silver are rather similar to the Germanic ones (e.g. Russian серебро [serebró], Polish srebro, Lithuanian sidãbras), as is the Celtiberian form silabur. They may have a common Indo-European origin, although their morphology rather suggest a non-Indo-European Wanderwort. Some scholars have thus proposed a Paleo-Hispanic origin, pointing to the Basque form zilharr as an evidence. The chemical symbol Ag is from the Latin word for "silver", argentum (compare Ancient Greek ἄργυρος, árgyros), from the Proto-Indo-European root *h₂erǵ- (formerly reconstructed as *arǵ-), meaning "white" or "shining". This was the usual Proto-Indo-European word for the metal, whose reflexes are missing in Germanic and Balto-Slavic. Silver was one of the seven metals of antiquity that were known to prehistoric humans and whose discovery is thus lost to history. In particular, the three metals of group 11, copper, silver, and gold, occur in the elemental form in nature and were probably used as the first primitive forms of money as opposed to simple bartering. However, unlike copper, silver did not lead to the growth of metallurgy on account of its low structural strength, and was more often used ornamentally or as money. Since silver is more reactive than gold, supplies of native silver were much more limited than those of gold. For example, silver was more expensive than gold in Egypt until around the fifteenth century BC: the Egyptians are thought to have separated gold from silver by heating the metals with salt, and then reducing the silver chloride produced to the metal. The situation changed with the discovery of cupellation, a technique that allowed silver metal to be extracted from its ores. While slag heaps found in Asia Minor and on the islands of the Aegean Sea indicate that silver was being separated from lead as early as the 4th millennium BC, and one of the earliest silver extraction centres in Europe was Sardinia in the early Chalcolithic period, these techniques did not spread widely until later, when it spread throughout the region and beyond. The origins of silver production in India, China, and Japan were almost certainly equally ancient, but are not well-documented due to their great age. When the Phoenicians first came to what is now Spain, they obtained so much silver that they could not fit it all on their ships, and as a result used silver to weight their anchors instead of lead. By the time of the Greek and Roman civilizations, silver coins were a staple of the economy: the Greeks were already extracting silver from galena by the 7th century BC, and the rise of Athens was partly made possible by the nearby silver mines at Laurium, from which they extracted about 30 tonnes a year from 600 to 300 BC. The stability of the Roman currency relied to a high degree on the supply of silver bullion, mostly from Spain, which Roman miners produced on a scale unparalleled before the discovery of the New World. Reaching a peak production of 200 tonnes per year, an estimated silver stock of 10,000 tonnes circulated in the Roman economy in the middle of the second century AD, five to ten times larger than the combined amount of silver available to medieval Europe and the Abbasid Caliphate around AD 800. The Romans also recorded the extraction of silver in central and northern Europe in the same time period. This production came to a nearly complete halt with the fall of the Roman Empire, not to resume until the time of Charlemagne: by then, tens of thousands of tonnes of silver had already been extracted. Central Europe became the centre of silver production during the Middle Ages, as the Mediterranean deposits exploited by the ancient civilisations had been exhausted. Silver mines were opened in Bohemia, Saxony, Alsace, the Lahn region, Siegerland, Silesia, Hungary, Norway, Steiermark, Schwaz, and the southern Black Forest. Most of these ores were quite rich in silver and could simply be separated by hand from the remaining rock and then smelted; some deposits of native silver were also encountered. Many of these mines were soon exhausted, but a few of them remained active until the Industrial Revolution, before which the world production of silver was around a meagre 50 tonnes per year. In the Americas, high temperature silver-lead cupellation technology was developed by pre-Inca civilizations as early as AD 60–120; silver deposits in India, China, Japan, and pre-Columbian America continued to be mined during this time. With the discovery of America and the plundering of silver by the Spanish conquistadors, Central and South America became the dominant producers of silver until around the beginning of the 18th century, particularly Peru, Bolivia, Chile, and Argentina: the last of these countries later took its name from that of the metal that composed so much of its mineral wealth. The silver trade gave way to a global network of exchange. As one historian put it, silver "went round the world and made the world go round." Much of this silver ended up in the hands of the Chinese. A Portuguese merchant in 1621 noted that silver "wanders throughout all the world... before flocking to China, where it remains as if at its natural center." Still, much of it went to Spain, allowing Spanish rulers to pursue military and political ambitions in both Europe and the Americas. "New World mines", concluded several historians, "supported the Spanish empire." In the 19th century, primary production of silver moved to North America, particularly Canada, Mexico, and Nevada in the United States: some secondary production from lead and zinc ores also took place in Europe, and deposits in Siberia and the Russian Far East as well as in Australia were mined. Poland emerged as an important producer during the 1970s after the discovery of copper deposits that were rich in silver, before the centre of production returned to the Americas the following decade. Today, Peru and Mexico are still among the primary silver producers, but the distribution of silver production around the world is quite balanced and about one-fifth of the silver supply comes from recycling instead of new production. Silver plays a certain role in mythology and has found various usage as a metaphor and in folklore. The Greek poet Hesiod's Works and Days (lines 109–201) lists different ages of man named after metals like gold, silver, bronze and iron to account for successive ages of humanity. Ovid's Metamorphoses contains another retelling of the story, containing an illustration of silver's metaphorical use of signifying the second-best in a series, better than bronze but worse than gold: But when good Saturn, banish'd from above, Was driv'n to Hell, the world was under Jove. Succeeding times a silver age behold, Excelling brass, but more excell'd by gold. In folklore, silver was commonly thought to have mystic powers: for example, a bullet cast from silver is often supposed in such folklore the only weapon that is effective against a werewolf, witch, or other monsters. From this the idiom of a silver bullet developed into figuratively referring to any simple solution with very high effectiveness or almost miraculous results, as in the widely discussed software engineering paper "No Silver Bullet." Other powers attributed to silver include detection of poison and facilitation of passage into the mythical realm of fairies. Silver production has also inspired figurative language. Clear references to cupellation occur throughout the Old Testament of the Bible, such as in Jeremiah's rebuke to Judah: "The bellows are burned, the lead is consumed of the fire; the founder melteth in vain: for the wicked are not plucked away. Reprobate silver shall men call them, because the Lord hath rejected them." (Jeremiah 6:19–20) Jeremiah was also aware of sheet silver, exemplifying the malleability and ductility of the metal: "Silver spread into plates is brought from Tarshish, and gold from Uphaz, the work of the workman, and of the hands of the founder: blue and purple is their clothing: they are all the work of cunning men." (Jeremiah 10:9) Silver also has more negative cultural meanings: the idiom thirty pieces of silver, referring to a reward for betrayal, references the bribe Judas Iscariot is said in the New Testament to have taken from Jewish leaders in Jerusalem to turn Jesus of Nazareth over to soldiers of the high priest Caiaphas. Ethically, silver also symbolizes greed and degradation of consciousness; this is the negative aspect, the perverting of its value. The abundance of silver in the Earth's crust is 0.08 parts per million, almost exactly the same as that of mercury. It mostly occurs in sulfide ores, especially acanthite and argentite, Ag2S. Argentite deposits sometimes also contain native silver when they occur in reducing environments, and when in contact with salt water they are converted to chlorargyrite (including horn silver), AgCl, which is prevalent in Chile and New South Wales. Most other silver minerals are silver pnictides or chalcogenides; they are generally lustrous semiconductors. Most true silver deposits, as opposed to argentiferous deposits of other metals, came from Tertiary period vulcanism. The principal sources of silver are the ores of copper, copper-nickel, lead, and lead-zinc obtained from Peru, Bolivia, Mexico, China, Australia, Chile, Poland and Serbia. Peru, Bolivia and Mexico have been mining silver since 1546, and are still major world producers. Top silver-producing mines are Cannington (Australia), Fresnillo (Mexico), San Cristóbal (Bolivia), Antamina (Peru), Rudna (Poland), and Penasquito (Mexico). Top near-term mine development projects through 2015 are Pascua Lama (Chile), Navidad (Argentina), Jaunicipio (Mexico), Malku Khota (Bolivia), and Hackett River (Canada). In Central Asia, Tajikistan is known to have some of the largest silver deposits in the world. Silver is usually found in nature combined with other metals, or in minerals that contain silver compounds, generally in the form of sulfides such as galena (lead sulfide) or cerussite (lead carbonate). So the primary production of silver requires the smelting and then cupellation of argentiferous lead ores, a historically important process. Lead melts at 327 °C, lead oxide at 888 °C and silver melts at 960 °C. To separate the silver, the alloy is melted again at the high temperature of 960 °C to 1000 °C in an oxidizing environment. The lead oxidises to lead monoxide, then known as litharge, which captures the oxygen from the other metals present. The liquid lead oxide is removed or absorbed by capillary action into the hearth linings. Today, silver metal is primarily produced instead as a secondary byproduct of electrolytic refining of copper, lead, and zinc, and by application of the Parkes process on lead bullion from ore that also contains silver. In such processes, silver follows the non-ferrous metal in question through its concentration and smelting, and is later purified out. For example, in copper production, purified copper is electrolytically deposited on the cathode, while the less reactive precious metals such as silver and gold collect under the anode as the so-called "anode slime". This is then separated and purified of base metals by treatment with hot aerated dilute sulfuric acid and heating with lime or silica flux, before the silver is purified to over 99.9% purity via electrolysis in nitrate solution. Commercial-grade fine silver is at least 99.9% pure, and purities greater than 99.999% are available. In 2022, Mexico was the top producer of silver (6,300 tonnes or 24.2% of the world's total of 26,000 t), followed by China (3,600 t) and Peru (3,100 t). Silver concentration is low in seawater (pmol/L). Levels vary by depth and between water bodies. Dissolved silver concentrations range from 0.3 pmol/L in coastal surface waters to 22.8 pmol/L in pelagic deep waters. Analyzing the presence and dynamics of silver in marine environments is difficult due to these particularly low concentrations and complex interactions in the environment. Although a rare trace metal, concentrations are greatly impacted by fluvial, aeolian, atmospheric, and upwelling inputs, as well as anthropogenic inputs via discharge, waste disposal, and emissions from industrial companies. Other internal processes such as decomposition of organic matter may be a source of dissolved silver in deeper waters, which feeds into some surface waters through upwelling and vertical mixing. In the Atlantic and Pacific, silver concentrations are minimal at the surface but rise in deeper waters. Silver is taken up by plankton in the photic zone, remobilized with depth, and enriched in deep waters. Silver is transported from the Atlantic to the other oceanic water masses. In North Pacific waters, silver is remobilized at a slower rate and increasingly enriched compared to deep Atlantic waters. Silver has increasing concentrations that follow the major oceanic conveyor belt that cycles water and nutrients from the North Atlantic to the South Atlantic to the North Pacific. There is not an extensive amount of data focused on how marine life is affected by silver despite the likely deleterious effects it could have on organisms through bioaccumulation, association with particulate matters, and sorption. Not until about 1984 did scientists begin to understand the chemical characteristics of silver and the potential toxicity. In fact, mercury is the only other trace metal that surpasses the toxic effects of silver; however, the full extent of silver toxicity is not expected in oceanic conditions because of its ability to transfer into nonreactive biological compounds. In one study, the presence of excess ionic silver and silver nanoparticles caused bioaccumulation effects on zebrafish organs and altered the chemical pathways within their gills. In addition, very early experimental studies demonstrated how the toxic effects of silver fluctuate with salinity and other parameters, as well as between life stages and different species such as finfish, molluscs, and crustaceans. Another study found raised concentrations of silver in the muscles and liver of dolphins and whales, indicating pollution of this metal within recent decades. Silver is not an easy metal for an organism to eliminate and elevated concentrations can cause death. The earliest known coins were minted in the kingdom of Lydia in Asia Minor around 600 BC. The coins of Lydia were made of electrum, which is a naturally occurring alloy of gold and silver, that was available within the territory of Lydia. Since that time, silver standards, in which the standard economic unit of account is a fixed weight of silver, have been widespread throughout the world until the 20th century. Notable silver coins through the centuries include the Greek drachma, the Roman denarius, the Islamic dirham, the karshapana from ancient India and rupee from the time of the Mughal Empire (grouped with copper and gold coins to create a trimetallic standard), and the Spanish dollar. The ratio between the amount of silver used for coinage and that used for other purposes has fluctuated greatly over time; for example, in wartime, more silver tends to have been used for coinage to finance the war. Today, silver bullion has the ISO 4217 currency code XAG, one of only four precious metals to have one (the others being palladium, platinum, and gold). Silver coins are produced from cast rods or ingots, rolled to the correct thickness, heat-treated, and then used to cut blanks from. These blanks are then milled and minted in a coining press; modern coining presses can produce 8000 silver coins per hour. Silver prices are normally quoted in troy ounces. One troy ounce is equal to 31.1034768 grams. The London silver fix is published every working day at noon London time. This price is determined by several major international banks and is used by London bullion market members for trading that day. Prices are most commonly shown as the United States dollar (USD), the Pound sterling (GBP), and the Euro (EUR). The major use of silver besides coinage throughout most of history was in the manufacture of jewellery and other general-use items, and this continues to be a major use today. Examples include table silver for cutlery, for which silver is highly suited due to its antibacterial properties. Western concert flutes are usually plated with or made out of sterling silver; in fact, most silverware is only silver-plated rather than made out of pure silver; the silver is normally put in place by electroplating. Silver-plated glass (as opposed to metal) is used for mirrors, vacuum flasks, and Christmas tree decorations. Because pure silver is very soft, most silver used for these purposes is alloyed with copper, with finenesses of 925/1000, 835/1000, and 800/1000 being common. One drawback is the easy tarnishing of silver in the presence of hydrogen sulfide and its derivatives. Including precious metals such as palladium, platinum, and gold gives resistance to tarnishing but is quite costly; base metals like zinc, cadmium, silicon, and germanium do not totally prevent corrosion and tend to affect the lustre and colour of the alloy. Electrolytically refined pure silver plating is effective at increasing resistance to tarnishing. The usual solutions for restoring the lustre of tarnished silver are dipping baths that reduce the silver sulfide surface to metallic silver, and cleaning off the layer of tarnish with a paste; the latter approach also has the welcome side effect of polishing the silver concurrently. In medicine, silver is incorporated into wound dressings and used as an antibiotic coating in medical devices. Wound dressings containing silver sulfadiazine or silver nanomaterials are used to treat external infections. Silver is also used in some medical applications, such as urinary catheters (where tentative evidence indicates it reduces catheter-related urinary tract infections) and in endotracheal breathing tubes (where evidence suggests it reduces ventilator-associated pneumonia). The silver ion is bioactive and in sufficient concentration readily kills bacteria in vitro. Silver ions interfere with enzymes in the bacteria that transport nutrients, form structures, and synthesise cell walls; these ions also bond with the bacteria's genetic material. Silver and silver nanoparticles are used as an antimicrobial in a variety of industrial, healthcare, and domestic application: for example, infusing clothing with nanosilver particles thus allows them to stay odourless for longer. Bacteria can, however, develop resistance to the antimicrobial action of silver. Silver compounds are taken up by the body like mercury compounds, but lack the toxicity of the latter. Silver and its alloys are used in cranial surgery to replace bone, and silver–tin–mercury amalgams are used in dentistry. Silver diammine fluoride, the fluoride salt of a coordination complex with the formula [Ag(NH3)2]F, is a topical medicament (drug) used to treat and prevent dental caries (cavities) and relieve dentinal hypersensitivity. Silver is very important in electronics for conductors and electrodes on account of its high electrical conductivity even when tarnished. Bulk silver and silver foils were used to make vacuum tubes, and continue to be used today in the manufacture of semiconductor devices, circuits, and their components. For example, silver is used in high quality connectors for RF, VHF, and higher frequencies, particularly in tuned circuits such as cavity filters where conductors cannot be scaled by more than 6%. Printed circuits and RFID antennas are made with silver paints, Powdered silver and its alloys are used in paste preparations for conductor layers and electrodes, ceramic capacitors, and other ceramic components. Silver-containing brazing alloys are used for brazing metallic materials, mostly cobalt, nickel, and copper-based alloys, tool steels, and precious metals. The basic components are silver and copper, with other elements selected according to the specific application desired: examples include zinc, tin, cadmium, palladium, manganese, and phosphorus. Silver provides increased workability and corrosion resistance during usage. Silver is useful in the manufacture of chemical equipment on account of its low chemical reactivity, high thermal conductivity, and being easily workable. Silver crucibles (alloyed with 0.15% nickel to avoid recrystallisation of the metal at red heat) are used for carrying out alkaline fusion. Copper and silver are also used when doing chemistry with fluorine. Equipment made to work at high temperatures is often silver-plated. Silver and its alloys with gold are used as wire or ring seals for oxygen compressors and vacuum equipment. Silver metal is a good catalyst for oxidation reactions; in fact it is somewhat too good for most purposes, as finely divided silver tends to result in complete oxidation of organic substances to carbon dioxide and water, and hence coarser-grained silver tends to be used instead. For instance, 15% silver supported on α-Al2O3 or silicates is a catalyst for the oxidation of ethylene to ethylene oxide at 230–270 °C. Dehydrogenation of methanol to formaldehyde is conducted at 600–720 °C over silver gauze or crystals as the catalyst, as is dehydrogenation of isopropanol to acetone. In the gas phase, glycol yields glyoxal and ethanol yields acetaldehyde, while organic amines are dehydrated to nitriles. Before the advent of digital photography, which is now dominant, the photosensitivity of silver halides was exploited for use in traditional film photography. The photosensitive emulsion used in black-and-white photography is a suspension of silver halide crystals in gelatin, possibly mixed in with some noble metal compounds for improved photosensitivity, developing, and tuning. Colour photography requires the addition of special dye components and sensitisers, so that the initial black-and-white silver image couples with a different dye component. The original silver images are bleached off and the silver is then recovered and recycled. Silver nitrate is the starting material in all cases. The market for silver nitrate and silver halides for photography has rapidly declined with the rise of digital cameras. From the peak global demand for photographic silver in 1999 (267,000,000 troy ounces or 8,304.6 tonnes) the market contracted almost 70% by 2013. Nanosilver particles, between 10 and 100 nanometres in size, are used in many applications. They are used in conductive inks for printed electronics, and have a much lower melting point than larger silver particles of micrometre size. They are also used medicinally in antibacterials and antifungals in much the same way as larger silver particles. In addition, according to the European Union Observatory for Nanomaterials (EUON), silver nanoparticles are used both in pigments, as well as cosmetics. Pure silver metal is used as a food colouring. It has the E174 designation and is approved in the European Union. Traditional Indian and Pakistani dishes sometimes include decorative silver foil known as vark, and in various other cultures, silver dragée are used to decorate cakes, cookies, and other dessert items. Photochromic lenses include silver halides, so that ultraviolet light in natural daylight liberates metallic silver, darkening the lenses. The silver halides are reformed in lower light intensities. Colourless silver chloride films are used in radiation detectors. Zeolite sieves incorporating Ag ions are used to desalinate seawater during rescues, using silver ions to precipitate chloride as silver chloride. Silver is also used for its antibacterial properties for water sanitisation, but the application of this is limited by limits on silver consumption. Colloidal silver is similarly used to disinfect closed swimming pools; while it has the advantage of not giving off a smell like hypochlorite treatments do, colloidal silver is not effective enough for more contaminated open swimming pools. Small silver iodide crystals are used in cloud seeding to cause rain. The Texas Legislature designated silver the official precious metal of Texas in 2007. Silver compounds have low toxicity compared to those of most other heavy metals, as they are poorly absorbed by the human body when ingested, and that which does get absorbed is rapidly converted to insoluble silver compounds or complexed by metallothionein. However, silver fluoride and silver nitrate are caustic and can cause tissue damage, resulting in gastroenteritis, diarrhoea, falling blood pressure, cramps, paralysis, and respiratory arrest. Animals repeatedly dosed with silver salts have been observed to experience anaemia, slowed growth, necrosis of the liver, and fatty degeneration of the liver and kidneys; rats implanted with silver foil or injected with colloidal silver have been observed to develop localised tumours. Parenterally admistered colloidal silver causes acute silver poisoning. Some waterborne species are particularly sensitive to silver salts and those of the other precious metals; in most situations, however, silver does not pose serious environmental hazards. In large doses, silver and compounds containing it can be absorbed into the circulatory system and become deposited in various body tissues, leading to argyria, which results in a blue-grayish pigmentation of the skin, eyes, and mucous membranes. Argyria is rare, and so far as is known, does not otherwise harm a person's health, though it is disfiguring and usually permanent. Mild forms of argyria are sometimes mistaken for cyanosis, a blue tint on skin, caused by lack of oxygen. Metallic silver, like copper, is an antibacterial agent, which was known to the ancients and first scientifically investigated and named the oligodynamic effect by Carl Nägeli. Silver ions damage the metabolism of bacteria even at such low concentrations as 0.01–0.1 milligrams per litre; metallic silver has a similar effect due to the formation of silver oxide. This effect is lost in the presence of sulfur due to the extreme insolubility of silver sulfide. Some silver compounds are very explosive, such as the nitrogen compounds silver azide, silver amide, and silver fulminate, as well as silver acetylide, silver oxalate, and silver(II) oxide. They can explode on heating, force, drying, illumination, or sometimes spontaneously. To avoid the formation of such compounds, ammonia and acetylene should be kept away from silver equipment. Salts of silver with strongly oxidising acids such as silver chlorate and silver nitrate can explode on contact with materials that can be readily oxidised, such as organic compounds, sulfur and soot.
:''This article is about Popes in general. For the current Pope, see [[Pope Francis]].'' [[File:Franciscus, 2023 (cropped).jpg|thumb|right|[[Pope Francis|Francis]], the current pope since 2013]] [[File:Chiesa di San Silvestro - San Silvestro statue.jpg|thumb|Statue of [[Pope Sylvester I]] before a church in [[Pisa]]. He was the bishop of Rome 314–335]] The '''Pope''' is the head of the [[Roman Catholic Church]] in religious contexts, and politically of the [[Papal States]] and later [[Vatican City|Vatican City State]].<ref name=Reese10/><ref name="catholic">[http://www.newadvent.org/cathen/12260a.htm "Pope,"] 'Catholic Encyclopedia'' (2009); retrieved 2013-4-1.</ref> His official [[:wikt:title|title]] is the '''Bishop of Rome'''.<ref name=Reese10>Thomas J. Reese, ''Inside the Vatican: The Politics and Organization of the Catholic Church'' (Cambridge, MA: Harvard University Press, 1996), p. 10</ref> The current Pope is [[Pope Francis]].<ref>[http://www.newadvent.org/cathen/12272b.htm "List of Popes,"] 'Catholic Encyclopedia''; retrieved 2013-4-1.</ref> Popes are elected by [[Cardinal (Catholicism)|Cardinals]] of the Catholic Church. Once they are elected, they hold the position until they die or resign. The Pope can’t be an organ donor.<ref>{{Cite web|last=Herald|first=The Catholic|date=2011-02-05|title=Pope cannot be organ donor, Vatican official says|url=https://catholicherald.org/news/nation-and-world/pope-cannot-be-organ-donor-vatican-official-says/|access-date=2024-09-07|website=Catholic Herald|language=en-US}}</ref> Usually they do not resign, though; [[Pope Benedict XVI]] is the only Pope to resign in the last 500 years. A newly elected Pope chooses a [[regnal name]]. Everyone is told this new name when the [[Habemus Papam]] is read out. The current pope (Francis) was called [[Jorge Bergoglio]] before he became a pope. The name ''Pope'' comes from the [[Greek language|Greek]] word ''pappas'', meaning "father".<ref name="catholic"/> Catholic doctrine holds that when making statements ''ex cathedra'', that is official statements teaching about faith and morals, the Pope is ''infallible'' - which means [[God]] will not allow his followers to be misled by allowing their leader to make a wrong statement. Only two of any Pope's statements have been ''ex cathedra''.<ref>Apostolic constitution ''Munificentissimus Deus'', written by Pope Pius XII, 1 November 1950 {{cite web|url=http://www.vatican.va/holy_father/pius_xii/apost_constitutions/documents/hf_p-xii_apc_19501101_munificentissimus-deus_en.html|title=Munificentissimus Deus}}; Apostolic constitution ''Ineffabilis Deus'', written by Pope Pius IX, 8 December 1854.</ref> Popes today travel to many countries around the world preaching as well as one of the two ways to represent the [[Holy See]], the other being through the presence of an [[Apostolic Nunciature|apostolic nuncio]] on some Catholic-related events serving as the Pope's representative or spokesperson to a country. The Pope and the Bishop of Urgell of [[Andorra]] are the only Catholic religious figures in the world who both leads the church and [[government]]. Like the bishops, archbishops, and cardinals of the Catholic Church, he wears a big hat called a [[mitre]] and holds a staff called a [[crosier]]. As head of the [[Holy See]] having governance over the Catholic Church and Vatican City, the pope is protected by Swiss Guards and Vatican Gendarmerie every time he made public appearances within Vatican and around the world to protect him from assassinations and other events that are against the Holy See, making him the only Catholic religious figure with security guards for protection. == Recent popes == Some recent Popes, and the time they were Pope: * [[Pope John XXIII]], 1958-1963 * [[Pope Paul VI]], 1963-1978 * [[Pope John Paul I]], 1978 * [[Pope John Paul II]], 1978-2005 * [[Pope Benedict XVI]], 2005-2013<ref>Lizzy Davies. [http://www.guardian.co.uk/world/2013/feb/11/pope-benedict-xvi-resigns-age "Pope Benedict XVI resigns,"] ''Guardian'' UK). 11 February 2013; retrieved 2013-2-11.</ref> * [[Pope Francis]], current pope == The Popes in Avignon == During parts of the Middle Ages, the French kings had a lot of influence in Europe. For this reason, seven popes (and two anti-popes) lived in [[Avignon]], rather than Rome. The [[Avignon Papacy]] was from 1309 to 1377. During that time, the popes were known for their greed and [[corruption]].<ref name="R:RCC">[[Will Durant|Durant, Will]]. The Reformation. New York: Simon and Schuster. 1957. "Chapter I. The Roman Catholic Church." 1300-1517. p. 3-25</ref> These popes were [[ally|allies]] of France; the enemies of France were also their enemies.<ref name="R:EWCGR">[[Will Durant|Durant, Will]]. The Reformation. New York: Simon and Schuster. 1957. "Chapter II. England: Wyclif, Chaucer, and the Great Revolt." 1308-1400. p. 26-57</ref> The Bishops of Rome who lived in Avignon were: #[[Pope Clement V]]: 1305–1314 #[[Pope John XXII]]: 1316–1334 #[[Pope Benedict XII]]: 1334–1342 #[[Pope Clement VI]]: 1342–1352 #[[Pope Innocent VI]]: 1352–1362 #[[Pope Urban V]]: 1362–1370 #[[Pope Gregory XI]]: 1370–1378 Two antipopes were based in Avignon as well: *[[Antipope Clement VII|Clement VII]]: 1378–1394 *[[Antipope Benedict XIII|Benedict XIII]]: 1394–1423 (expelled from Avignon in 1403) [[Antipope]]s were people that were elected by small groups who did not like the official choice. [[Catherine of Siena]] convinced pope Gregory XI to move back to Rome. Unfortunately, he died shortly after moving. The cardinals then elected [[Pope Urban VI|Urban VI]] to be the next pope. The French cardinals did not recognise this election as legitimate. They declared the papal see as vacant; which led to the [[Western Schism]]. The schism lasted until the [[Council of Constance]] in 1417. During this time, there was a pope in Rome, an Antipope in Avignon, and for some time, a second antipope. Each of the three was recognised as legitimate pope by different European powers. This led to a big split in the church as a whole. The council elected [[Pope Martin V]] as a new pope, recognised by all parties. ==Related pages== * [[List of popes]] * [[Abdication]] * ''[[Habemus Papam]]'' * [[Papal name]] == References == [[File:Emblem Holy See.svg|thumb|120px|right|[[wikt:emblem|Emblem]] of the [[popes]] ]] {{reflist}} ==Other websites== {{commonscat-inline|Popes}} {{Popes}} {{Catholic Church ranks}} {{authority control}} [[Category:Popes| ]] [[Category:Heads of state]] [[Category:Christian religious occupations]]
The pope (Latin: papa, from Greek: πάππας, romanized: páppas, lit. 'father'), also known as the supreme pontiff (pontifex maximus or summus pontifex), Roman pontiff (Romanus pontifex) or sovereign pontiff, is the bishop of Rome (or historically the patriarch of Rome), head of the worldwide Catholic Church, and has also served as the head of state or sovereign of the Papal States and later the Vatican City State since the eighth century. From a Catholic viewpoint, the primacy of the bishop of Rome is largely derived from his role as the apostolic successor to Saint Peter, to whom primacy was conferred by Jesus, who gave Peter the Keys of Heaven and the powers of "binding and loosing", naming him as the "rock" upon which the Church would be built. The current pope is Francis, who was elected on 13 March 2013. While his office is called the papacy, the jurisdiction of the episcopal see is called the Holy See. It is the Holy See that is the sovereign entity by international law headquartered in the distinctively independent Vatican City State, a city-state which forms a geographical enclave within the conurbation of Rome, established by the Lateran Treaty in 1929 between Italy and the Holy See to ensure its temporal and spiritual independence. The Holy See is recognized by its adherence at various levels to international organizations and by means of its diplomatic relations and political accords with many independent states. According to Catholic tradition, the apostolic see of Rome was founded by Saint Peter and Saint Paul in the first century. The papacy is one of the most enduring institutions in the world and has had a prominent part in human history. In ancient times, the popes helped spread Christianity and intervened to find resolutions in various doctrinal disputes. In the Middle Ages, they played a role of secular importance in Western Europe, often acting as arbitrators between Christian monarchs. In addition to the expansion of Christian faith and doctrine, modern popes are involved in ecumenism and interfaith dialogue, charitable work, and the defense of human rights. Over time, the papacy accrued broad secular and political influence, eventually rivaling those of territorial rulers. In recent centuries, the temporal authority of the papacy has declined and the office is now largely focused on religious matters. By contrast, papal claims of spiritual authority have been increasingly firmly expressed over time, culminating in 1870 with the proclamation of the dogma of papal infallibility for rare occasions when the pope speaks ex cathedra—literally "from the chair (of Saint Peter)"—to issue a formal definition of faith or morals. The pope is considered one of the world's most powerful people due to the extensive diplomatic, cultural, and spiritual influence of his position on both 1.3 billion Catholics and those outside the Catholic faith, and because he heads the world's largest non-government provider of education and health care, with a vast network of charities. The word pope derives from Greek πάππας ('páppas'), meaning 'father'. In the early centuries of Christianity, this title was applied, especially in the East, to all bishops and other senior clergy, and later became reserved in the West to the bishop of Rome during the reign of Pope Leo I (440–461), a reservation made official only in the 11th century. The earliest record of the use of the title of 'pope' was in regard to the by-then-deceased patriarch of Alexandria, Heraclas (232–248). The earliest recorded use of the title "pope" in English dates to the mid-10th century, when it was used in reference to the 7th century Roman Pope Vitalian in an Old English translation of Bede's Historia ecclesiastica gentis Anglorum. The Catholic Church teaches that the pastoral office, the office of shepherding the Church, that was held by the apostles, as a group or "college" with Saint Peter as their head, is now held by their successors, the bishops, with the bishop of Rome (the pope) as their head. Thus is derived another title by which the pope is known, that of "supreme pontiff". The Catholic Church teaches that Jesus personally appointed Peter as the visible head of the Church, and the Catholic Church's dogmatic constitution Lumen gentium makes a clear distinction between apostles and bishops, presenting the latter as the successors of the former, with the pope as successor of Peter, in that he is head of the bishops as Peter was head of the apostles. Some historians argue against the notion that Peter was the first bishop of Rome, noting that the episcopal see in Rome can be traced back no earlier than the 3rd century. The writings of the Church Father Irenaeus, who wrote around 180 AD, reflect a belief that Peter "founded and organized" the Church at Rome. Moreover, Irenaeus was not the first to write of Peter's presence in the early Roman Church. The Church of Rome wrote in a letter to the Corinthians (which is traditionally attributed to Clement of Rome c. 96) about the persecution of Christians in Rome as the "struggles in our time" and presented to the Corinthians its heroes, "first, the greatest and most just columns", the "good apostles" Peter and Paul. St. Ignatius of Antioch wrote shortly after Clement; in his letter from the city of Smyrna to the Romans, he said he would not command them as Peter and Paul did. Given this and other evidence, such as Emperor Constantine's erection of the "Old St. Peter's Basilica" on the location of St. Peter's tomb, as held and given to him by Rome's Christian community, many scholars agree that Peter was martyred in Rome under Nero, although some scholars argue that he may have been martyred in Palestine. Though open to historical debate, first-century Christian communities may have had a group of presbyter-bishops functioning as guides of their local churches. Gradually, episcopal sees were established in metropolitan areas. Antioch may have developed such a structure before Rome. In Rome, there were over time at various junctures rival claimants to be the rightful bishop, though again Irenaeus stressed the validity of one line of bishops from the time of St. Peter up to his contemporary Pope Victor I and listed them. Some writers claim that the emergence of a single bishop in Rome probably did not occur until the middle of the 2nd century. In their view, Linus, Cletus and Clement were possibly prominent presbyter-bishops, but not necessarily monarchical bishops. Documents of the 1st century and early second century indicate that the bishop of Rome had some kind of pre-eminence and prominence in the Church as a whole, as even a letter from the bishop, or patriarch, of Antioch acknowledged the bishop of Rome as "a first among equals", though the detail of what this meant is unclear. Sources suggest that at first, the terms 'episcopos' and 'presbyter' were used interchangeably, with the consensus among scholars being that by the turn of the 1st and 2nd centuries, local congregations were led by bishops and presbyters, whose duties of office overlapped or were indistinguishable from one another. Some say that there was probably "no single 'monarchical' bishop in Rome before the middle of the 2nd century...and likely later." In the early Christian era, Rome and a few other cities had claims on the leadership of worldwide Church. James the Just, known as "the brother of the Lord", served as head of the Jerusalem church, which is still honored as the "Mother Church" in Orthodox tradition. Alexandria had been a center of Jewish learning and became a center of Christian learning. Rome had a large congregation early in the apostolic period whom Paul the Apostle addressed in his Epistle to the Romans, and according to tradition Paul was martyred there. During the 1st century of the Church (c. 30–130), the Roman capital became recognized as a Christian center of exceptional importance. The church there, at the end of the century, wrote an epistle to the Church in Corinth intervening in a major dispute, and apologizing for not having taken action earlier. There are a few other references of that time to recognition of the authoritative primacy of the Roman See outside of Rome. In the Ravenna Document of 13 October 2007, theologians chosen by the Catholic and the Eastern Orthodox Churches stated: "41. Both sides agree ... that Rome, as the Church that 'presides in love' according to the phrase of St Ignatius of Antioch, occupied the first place in the taxis, and that the bishop of Rome was therefore the protos among the patriarchs. Translated into English, the statement means "first among equals". What form that should take is still a matter of disagreement, just as it was when the Catholic and Orthodox Churches split in the Great East-West Schism. They also disagree on the interpretation of the historical evidence from this era regarding the prerogatives of the bishop of Rome as protos, a matter that was already understood in different ways in the first millennium." In AD 195, Pope Victor I, in what is seen as an exercise of Roman authority over other churches, excommunicated the Quartodecimans for observing Easter on the 14th of Nisan, the date of the Jewish Passover, a tradition handed down by John the Evangelist (see Easter controversy). Celebration of Easter on a Sunday, as insisted on by the pope, is the system that has prevailed (see computus). The Edict of Milan in 313 granted freedom to all religions in the Roman Empire, beginning the Peace of the Church. In 325, the First Council of Nicaea condemned Arianism, declaring trinitarianism dogmatic, and in its sixth canon recognized the special role of the Sees of Rome, Alexandria, and Antioch. Great defenders of Trinitarian faith included the popes, especially Liberius, who was exiled to Berea by Constantius II for his Trinitarian faith, Damasus I, and several other bishops. In 380, the Edict of Thessalonica declared Nicene Christianity to be the state religion of the empire, with the name "Catholic Christians" reserved for those who accepted that faith. While the civil power in the Eastern Roman Empire controlled the church, and the patriarch of Constantinople, the capital, wielded much power, in the Western Roman Empire, the bishops of Rome were able to consolidate the influence and power they already possessed. After the fall of the Western Roman Empire, barbarian tribes were converted to Arian Christianity or Nicene Christianity; Clovis I, king of the Franks, was the first important barbarian ruler to convert to the mainstream church rather than Arianism, allying himself with the papacy. Other tribes, such as the Visigoths, later abandoned Arianism in favour of the established church. After the fall of the Western Roman Empire, the pope served as a source of authority and continuity. Pope Gregory I (c. 540–604) administered the church with strict reform. From an ancient senatorial family, Gregory worked with the stern judgement and discipline typical of ancient Roman rule. Theologically, he represents the shift from the classical to the medieval outlook; his popular writings are full of dramatic miracles, potent relics, demons, angels, ghosts, and the approaching end of the world. Gregory's successors were largely dominated by the exarch of Ravenna, the Byzantine emperor's representative in the Italian Peninsula. These humiliations, the weakening of the Byzantine Empire in the face of the Muslim conquests, and the inability of the emperor to protect the papal estates against the Lombards, made Pope Stephen II turn from Emperor Constantine V. He appealed to the Franks to protect his lands. Pepin the Short subdued the Lombards and donated Italian land to the papacy. When Pope Leo III crowned Charlemagne (800) as emperor, he established the precedent that, in Western Europe, no man would be emperor without being crowned by a pope. The low point of the papacy was 867–1049. This period includes the Saeculum obscurum, the Crescentii era, and the Tusculan Papacy. The papacy came under the control of vying political factions. Popes were variously imprisoned, starved, killed, and deposed by force. The family of a certain papal official made and unmade popes for fifty years. The official's great-grandson, Pope John XII, held orgies of debauchery in the Lateran Palace. Emperor Otto I had John accused in an ecclesiastical court, which deposed him and elected a layman as Pope Leo VIII. John mutilated the Imperial representatives in Rome and had himself reinstated as pope. Conflict between the Emperor and the papacy continued, and eventually dukes in league with the emperor were buying bishops and popes almost openly. In 1049, Leo IX traveled to the major cities of Europe to deal with the church's moral problems firsthand, notably simony and clerical marriage and concubinage. With his long journey, he restored the prestige of the papacy in Northern Europe. From the 7th century it became common for European monarchies and nobility to found churches and perform investiture or deposition of clergy in their states and fiefdoms, their personal interests causing corruption among the clergy. This practice had become common because often the prelates and secular rulers were also participants in public life. To combat this and other practices that had been seen as corrupting the Church between the years 900 and 1050, centres emerged promoting ecclesiastical reform, the most important being the Abbey of Cluny, which spread its ideals throughout Europe. This reform movement gained strength with the election of Pope Gregory VII in 1073, who adopted a series of measures in the movement known as the Gregorian Reform, in order to fight strongly against simony and the abuse of civil power and try to restore ecclesiastical discipline, including clerical celibacy. This conflict between popes and secular autocratic rulers such as the Holy Roman Emperor Henry IV and King Henry I of England, known as the Investiture controversy, was only resolved in 1122, by the Concordat of Worms, in which Pope Callixtus II decreed that clerics were to be invested by clerical leaders, and temporal rulers by lay investiture. Soon after, Pope Alexander III began reforms that would lead to the establishment of canon law. Since the beginning of the 7th century, Islamic conquests had succeeded in controlling much of the southern Mediterranean, and represented a threat to Christianity. In 1095, the Byzantine emperor, Alexios I Komnenos, asked for military aid from Pope Urban II in the ongoing Byzantine–Seljuq wars. Urban, at the council of Clermont, called the First Crusade to assist the Byzantine Empire to regain the old Christian territories, especially Jerusalem. With the East–West Schism, the Eastern Orthodox Church and the Catholic Church split definitively in 1054. This fracture was caused more by political events than by slight divergences of creed. Popes had galled the Byzantine emperors by siding with the king of the Franks, crowning a rival Roman emperor, appropriating the Exarchate of Ravenna, and driving into Greek Italy. In the Middle Ages, popes struggled with monarchs over power. From 1309 to 1377, the pope resided not in Rome but in Avignon. The Avignon Papacy was notorious for greed and corruption. During this period, the pope was effectively an ally of the Kingdom of France, alienating France's enemies, such as the Kingdom of England. The pope was understood to have the power to draw on the Treasury of Merit built up by the saints and by Christ, so that he could grant indulgences, reducing one's time in purgatory. The concept that a monetary fine or donation accompanied contrition, confession, and prayer eventually gave way to the common assumption that indulgences depended on a simple monetary contribution. The popes condemned misunderstandings and abuses, but were too pressed for income to exercise effective control over indulgences. Popes also contended with the cardinals, who sometimes attempted to assert the authority of Catholic Ecumenical Councils over the pope's. Conciliarism holds that the supreme authority of the church lies with a General Council, not with the pope. Its foundations were laid early in the 13th century, and it culminated in the 15th century with Jean Gerson as its leading spokesman. The failure of Conciliarism to gain broad acceptance after the 15th century is taken as a factor in the Protestant Reformation. Various Antipopes challenged papal authority, especially during the Western Schism (1378–1417). It came to a close when the Council of Constance, at the high-point of Concilliarism, decided among the papal claimants. The Eastern Church continued to decline with the Eastern Roman (Byzantine) Empire, undercutting Constantinople's claim to equality with Rome. Twice an Eastern Emperor tried to force the Eastern Church to reunify with the West. First in the Second Council of Lyon (1272–1274) and secondly in the Council of Florence (1431–1449). Papal claims of superiority were a sticking point in reunification, which failed in any event. In the 15th century, the Ottoman Empire captured Constantinople and ended the Byzantine Empire. Protestant Reformers criticized the papacy as corrupt and characterized the pope as the antichrist. Popes instituted a Catholic Reformation (1560–1648), which addressed the challenges of the Protestant Reformation and instituted internal reforms. Pope Paul III initiated the Council of Trent (1545–1563), whose definitions of doctrine and whose reforms sealed the triumph of the papacy over elements in the church that sought conciliation with Protestants and opposed papal claims. Gradually forced to give up secular power to the increasingly assertive European nation states, the popes focused on spiritual issues. In 1870, the First Vatican Council proclaimed the dogma of papal infallibility for the most solemn occasions when the pope speaks ex cathedra when issuing a definition of faith or morals. Later the same year, Victor Emmanuel II of Italy seized Rome from the pope's control and substantially completed the unification of Italy. In 1929, the Lateran Treaty between the Kingdom of Italy and the Holy See established Vatican City as an independent city-state, guaranteeing papal independence from secular rule. In 1950, Pope Pius XII defined the Assumption of Mary as dogma, the only time a pope has spoken ex cathedra since papal infallibility was explicitly declared. The Primacy of St. Peter, the controversial doctrinal basis of the pope's authority, continues to divide the eastern and western churches and to separate Protestants from Rome. The writings of several Early Church fathers contain references to the authority and unique position held by the bishops of Rome, providing valuable insight into the recognition and significance of the papacy during the early Christian era. These sources attest to the acknowledgement of the bishop of Rome as an influential figure within the Church, with some emphasizing the importance of adherence to Rome's teachings and decisions. Such references served to establish the concept of papal primacy and have continued to inform Catholic theology and practice. Cyprian of Carthage, in his letters, recognized the bishop of Rome as the successor of St. Peter in his Letter 55 (c. 251 AD), which is addressed to Pope Cornelius, and affirmed his unique authority in the early Christian Church. Cornelius [the Bishop of Rome] was made bishop by the choice of God and of His Christ, by the favorable witness of almost all the clergy, by the votes of the people who were present, and by the assembly of ancient priests and good men. And he was made bishop when no one else had been made bishop before him when the position of Fabian, that is to say, the position of Peter and the office of the bishop's chair, was vacant. But the position once has been filled by the will of God and that appointment has been ratified by the consent of us all, if anyone wants to be made bishop after that, it has to be done outside the church; if a man does not uphold the unity of the Church's unity, it is not possible for him to have the Church's ordination. Irenaeus of Lyons, a prominent Christian theologian of the second century, provided a list of early popes in his work Against Heresies III. The list covers the period from Saint Peter to Pope Eleutherius who served from 174 to 189 AD. The blessed apostles [Peter and Paul], then, having founded and built up the Church [in Rome], committed into the hands of Linus the office of the episcopate. Of this Linus, Paul makes mention in the Epistles to Timothy. To him succeeded Anacletus; and after him, in the third place from the apostles, Clement was allotted the bishopric. [...] To this Clement there succeeded Eviristus. Alexander followed Evaristus; then, sixth from the apostles, Sixtus was appointed; after him, Telephorus, who was gloriously martyred; then Hyginus; after him, Pius; then after him, Anicetus. Soter having succeeded Anicetus, Eleutherius does now, in the twelfth place from the apostles, hold the inheritance of the episcopate. Ignatius of Antioch wrote in his "Epistle to the Romans" that the church in Rome is "the church that presides over love". ...the Church which is beloved and enlightened by the will of Him that wills all things which are according to the love of Jesus Christ our God, which also presides in the place of the region of the Romans, worthy of God, worthy of honour, worthy of the highest happiness, worthy of praise, worthy of obtaining her every desire, worthy of being deemed holy, and which presides over love, is named from Christ, and from the Father, which I also salute in the name of Jesus Christ, the Son of the Father: to those who are united, both according to the flesh and spirit, to every one of His commandments; Augustine of Hippo, in his Letter 53, wrote a list of 38 popes from Saint Peter to Siricius. The order of this list differs from the lists of Irenaeus and the Annuario Pontificio in that Augustine's list claims that Linus was succeeded by Clement and Clement was succeeded by Anacletus as in the list of Eusebius, while the other two lists switch the positions of Clement and Anacletus. For if the lineal succession of bishops is to be taken into account, with how much more certainty and benefit to the Church do we reckon back till we reach Peter himself, to whom, as bearing in a figure the whole Church, the Lord said: Upon this rock will I build my Church, and the gates of hell shall not prevail against it! Matthew 16:18. The successor of Peter was Linus, and his successors in unbroken continuity were these:— Clement, Anacletus, Evaristus... Eusebius mentions Linus as Saint Peter's successor and Clement as the third bishop of Rome in his book Church History. As recorded by Eusebius, Clement worked with Saint Paul as his "co-laborer". As to the rest of his followers, Paul testifies that Crescens was sent to Gaul; but Linus, whom he mentions in the Second Epistle to Timothy as his companion at Rome, was Peter’s successor in the episcopate of the church there, as has already been shown. Clement also, who was appointed third bishop of the church at Rome, was, as Paul testifies, his co-laborer and fellow-soldier. Tertullian wrote in his work "The Prescription Against Heretics" about the authority of the church in Rome. In this work, Tertullian said that the Church in Rome has the authority of the Apostles because of its apostolic foundation. Since, moreover, you are close upon Italy, you have Rome, from which there comes even into our own hands the very authority (of apostles themselves). How happy is its church, on which apostles poured forth all their doctrine along with their blood! Where Peter endures a passion like his Lord's! Where Paul wins his crown in a death like John's where the Apostle John was first plunged, unhurt, into boiling oil, and thence remitted to his island-exile! According to the same book, Clement of Rome was ordained by Saint Peter as the bishop of Rome. For this is the manner in which the apostolic churches transmit their registers: as the church of Smyrna, which records that Polycarp was placed therein by John; as also the church of Rome, which makes Clement to have been ordained in like manner by Peter. Optatus the bishop of Milevis in Numidia (today's Algeria) and a contemporary of the Donatist schism, presents a detailed analysis of the origins, beliefs, and practices of the Donatists, as well as the events and debates surrounding the schism, in his book The Schism of the Donatists (367 A.D). In the book, Optatus wrote about the position of the bishop of Rome in maintaining the unity of the Church. You cannot deny that you are aware that in the city of Rome the episcopal chair was given first to Peter; the chair in which Peter sat, the same who was head—that is why he is also called Cephas [‘Rock’]—of all the apostles; the one chair in which unity is maintained by all The Catholic Church teaches that, within the Christian community, the bishops as a body have succeeded to the body of the apostles (apostolic succession) and the bishop of Rome has succeeded to Saint Peter. Scriptural texts proposed in support of Peter's special position in relation to the church include: I tell you, you are Peter, and on this rock I will build my church, and the gates of hell shall not prevail against it. I will give you the keys of the kingdom of heaven, and whatever you bind on earth shall be bound in heaven, and whatever you loose on earth shall be loosed in heaven. Simon, Simon, behold, Satan demanded to have you, that he might sift you like wheat, but I have prayed for you that your faith may not fail. And when you have turned again, strengthen your brothers. Feed my sheep. The symbolic keys in the Papal coats of arms are a reference to the phrase "the keys of the kingdom of heaven" in the first of these texts. Some Protestant writers have maintained that the "rock" that Jesus speaks of in this text is Jesus himself or the faith expressed by Peter. This idea is undermined by the Biblical usage of "Cephas", which is the masculine form of "rock" in Aramaic, to describe Peter. The Encyclopædia Britannica comments that "the consensus of the great majority of scholars today is that the most obvious and traditional understanding should be construed, namely, that rock refers to the person of Peter". According to the Catholic church, the pope is also the new Eliakim, a figure in the Old Testament of the Bible who directed the affairs of the royal court, managed the palace staff, and handled state affairs. Isaiah also describes him as having the key to the house of David, which symbolizes his authority and power. Both Matthew 16:18-19 and Isaiah 22:22 show similarities between Eliakim and Peter getting keys which symbolise power. Eliakim gets the power to close and open, while Peter gets the power to bind and loose. According to the Book of Isaiah, Eliakim receives the keys and power to close and open. I will place the key of the House of David on his shoulder; what he opens, no one will shut, what he shuts, no one will open. According to book of Matthew Peter also gets keys and power to bind and loose. I will give you the keys of the kingdom of heaven, and whatever you bind on earth shall be bound in heaven, and whatever you loose on earth shall be loosed in heaven.” In the Books of Isaiah 22:3 and Matthew 16:18, both Eliakim and Peter are compared to an object. Eliakim to a peg (a structure that is driven into a wall or other structure to provide support and stability) while Peter to a rock. And I will fasten him like a peg in a secure place, and he will become a throne of honor to his father's house. In Matthew 16:18 Peter was compared to a rock. And I tell you, you are Peter, and on this rock I will build my church, and the gates of hell shall not prevail against it The pope was originally chosen by those senior clergymen resident in and near Rome. In 1059, the electorate was restricted to the cardinals of the Holy Roman Church, and the individual votes of all cardinal electors were made equal in 1179. The electors are now limited to those who have not reached 80 on the day before the death or resignation of a pope. The pope does not need to be a cardinal elector or indeed a cardinal; since the pope is the bishop of Rome, only those who can be ordained a bishop can be elected, which means that any male baptized Catholic is eligible. The last to be elected when not yet a bishop was Gregory XVI in 1831, the last to be elected when not even a priest was Leo X in 1513, and the last to be elected when not a cardinal was Urban VI in 1378. If someone who is not a bishop is elected, he must be given episcopal ordination before the election is announced to the people. The Second Council of Lyon was convened on 7 May 1274, to regulate the election of the pope. This Council decreed that the cardinal electors must meet within ten days of the pope's death, and that they must remain in seclusion until a pope has been elected; this was prompted by the three-year sede vacante following the death of Clement IV in 1268. By the mid-16th century, the electoral process had evolved into its present form, allowing for variation in the time between the death of the pope and the meeting of the cardinal electors. Traditionally, the vote was conducted by acclamation, by selection (by committee), or by plenary vote. Acclamation was the simplest procedure, consisting entirely of a voice vote. The election of the pope almost always takes place in the Sistine Chapel, in a sequestered meeting called a "conclave" (so called because the cardinal electors are theoretically locked in, cum clave, i.e., with key, until they elect a new pope). Three cardinals are chosen by lot to collect the votes of absent cardinal electors (by reason of illness), three are chosen by lot to count the votes, and three are chosen by lot to review the count of the votes. The ballots are distributed and each cardinal elector writes the name of his choice on it and pledges aloud that he is voting for "one whom under God I think ought to be elected" before folding and depositing his vote on a plate atop a large chalice placed on the altar. For the Papal conclave, 2005, a special urn was used for this purpose instead of a chalice and plate. The plate is then used to drop the ballot into the chalice, making it difficult for electors to insert multiple ballots. Before being read, the ballots are counted while still folded; if the number of ballots does not match the number of electors, the ballots are burned unopened and a new vote is held. Otherwise, each ballot is read aloud by the presiding Cardinal, who pierces the ballot with a needle and thread, stringing all the ballots together and tying the ends of the thread to ensure accuracy and honesty. Balloting continues until someone is elected by a two-thirds majority. (With the promulgation of Universi Dominici Gregis in 1996, a simple majority after a deadlock of twelve days was allowed, but this was revoked by Pope Benedict XVI by motu proprio in 2007.) One of the most prominent aspects of the papal election process is the means by which the results of a ballot are announced to the world. Once the ballots are counted and bound together, they are burned in a special stove erected in the Sistine Chapel, with the smoke escaping through a small chimney visible from Saint Peter's Square. The ballots from an unsuccessful vote are burned along with a chemical compound to create black smoke, or fumata nera. (Traditionally, wet straw was used to produce the black smoke, but this was not completely reliable. The chemical compound is more reliable than the straw.) When a vote is successful, the ballots are burned alone, sending white smoke (fumata bianca) through the chimney and announcing to the world the election of a new pope. Starting with the Papal conclave, 2005, church bells are also rung as a signal that a new pope has been chosen. The dean of the College of Cardinals then asks two solemn questions of the man who has been elected. First he asks, "Do you freely accept your election as supreme pontiff?" If he replies with the word "Accepto", his reign begins at that instant. In practice, any cardinal who intends not to accept will explicitly state this before he receives a sufficient number of votes to become pope. The dean asks next, "By what name shall you be called?" The new pope announces the regnal name he has chosen. If the dean himself is elected pope, the vice dean performs this task. The new pope is led to the Room of Tears, a dressing room where three sets of white papal vestments (immantatio) await in three sizes. Donning the appropriate vestments and reemerging into the Sistine Chapel, the new pope is given the "Fisherman's Ring" by the camerlengo of the Holy Roman Church. The pope assumes a place of honor as the rest of the cardinals wait in turn to offer their first "obedience" (adoratio) and to receive his blessing. The cardinal protodeacon announces from a balcony over St. Peter's Square the following proclamation: Annuntio vobis gaudium magnum! Habemus Papam! ("I announce to you a great joy! We have a pope!"). He announces the new pope's Christian name along with his newly chosen regnal name. Until 1978, the pope's election was followed in a few days by the papal coronation, which started with a procession with great pomp and circumstance from the Sistine Chapel to St. Peter's Basilica, with the newly elected pope borne in the sedia gestatoria. After a solemn Papal Mass, the new pope was crowned with the triregnum (papal tiara) and he gave for the first time as pope the famous blessing Urbi et Orbi ("to the City [Rome] and to the World"). Another renowned part of the coronation was the lighting of a bundle of flax at the top of a gilded pole, which would flare brightly for a moment and then promptly extinguish, as he said, Sic transit gloria mundi ("Thus passes worldly glory"). A similar warning against papal hubris made on this occasion was the traditional exclamation, "Annos Petri non-videbis", reminding the newly crowned pope that he would not live to see his rule lasting as long as that of St. Peter. According to tradition, he headed the church for 35 years and has thus far been the longest-reigning pope in the history of the Catholic Church. The Latin term, sede vacante ("while the see is vacant"), refers to a papal interregnum, the period between the death or resignation of a pope and the election of his successor. From this term is derived the term sedevacantism, which designates a category of dissident Catholics who maintain that there is no canonically and legitimately elected pope, and that there is therefore a sede vacante. For centuries, from 1378 on, those elected to the papacy were predominantly Italians. Prior to the election of the Polish-born John Paul II in 1978, the last non-Italian was Adrian VI of the Netherlands, elected in 1522. John Paul II was followed by election of the German-born Benedict XVI, who was in turn followed by Argentine-born Francis, the first non-European after 1272 years and the first Latin American (albeit of Italian ancestry). The current regulations regarding a papal interregnum—that is, a sede vacante ("vacant seat")—were promulgated by Pope John Paul II in his 1996 document Universi Dominici Gregis. During the sede vacante period, the College of Cardinals is collectively responsible for the government of the Church and of the Vatican itself, under the direction of the Camerlengo of the Holy Roman Church. Canon law specifically forbids the cardinals from introducing any innovation in the government of the Church during the vacancy of the Holy See. Any decision that requires the assent of the pope has to wait until the new pope has been elected and accepts office. In recent centuries, when a pope was judged to have died, it was reportedly traditional for the cardinal camerlengo to confirm the death ceremonially by gently tapping the pope's head thrice with a silver hammer, calling his birth name each time. This was not done on the deaths of popes John Paul I and John Paul II. The cardinal camerlengo retrieves the Ring of the Fisherman and cuts it in two in the presence of the cardinals. The pope's seals are defaced, to keep them from ever being used again, and his personal apartment is sealed. The body lies in state for several days before being interred in the crypt of a leading church or cathedral; all popes who have died in the 20th and 21st centuries have been interred in St. Peter's Basilica. A nine-day period of mourning (novendialis) follows the interment. It is highly unusual for a pope to resign. The 1983 Code of Canon Law states, "If it happens that the Roman Pontiff resigns his office, it is required for validity that the resignation is made freely and properly manifested but not that it is accepted by anyone." Benedict XVI, who vacated the Holy See on 28 February 2013, was the most recent to do so, and the first since Gregory XII's resignation in 1415. Popes adopt a new name on their accession, known as papal name, in Italian and Latin. Currently, after a new pope is elected and accepts the election, he is asked, "By what name shall you be called?" The new pope chooses the name by which he will be known from that point on. The senior cardinal deacon, or cardinal protodeacon, then appears on the balcony of Saint Peter's to proclaim the new pope by his birth name, and announce his papal name in Latin. It's customary when referring to popes to translate the regnal name into all local languages. For example, the current pope bears the papal name Papa Franciscus in Latin and Papa Francesco in Italian, but Papa Francisco in his native Spanish, Pope Francis in English, etc. The official list of titles of the pope, in the order in which they are given in the Annuario Pontificio, is: Bishop of Rome, Vicar of Jesus Christ, Successor of the Prince of the Apostles, Supreme Pontiff of the Universal Church, Primate of Italy, Archbishop and Metropolitan of the Roman Province, Sovereign of the Vatican City State, Servant of the servants of God. The best-known title, that of "pope", does not appear in the official list, but is commonly used in the titles of documents, and appears, in abbreviated form, in their signatures. Thus Paul VI signed as "Paulus PP. VI", the "PP." standing for "papa pontifex" ("pope and pontiff"). The title "pope" was from the early 3rd century an honorific designation used for any bishop in the West. In the East, it was used only for the bishop of Alexandria. Marcellinus (d. 304) is the first bishop of Rome shown in sources to have had the title "pope" used of him. From the 6th century, the imperial chancery of Constantinople normally reserved this designation for the bishop of Rome. From the early 6th century, it began to be confined in the West to the bishop of Rome, a practice that was firmly in place by the 11th century. In Eastern Christianity, where the title "pope" is used also of the bishop of Alexandria, the bishop of Rome is often referred to as the "pope of Rome", regardless of whether the speaker or writer is in communion with Rome or not. "Vicar of Jesus Christ" (Vicarius Iesu Christi) is one of the official titles of the pope given in the Annuario Pontificio. It is commonly used in the slightly abbreviated form "vicar of Christ" (vicarius Christi). While it is only one of the terms with which the pope is referred to as "vicar", it is "more expressive of his supreme headship of the Church on Earth, which he bears in virtue of the commission of Christ and with vicarial power derived from him", a vicarial power believed to have been conferred on Saint Peter when Christ said to him: "Feed my lambs...Feed my sheep". The first record of the application of this title to a bishop of Rome appears in a synod of 495 with reference to Gelasius I. But at that time, and down to the 9th century, other bishops too referred to themselves as vicars of Christ, and for another four centuries this description was sometimes used of kings and even judges, as it had been used in the 5th and 6th centuries to refer to the Byzantine emperor. Earlier still, in the 3rd century, Tertullian used "vicar of Christ" to refer to the Holy Spirit sent by Jesus. Its use specifically for the pope appears in the 13th century in connection with the reforms of Pope Innocent III, as can be observed already in his 1199 letter to Leo I, King of Armenia. Other historians suggest that this title was already used in this way in association with the pontificate of Eugene III (1145–1153). This title "vicar of Christ" is thus not used of the pope alone and has been used of all bishops since the early centuries. The Second Vatican Council referred to all bishops as "vicars and ambassadors of Christ", and this description of the bishops was repeated by John Paul II in his encyclical Ut unum sint, 95. The difference is that the other bishops are vicars of Christ for their own local churches, the pope is vicar of Christ for the whole Church. On at least one occasion the title "vicar of God" (a reference to Christ as God) was used of the pope. The title "vicar of Peter" (vicarius Petri) is used only of the pope, not of other bishops. Variations of it include: "Vicar of the Prince of the Apostles" (Vicarius Principis Apostolorum) and "Vicar of the Apostolic See" (Vicarius Sedis Apostolicae). Saint Boniface described Pope Gregory II as vicar of Peter in the oath of fealty that he took in 722. In today's Roman Missal, the description "vicar of Peter" is found also in the collect of the Mass for a saint who was a pope. The term "pontiff" is derived from the Latin: pontifex, which literally means "bridge builder" (pons + facere) and which designated a member of the principal college of priests in pagan Rome. The Latin word was translated into ancient Greek variously: as Ancient Greek: ἱεροδιδάσκαλος, Ancient Greek: ἱερονόμος, Ancient Greek: ἱεροφύλαξ, Ancient Greek: ἱεροφάντης (hierophant), or Ancient Greek: ἀρχιερεύς (archiereus, high priest) The head of the college was known as the Latin: Pontifex Maximus (the greatest pontiff). In Christian use, pontifex appears in the Vulgate translation of the New Testament to indicate the High Priest of Israel (in the original Koine Greek, ἀρχιερεύς). The term came to be applied to any Christian bishop, but since the 11th century commonly refers specifically to the bishop of Rome, who is more strictly called the "Roman Pontiff". The use of the term to refer to bishops in general is reflected in the terms "Roman Pontifical" (a book containing rites reserved for bishops, such as confirmation and ordination), and "pontificals" (the insignia of bishops). The Annuario Pontificio lists as one of the official titles of the pope that of "Supreme Pontiff of the Universal Church" (Latin: Summus Pontifex Ecclesiae Universalis). He is also commonly called the supreme pontiff or the sovereign pontiff (Latin: summus pontifex). Pontifex Maximus, similar in meaning to Summus Pontifex, is a title commonly found in inscriptions on papal buildings, paintings, statues and coins, usually abbreviated as "Pont. Max" or "P.M." The office of Pontifex Maximus, or head of the College of Pontiffs, was held by Julius Caesar and thereafter, by the Roman emperors, until Gratian (375–383) relinquished it. Tertullian, when he had become a Montanist, used the title derisively of either the pope or the bishop of Carthage. The popes began to use this title regularly only in the 15th century. Although the description "servant of the servants of God" (Latin: servus servorum Dei) was also used by other Church leaders, including Augustine of Hippo and Benedict of Nursia, it was first used extensively as a papal title by Gregory the Great, reportedly as a lesson in humility for the patriarch of Constantinople, John the Faster, who had assumed the title "ecumenical patriarch". It became reserved for the pope in the 12th century and is used in papal bulls and similar important papal documents. From 1863 until 2005, the Annuario Pontificio also included the title "patriarch of the West". This title was first used by Pope Theodore I in 642, and was only used occasionally. Indeed, it did not begin to appear in the pontifical yearbook until 1863. On 22 March 2006, the Vatican released a statement explaining this omission on the grounds of expressing a "historical and theological reality" and of "being useful to ecumenical dialogue". The title patriarch of the West symbolized the pope's special relationship with, and jurisdiction over, the Latin Church—and the omission of the title neither symbolizes in any way a change in this relationship, nor distorts the relationship between the Holy See and the Eastern Churches, as solemnly proclaimed by the Second Vatican Council. Other titles commonly used are "His Holiness" (either used alone or as an honorific prefix as in "His Holiness Pope Francis"; and as "Your Holiness" as a form of address), "Holy Father". In Spanish and Italian, "Beatísimo/Beatissimo Padre" (Most Blessed Father) is often used in preference to "Santísimo/Santissimo Padre" (Most Holy Father). In the medieval period, "Dominus Apostolicus" ("the Apostolic Lord") was also used. Pope Francis signs some documents with his name alone, either in Latin ("Franciscus", as in an encyclical dated 29 June 2013) or in another language. Other documents he signs in accordance with the tradition of using Latin only and including the abbreviated form "PP.", for the Latin Papa ("Pope"). Popes who have an ordinal numeral in their name traditionally place the abbreviation "PP." before the ordinal numeral, as in "Benedictus PP. XVI" (Pope Benedict XVI), except in papal bulls of canonization and decrees of ecumenical councils, which a pope signs with the formula, "Ego N. Episcopus Ecclesiae catholicae", without the numeral, as in "Ego Benedictus Episcopus Ecclesiae catholicae" (I, Benedict, Bishop of the Catholic Church). In heraldry, each pope has his own personal coat of arms. Though unique for each pope, the arms have for several centuries been traditionally accompanied by two keys in saltire (i.e., crossed over one another so as to form an X) behind the escutcheon (shield) (one silver key and one gold key, tied with a red cord), and above them a silver triregnum with three gold crowns and red infulae (lappets—two strips of fabric hanging from the back of the triregnum which fall over the neck and shoulders when worn). This is blazoned: "two keys in saltire or and argent, interlacing in the rings or, beneath a tiara argent, crowned or". The 21st century has seen departures from this tradition. In 2005, Pope Benedict XVI, while maintaining the crossed keys behind the shield, omitted the papal tiara from his personal coat of arms, replacing it with a mitre with three horizontal lines. Beneath the shield he added the pallium, a papal symbol of authority more ancient than the tiara, the use of which is also granted to metropolitan archbishops as a sign of communion with the See of Rome. Although the tiara was omitted in the pope's personal coat of arms, the coat of arms of the Holy See, which includes the tiara, remained unaltered. In 2013, Pope Francis maintained the mitre that replaced the tiara, but omitted the pallium. The flag most frequently associated with the pope is the yellow and white flag of Vatican City, with the arms of the Holy See (blazoned: "Gules, two keys in saltire or and argent, interlacing in the rings or, beneath a tiara argent, crowned or") on the right-hand side (the "fly") in the white half of the flag (the left-hand side—the "hoist"—is yellow). The pope's escucheon does not appear on the flag. This flag was first adopted in 1808, whereas the previous flag had been red and gold. Although Pope Benedict XVI replaced the triregnum with a mitre on his personal coat of arms, it has been retained on the flag. Pope Pius V (reigned 1566–1572), is often credited with having originated the custom whereby the pope wears white, by continuing after his election to wear the white habit of the Dominican order. In reality, the basic papal attire was white long before. The earliest document that describes it as such is the Ordo XIII, a book of ceremonies compiled in about 1274. Later books of ceremonies describe the pope as wearing a red mantle, mozzetta, camauro and shoes, and a white cassock and stockings. Many contemporary portraits of 15th and 16th-century predecessors of Pius V show them wearing a white cassock similar to his. The status and authority of the pope in the Catholic Church was dogmatically defined by the First Vatican Council on 18 July 1870. In its Dogmatic Constitution of the Church of Christ, the council established the following canons: If anyone says that the blessed Apostle Peter was not established by the Lord Christ as the chief of all the apostles, and the visible head of the whole militant Church, or, that the same received great honour but did not receive from the same our Lord Jesus Christ directly and immediately the primacy in true and proper jurisdiction: let him be anathema. If anyone says that it is not from the institution of Christ the Lord Himself, or by divine right that the blessed Peter has perpetual successors in the primacy over the universal Church, or that the Roman Pontiff is not the successor of blessed Peter in the same primacy, let him be anathema. If anyone thus speaks, that the Roman pontiff has only the office of inspection or direction, but not the full and supreme power of jurisdiction over the universal Church, not only in things which pertain to faith and morals, but also in those which pertain to the discipline and government of the Church spread over the whole world; or, that he possesses only the more important parts, but not the whole plenitude of this supreme power; or that this power of his is not ordinary and immediate, or over the churches altogether and individually, and over the pastors and the faithful altogether and individually: let him be anathema. We, adhering faithfully to the tradition received from the beginning of the Christian faith, to the glory of God, our Saviour, the elevation of the Catholic religion and the salvation of Christian peoples, with the approbation of the sacred Council, teach and explain that the dogma has been divinely revealed: that the Roman Pontiff, when he speaks ex cathedra, that is, when carrying out the duty of the pastor and teacher of all Christians by his supreme apostolic authority he defines a doctrine of faith or morals to be held by the universal Church, through the divine assistance promised him in blessed Peter, operates with that infallibility with which the divine Redeemer wished that His church be instructed in defining doctrine on faith and morals; and so such definitions of the Roman Pontiff from himself, but not from the consensus of the Church, are unalterable. But if anyone presumes to contradict this definition of Ours, which may God forbid: let him be anathema. In its Dogmatic Constitution on the Church (1964), the Second Vatican Council declared: Among the principal duties of bishops the preaching of the Gospel occupies an eminent place. For bishops are preachers of the faith, who lead new disciples to Christ, and they are authentic teachers, that is, teachers endowed with the authority of Christ, who preach to the people committed to them the faith they must believe and put into practice, and by the light of the Holy Spirit illustrate that faith. They bring forth from the treasury of Revelation new things and old, making it bear fruit and vigilantly warding off any errors that threaten their flock. Bishops, teaching in communion with the Roman Pontiff, are to be respected by all as witnesses to divine and Catholic truth. In matters of faith and morals, the bishops speak in the name of Christ and the faithful are to accept their teaching and adhere to it with a religious assent. This religious submission of mind and will must be shown in a special way to the authentic magisterium of the Roman Pontiff, even when he is not speaking ex cathedra; that is, it must be shown so that his supreme magisterium is acknowledged with reverence, the judgments made by him are sincerely adhered to, according to his manifest mind and will. His mind and will in the matter may be known either from the character of the documents, from his frequent repetition of the same doctrine, or from his manner of speaking. ... this infallibility with which the Divine Redeemer willed His Church to be endowed in defining doctrine of faith and morals, extends as far as the deposit of Revelation extends, which must be religiously guarded and faithfully expounded. And this is the infallibility which the Roman Pontiff, the head of the College of Bishops, enjoys in virtue of his office, when, as the supreme shepherd and teacher of all the faithful, who confirms his brethren in their faith, by a definitive act he proclaims a doctrine of faith or morals. And therefore his definitions, of themselves, and not from the consent of the Church, are justly styled irreformable, since they are pronounced with the assistance of the Holy Spirit, promised to him in blessed Peter, and therefore they need no approval of others, nor do they allow an appeal to any other judgment. For then the Roman Pontiff is not pronouncing judgment as a private person, but as the supreme teacher of the universal Church, in whom the charism of infallibility of the Church itself is individually present, he is expounding or defending a doctrine of Catholic faith. The infallibility promised to the Church resides also in the body of Bishops, when that body exercises the supreme magisterium with the successor of Peter. To these definitions the assent of the Church can never be wanting, on account of the activity of that same Holy Spirit, by which the whole flock of Christ is preserved and progresses in unity of faith. On 11 October 2012, on the occasion of the 50th anniversary of the opening of the Second Vatican Council 60 prominent theologians, (including Hans Küng), put out a Declaration, stating that the intention of Vatican II to balance authority in the Church has not been realised. "Many of the key insights of Vatican II have not at all, or only partially, been implemented... A principal source of present-day stagnation lies in misunderstanding and abuse affecting the exercise of authority in our Church." The pope's official seat is in the Archbasilica of Saint John Lateran, considered the cathedral of the Diocese of Rome, and his official residence is the Apostolic Palace. He also possesses a summer residence at Castel Gandolfo, situated on the site of the ancient city of Alba Longa. The names "Holy See" and "Apostolic See" are ecclesiastical terminology for the ordinary jurisdiction of the bishop of Rome (including the Roman Curia); the pope's various honors, powers, and privileges within the Catholic Church and the international community derive from his Episcopate of Rome in lineal succession from the Saint Peter, one of the twelve apostles. Consequently, Rome has traditionally occupied a central position in the Catholic Church, although this is not necessarily so. The pope derives his pontificate from being the bishop of Rome but is not required to live there; according to the Latin formula ubi Papa, ibi Curia, wherever the pope resides is the central government of the Church. As such, between 1309 and 1378, the popes lived in Avignon, France, a period often called the "Babylonian captivity" in allusion to the Biblical narrative of Jews of the ancient Kingdom of Judah living as captives in Babylonia. Though the pope is the diocesan bishop of Rome, he delegates most of the day-to-day work of leading the diocese to the cardinal vicar, who assures direct episcopal oversight of the diocese's pastoral needs, not in his own name but in that of the pope. The current cardinal vicar is Angelo De Donatis, who was appointed to the office in June 2017. Though the progressive Christianisation of the Roman Empire in the 4th century did not confer upon bishops civil authority within the state, the gradual withdrawal of imperial authority during the 5th century left the pope the senior imperial civilian official in Rome, as bishops were increasingly directing civil affairs in other cities of the Western Empire. This status as a secular and civil ruler was vividly displayed by Pope Leo I's confrontation with Attila in 452. The first expansion of papal rule outside of Rome came in 728 with the Donation of Sutri, which in turn was substantially increased in 754, when the Frankish ruler Pippin the Younger gave to the pope the land from his conquest of the Lombards. The pope may have utilized the forged Donation of Constantine to gain this land, which formed the core of the Papal States. This document, accepted as genuine until the 15th century, states that Constantine the Great placed the entire Western Empire of Rome under papal rule. In 800, Pope Leo III crowned the Frankish ruler Charlemagne as Roman emperor, a major step toward establishing what later became known as the Holy Roman Empire; from that date onward the popes claimed the prerogative to crown the emperor, though the right fell into disuse after the coronation of Charles V in 1530. Pius VII was present at the coronation of Napoleon I in 1804 but did not actually perform the crowning. As mentioned above, the pope's sovereignty over the Papal States ended in 1870 with their annexation by Italy. Popes like Alexander VI, an ambitious if spectacularly corrupt politician, and Julius II, a formidable general and statesman, were not afraid to use power to achieve their own ends, which included increasing the power of the papacy. This political and temporal authority was demonstrated through the papal role in the Holy Roman Empire (especially prominent during periods of contention with the emperors, such as during the pontificates of Pope Gregory VII and Pope Alexander III). Papal bulls, interdict, and excommunication (or the threat thereof) have been used many times to exercise papal power. The bull Laudabiliter in 1155 authorized King Henry II of England to invade Ireland. In 1207, Innocent III placed England under interdict until King John made his kingdom a fiefdom to the pope, complete with yearly tribute, saying, "we offer and freely yield...to our lord Pope Innocent III and his catholic successors, the whole kingdom of England and the whole kingdom of Ireland with all their rights and appurtenences for the remission of our sins". The Bull Inter caetera in 1493 led to the Treaty of Tordesillas in 1494, which divided the world into areas of Spanish and Portuguese rule. The bull Regnans in Excelsis in 1570 excommunicated Queen Elizabeth I of England and declared that all her subjects were released from allegiance to her. The bull Inter gravissimas in 1582 established the Gregorian calendar. In recent decades, although the papacy has become less directly involved in politics, popes have nevertheless retained significant political influence. They have also served as mediators, with the support of the Catholic establishment. John Paul II, a native of Poland, was regarded as influential in the fall of Communism in Eastern Europe. He also mediated the Beagle conflict between Argentina and Chile, two predominantly Catholic countries. In the 21st century, Francis played a role in brokering the 2015 improvement in relations between the United States and Cuba. Under international law, a serving head of state has sovereign immunity from the jurisdiction of the courts of other countries, though not from that of international tribunals. This immunity is sometimes loosely referred to as "diplomatic immunity", which is, strictly speaking, the immunity enjoyed by the diplomatic representatives of a head of state. International law treats the Holy See, essentially the central government of the Catholic Church, as the juridical equal of a state. It is distinct from the state of Vatican City, existing for many centuries before the foundation of the latter. (It is common for publications and news media to use "the Vatican", "Vatican City", and even "Rome" as metonyms for the Holy See.) Most countries of the world maintain the same form of diplomatic relations with the Holy See that they entertain with other states. Even countries without those diplomatic relations participate in international organizations of which the Holy See is a full member. It is as head of the state-equivalent worldwide religious jurisdiction of the Holy See (not of the territory of Vatican City) that the U.S. Justice Department ruled that the pope enjoys head-of-state immunity. This head-of-state immunity, recognized by the United States, must be distinguished from that envisaged under the United States' Foreign Sovereign Immunities Act of 1976, which, while recognizing the basic immunity of foreign governments from being sued in American courts, lays down nine exceptions, including commercial activity and actions in the United States by agents or employees of the foreign governments. It was in relation to the latter that, in November 2008, the United States Court of Appeals in Cincinnati decided that a case over sexual abuse by Catholic priests could proceed, provided the plaintiffs could prove that the bishops accused of negligent supervision were acting as employees or agents of the Holy See and were following official Holy See policy. In April 2010, there was press coverage in Britain concerning a proposed plan by atheist campaigners and a prominent barrister to have Pope Benedict XVI arrested and prosecuted in the UK for alleged offences, dating from several decades before, in failing to take appropriate action regarding Catholic sex abuse cases and concerning their disputing his immunity from prosecution in that country. This was generally dismissed as "unrealistic and spurious". Another barrister said that it was a "matter of embarrassment that a senior British lawyer would want to allow himself to be associated with such a silly idea". Sovereign immunity does not apply to disputes relating to commercial transactions, and governmental units of the Holy See can face trial in foreign commercial courts. The first such trial to take place in the English courts is likely to occur in 2022 or 2023. The pope's claim to authority is either disputed or rejected outright by other churches, for various reasons. Other traditional Christian churches (Assyrian Church of the East, the Oriental Orthodox Church, the Eastern Orthodox Church, the Old Catholic Church, the Anglican Communion, the Independent Catholic churches, etc.) accept the doctrine of Apostolic succession and, to varying extents, papal claims to a primacy of honour, while generally rejecting the pope as the successor to Peter in any other sense than that of other bishops. Primacy is regarded as a consequence of the pope's position as bishop of the original capital city of the Roman Empire, a definition explicitly spelled out in the 28th canon of the Council of Chalcedon. These churches see no foundation to papal claims of universal immediate jurisdiction, or to claims of papal infallibility. Several of these churches refer to such claims as ultramontanism. In 1973, the United States Conference of Catholic Bishops' Committee on Ecumenical and Interreligious Affairs and the USA National Committee of the Lutheran World Federation in the official Catholic–Lutheran dialogue included this passage in a larger statement on papal primacy: In calling the pope the "Antichrist", the early Lutherans stood in a tradition that reached back into the eleventh century. Not only dissidents and heretics but even saints had called the bishop of Rome the "Antichrist" when they wished to castigate his abuse of power. What Lutherans understood as a papal claim to unlimited authority over everything and everyone reminded them of the apocalyptic imagery of Daniel 11, a passage that even prior to the Reformation had been applied to the pope as the Antichrist of the last days. Protestant denominations of Christianity reject the claims of Petrine primacy of honor, Petrine primacy of jurisdiction, and papal infallibility. These denominations vary from denying the legitimacy of the pope's claim to authority, to believing that the pope is the Antichrist from 1 John 2:18, the Man of Sin from 2 Thessalonians 2:3–12, and the Beast out of the Earth from Revelation 13:11–18. This sweeping rejection is held by, among others, some denominations of Lutherans: Confessional Lutherans hold that the pope is the Antichrist, stating that this article of faith is part of a quia ("because") rather than quatenus ("insofar as") subscription to the Book of Concord. In 1932, one of these Confessional churches, the Lutheran Church–Missouri Synod (LCMS), adopted A Brief Statement of the Doctrinal Position of the Missouri Synod, which a small number of Lutheran church bodies now hold. The Lutheran Churches of the Reformation, the Concordia Lutheran Conference, the Church of the Lutheran Confession, and the Illinois Lutheran Conference all hold to the Brief Statement, which the LCMS places on its website. The Wisconsin Evangelical Lutheran Synod (WELS), another Confessional Lutheran church that declares the Papacy to be the Antichrist, released its own statement, the "Statement on the Antichrist", in 1959. The WELS still holds to this statement. Historically, Protestants objected to the papacy's claim of temporal power over all secular governments, including territorial claims in Italy, the papacy's complex relationship with secular states such as the Roman and Byzantine empires, and the autocratic character of the papal office. In Western Christianity these objections both contributed to and are products of the Protestant Reformation. Groups sometimes form around antipopes, who claim the Pontificate without being canonically and properly elected to it. Traditionally, this term was reserved for claimants with a significant following of cardinals or other clergy. The existence of an antipope is usually due either to doctrinal controversy within the Church (heresy) or to confusion as to who is the legitimate pope at the time (schism). Briefly in the 15th century, three separate lines of popes claimed authenticity. In the earlier centuries of Christianity, the title "Pope", meaning "father", had been used by all bishops. Some popes used the term and others did not. Eventually, the title became associated especially with the bishop of Rome. In a few cases, the term is used for other Christian clerical authorities. In English, Catholic priests are still addressed as "father", but the term "pope" is reserved for the head of the church hierarchy. "Black Pope" is a name that was popularly, but unofficially, given to the superior general of the Society of Jesus due to the Jesuits' importance within the Church. This name, based on the black colour of his cassock, was used to suggest a parallel between him and the "White Pope" (since the time of Pius V the popes dress in white) and the cardinal prefect of the Congregation for the Evangelization of Peoples (formerly called the Sacred Congregation for the Propagation of the Faith), whose red cardinal's cassock gave him the name of the "Red Pope" in view of the authority over all territories that were not considered in some way Catholic. In the present time this cardinal has power over mission territories for Catholicism, essentially the Churches of Africa and Asia, but in the past his competence extended also to all lands where Protestants or Eastern Christianity was dominant. Some remnants of this situation remain, with the result that, for instance, New Zealand is still in the care of this Congregation. Since the papacy of Heraclas in the 3rd century, the bishop of Alexandria in both the Coptic Orthodox Church and the Greek Orthodox Church of Alexandria continues to be called "pope", the former being called "Coptic pope" or, more properly, "Pope and Patriarch of All Africa on the Holy Orthodox and Apostolic Throne of Saint Mark the Evangelist and Holy Apostle" and the latter called "Pope and Patriarch of Alexandria and All Africa". In the Bulgarian Orthodox Church, Russian Orthodox Church, Serbian Orthodox Church and Macedonian Orthodox Church, it is not unusual for a village priest to be called a "pope" ("поп" pop). This is different from the words used for the head of the Catholic Church (Bulgarian "папа" papa, Russian "папа римский" papa rimskiy). Some new religious movements within Christianity, especially those that have disassociated themselves from the Catholic Church yet retain a Catholic hierarchical framework, have used the designation "pope" for a founder or current leader. Examples include the African Legio Maria Church and the European Palmarian Catholic Church in Spain. The Cao Dai, a Vietnamese faith that duplicates the Catholic hierarchy, is similarly headed by a pope. The longest papal reigns of those whose reign lengths can be determined from contemporary historical data are the following: During the Western Schism, Avignon Pope Benedict XIII (1394–1423) ruled for 28 years, 7 months and 12 days, which would place him third in the above list. Since he is regarded as an anti-pope, he is not included there. There have been a number of popes whose reign lasted about a month or less. In the following list the number of calendar days includes partial days. Thus, for example, if a pope's reign commenced on 1 August and he died on 2 August, this counts as reigning for two calendar days. Stephen (23–26 March 752) died of stroke three days after his election, and before his consecration as a bishop. He is not recognized as a valid pope, but was added to the lists of popes in the 15th century as Stephen II, causing difficulties in enumerating later popes named Stephen. The Holy See's Annuario Pontificio, in its list of popes and antipopes, attaches a footnote to its mention of Pope Stephen II: On the death of Zachary the Roman priest Stephen was elected; but, since four days later he died, before his consecratio, which according to the canon law of the time was the true commencement of his pontificate, his name is not registered in the Liber Pontificalis nor in other lists of the popes. Published every year by the Roman Curia, the Annuario Pontificio attaches no consecutive numbers to the popes, stating that it is impossible to decide which side represented at various times the legitimate succession, in particular regarding Pope Leo VIII, Pope Benedict V and some mid-11th-century popes.
[[File:San Andreas Fault Aerial View.gif|thumb|right|The San Andreas Fault, a right-lateral strike-slip fault caused the massive 1906 [[San Francisco]] earthquake]] [[File:Sanandreas.jpg|thumb|Map of the San Andreas Fault, showing relative motion]] The '''San Andreas Fault''' is a right-moving ('dextral') strike-slip [[Fault (geology)|fault]]. It marks the boundary between the North American Plate on the east and the Pacific Plate on the west. The fault was the cause of the 1906 [[San Francisco]] [[earthquake]]. It first appeared about 20 million years ago.<ref>Atwater T. 1970. Implications of plate tectonics for the Cenozoic tectonic evolution of western North America.</ref><ref>Powell R.E. and Weldon,R.J. 1992. Evolution of the San Andreas fault. ''Annual Reviews of Earth and Planetary Science'', '''20''', 431–468.</ref> The San Andreas Fault is a zone several miles wide with multiple strands. The main active strand runs both on and off-shore between [[Cape Mendocino]] and the [[Sea of Cortez]]. From Cape Mendocino, it runs offshore to [[Tomales Bay]]. Then it goes southward through [[Bolinas Lagoon]], just west of the [[San Francisco]] Peninsula, to come onshore again at [[Daly City]], through the hills of the Peninsula. [[Crystal Springs|Crystal]] cum reservoir is formed by the fault itself. In the Santa Cruz mountains, it bends slightly eastward. This is the site of the 1989 [[Loma Prieta]] Earthquake. The fault continues south through the historic mission at [[San Juan Bautista]] and through the town of [[Hollister, California|Hollister]] (where active creep can be seen to offset sidewalks and even houses). The Transverse Ranges north of [[Santa Barbara]] are formed by compression. Strands of the fault snake under the [[Los Angeles]] Basin. From there it connects into the active spreading under the Sea of Cortez. The San Andreas fault was discovered by Professor Andrew Lawson in 1895, who climbed into the faulted [[serpentinite]] [[well]] where the south tower of the [[Golden Gate Bridge]] was being poured. In spite of the extreme deformation of the serpentinite, Lawson declared the bridge perfectly safe. The events of the 1906 San Francisco earthquake would seem to support his conclusion: the ruptured strand was not the one he observed in the footprint of the bridge. == References == {{Reflist}} [[Category:Geography of California]] [[Category:Faults]] [[Category:San Francisco Bay Area]]
The San Andreas Fault is a continental right-lateral strike-slip transform fault that extends roughly 1,200 kilometers (750 mi) through the Californias. It forms the tectonic boundary between the Pacific Plate and the North American Plate. Traditionally, for scientific purposes, the fault has been classified into three main segments (northern, central, and southern), each with different characteristics and a different degree of earthquake risk. The average slip rate along the entire fault ranges from 20 to 35 mm (0.79 to 1.38 in) per year. In the north, the fault terminates offshore near Eureka, California at the Mendocino Triple Junction, where three tectonic plates meet. It has been hypothesized that a major earthquake along the subduction zone could rupture the San Andreas Fault and vice versa. In the south, the fault terminates near Bombay Beach, California in the Salton Sea. Here, the plate motion is being reorganized from right-lateral to divergent. In this region (known as the Salton Trough), the plate boundary has been rifting and pulling apart, creating a new mid-ocean ridge that is an extension of the Gulf of California. Sediment deposited by the Colorado River is preventing the trough from being filled in with sea water from the gulf. The fault was first identified in 1895 by Professor Andrew Lawson of UC Berkeley. In the wake of the 1906 San Francisco earthquake, Lawson was tasked with deciphering the origin of the earthquake. He began by surveying and mapping offsets (such as fences or roads that had been sliced in half) along surface ruptures. When the location of these offsets were plotted on a map, he noted that they made a near perfect line on top of the fault he previously discovered. He concluded that the fault must have been the origin of the earthquake. This line ran through San Andreas Lake, a sag pond. The lake was created from an extensional step over in the fault, which created a natural depression where water could settle. A common misconception is that Lawson named the fault after this lake. However, according to some of his reports from 1895 and 1908, he actually named it after the surrounding San Andreas Valley. Following the 1906 San Francisco earthquake, Lawson also concluded that the fault extended all the way into Southern California. In 1953, geologist Thomas Dibblee concluded that hundreds of miles of lateral movement could occur along the fault. An NSF funded project called the San Andreas Fault Observatory at Depth (SAFOD) near Parkfield, California, involved drilling through the fault from 2004 to 2007. The aim was to collect core samples and make direct geophysical and geochemical observations to better understand fault behavior at depth. The northern segment of the fault runs from Hollister, through the Santa Cruz Mountains, epicenter of the 1989 Loma Prieta earthquake, then up the San Francisco Peninsula, where it was first identified by Professor Lawson in 1895, then offshore at Daly City near Mussel Rock. This is the approximate location of the epicenter of the 1906 San Francisco earthquake. The fault returns onshore at Bolinas Lagoon just north of Stinson Beach in Marin County. It returns underwater through the linear trough of Tomales Bay which separates the Point Reyes Peninsula from the mainland, runs just east of Bodega Head through Bodega Bay and back underwater, returning onshore at Fort Ross. (In this region around the San Francisco Bay Area several significant "sister faults" run more-or-less parallel, and each of these can create significantly destructive earthquakes.) From Fort Ross, the northern segment continues overland, forming in part a linear valley through which the Gualala River flows. It goes back offshore at Point Arena. After that, it runs underwater along the coast until it nears Cape Mendocino, where it begins to bend to the west, terminating at the Mendocino Triple Junction. The central segment of the San Andreas Fault runs in a northwestern direction from Parkfield to Hollister. While the southern section of the fault and the parts through Parkfield experience earthquakes, the rest of the central section of the fault exhibits a phenomenon called aseismic creep, where the fault slips continuously without causing earthquakes. It was formed by a transform boundary. The southern segment (also known as the Mojave segment) begins near Bombay Beach, California. Box Canyon, near the Salton Sea, contains upturned strata associated with that section of the fault. The fault then runs along the southern base of the San Bernardino Mountains, crosses through the Cajon Pass and continues northwest along the northern base of the San Gabriel Mountains. These mountains are a result of movement along the San Andreas Fault and are commonly called the Transverse Range. In Palmdale, a portion of the fault is easily examined at a roadcut for the Antelope Valley Freeway. The fault continues northwest alongside the Elizabeth Lake Road to the town of Elizabeth Lake. As it passes the towns of Gorman, Tejon Pass and Frazier Park, the fault begins to bend northward, forming the "Big Bend". This restraining bend is thought to be where the fault locks up in Southern California, with an earthquake-recurrence interval of roughly 140–160 years. Northwest of Frazier Park, the fault runs through the Carrizo Plain, a long, treeless plain where much of the fault is plainly visible. The Elkhorn Scarp defines the fault trace along much of its length within the plain. The southern segment, which stretches from Parkfield in Monterey County all the way to the Salton Sea, is capable of an 8.1-magnitude earthquake. At its closest, this fault passes about 35 miles (56 km) to the northeast of Los Angeles. Such a large earthquake on this southern segment would kill thousands of people in Los Angeles, San Bernardino, Riverside, and surrounding areas, and cause hundreds of billions of dollars in damage. The Pacific Plate, to the west of the fault, is moving in a northwest direction while the North American Plate to the east is moving toward the southwest, but relatively southeast under the influence of plate tectonics. The rate of slippage averages about 33 to 37 millimeters (1.3 to 1.5 in) a year across California. The southwestward motion of the North American Plate towards the Pacific is creating compressional forces along the eastern side of the fault. The effect is expressed as the Coast Ranges. The northwest movement of the Pacific Plate is also creating significant compressional forces which are especially pronounced where the North American Plate has forced the San Andreas to jog westward. This has led to the formation of the Transverse Ranges in Southern California, and to a lesser but still significant extent, the Santa Cruz Mountains (the location of the Loma Prieta earthquake in 1989). Studies of the relative motions of the Pacific and North American plates have shown that only about 75 percent of the motion can be accounted for in the movements of the San Andreas and its various branch faults. The rest of the motion has been found in an area east of the Sierra Nevada mountains called the Walker Lane or Eastern California Shear Zone. The reason for this is not clear. Several hypotheses have been offered and research is ongoing. One hypothesis – which gained interest following the Landers earthquake in 1992 – suggests the plate boundary may be shifting eastward away from the San Andreas towards Walker Lane. Assuming the plate boundary does not change as hypothesized, projected motion indicates that the landmass west of the San Andreas Fault, including Los Angeles, will eventually slide past San Francisco, then continue northwestward toward the Aleutian Trench, over a period of perhaps twenty million years. The San Andreas began to form in the mid Cenozoic about 30 Mya (million years ago). At this time, a spreading center between the Pacific Plate and the Farallon Plate (which is now mostly subducted, with remnants including the Juan de Fuca Plate, Rivera Plate, Cocos Plate, and the Nazca Plate) was beginning to reach the subduction zone off the western coast of North America. As the relative motion between the Pacific and North American Plates was different from the relative motion between the Farallon and North American Plates, the spreading ridge began to be "subducted", creating a new relative motion and a new style of deformation along the plate boundaries. These geological features are what are chiefly seen along San Andreas Fault. It also includes a possible driver for the deformation of the Basin and Range, separation of the Baja California Peninsula, and rotation of the Transverse Range. The main southern section of the San Andreas Fault proper has only existed for about 5 million years. The first known incarnation of the southern part of the fault was Clemens Well-Fenner-San Francisquito fault zone around 22–13 Ma. This system added the San Gabriel Fault as a primary focus of movement between 10–5 Ma. Currently, it is believed that the modern San Andreas will eventually transfer its motion toward a fault within the Eastern California Shear Zone. This complicated evolution, especially along the southern segment, is mostly caused by either the "Big Bend" and/or a difference in the motion vector between the plates and the trend of the fault and its surrounding branches. The fault was first identified in Northern California by UC Berkeley geology professor Andrew Lawson in 1895 and named by him after the surrounding San Andreas valley. Eleven years later, Lawson discovered that the San Andreas Fault stretched southward into southern California after reviewing the effects of the 1906 San Francisco earthquake. Large-scale (hundreds of miles) lateral movement along the fault was first proposed in a 1953 paper by geologists Mason Hill and Thomas Dibblee. This idea, which was considered radical at the time, has since been vindicated by modern plate tectonics. Seismologists discovered that the San Andreas Fault near Parkfield in central California consistently produces a magnitude 6.0 earthquake approximately once every 22 years. Following recorded seismic events in 1857, 1881, 1901, 1922, 1934, and 1966, scientists predicted that another earthquake should occur in Parkfield in 1993. It eventually occurred in 2004. Due to the frequency of predictable activity, Parkfield has become one of the most important areas in the world for large earthquake research. In 2004, work began just north of Parkfield on the San Andreas Fault Observatory at Depth (SAFOD). The goal of SAFOD is to drill a hole nearly 3 kilometres (1.9 mi) into the Earth's crust and into the San Andreas Fault. An array of sensors will be installed to record earthquakes that happen near this area. A 2023 study found a link between the water level in Lake Cahuilla (now the Salton Sea) and seismic activity along the southern San Andreas Fault. The study suggests that major earthquakes along this section of the fault coincided with high water levels in the lake. The hydrological load caused by high water levels can more than double the stress on the southern San Andreas Fault, which is likely sufficient for triggering earthquakes. This may explain the abnormally long period of time since the last major earthquake in the region since the lake has dried up. The San Andreas Fault System has been the subject of a flood of studies. In particular, scientific research performed during the last 23 years has given rise to about 3,400 publications. A study published in 2006 in the journal Nature by Yuri Fialko, an associate professor at the Cecil H. and Ida M. Green Institute of Geophysics and Planetary Physics at the Scripps Institution of Oceanography, found that the San Andreas fault has reached a sufficient stress level for an earthquake of magnitude greater than 7.0 on the moment magnitude scale to occur. This study also found that the risk of a large earthquake may be increasing more rapidly than scientists had previously believed. Moreover, the risk is currently concentrated on the southern section of the fault, i.e. the region around Los Angeles, because strong earthquakes have occurred relatively recently on the central (1857) and northern (1906) segments of the fault, while the southern section has not seen any similar rupture for at least 300 years. According to this study, a major earthquake on that southern section of the San Andreas fault would result in major damage to the Palm Springs–Indio metropolitan area and other cities in San Bernardino, Riverside and Imperial counties in California, and Mexicali Municipality in Baja California. It would be strongly felt (and potentially cause significant damage) throughout much of Southern California, including densely populated areas of Los Angeles County, Ventura County, Orange County, San Diego County, Ensenada Municipality and Tijuana Municipality, Baja California, San Luis Rio Colorado in Sonora and Yuma, Arizona. Older buildings would be especially prone to damage or collapse, as would buildings built on unconsolidated gravel or in coastal areas where water tables are high (and thus subject to soil liquefaction). Of the study, Fialko stated: All these data suggest that the fault is ready for the next big earthquake but exactly when the triggering will happen and when the earthquake will occur we cannot tell. It could be tomorrow or it could be 10 years or more from now. Nevertheless, in the 17 years since that publication there has not been a substantial quake in the Los Angeles area, and two major reports issued by the U.S. Geological Survey (USGS) have made variable predictions as to the risk of future seismic events. The ability to predict major earthquakes with sufficient precision to warrant increased precautions has remained elusive. The U.S. Geological Survey's most recent forecast, known as UCERF3 (Uniform California Earthquake Rupture Forecast 3), released in November 2013, estimated that an earthquake of magnitude 6.7 M or greater (i.e. equal to or greater than the 1994 Northridge earthquake) occurs about once every 6.7 years statewide. The same report also estimated there is a 7% probability that an earthquake of magnitude 8.0 or greater will occur in the next 30 years somewhere along the San Andreas Fault. A different USGS study in 2008 tried to assess the physical, social and economic consequences of a major earthquake in southern California. That study predicted that a magnitude 7.8 earthquake along the southern San Andreas Fault could cause about 1,800 deaths and $213 billion in damage. A 2008 paper, studying past earthquakes along the Pacific coastal zone, found a correlation in time between seismic events on the northern San Andreas Fault and the southern part of the Cascadia subduction zone (which stretches from Vancouver Island to northern California). Scientists believe quakes on the Cascadia subduction zone may have triggered most of the major quakes on the northern San Andreas within the past 3,000 years. The evidence also shows the rupture direction going from north to south in each of these time-correlated events. However the 1906 San Francisco earthquake seems to have been the exception to this correlation because the plate movement was mostly from south to north and it was not preceded by a major quake in the Cascadia zone. The San Andreas Fault has had some notable earthquakes in historic times:
[[File:3-Tasten-Maus Microsoft.jpg|thumb|Many mice have two buttons and a scroll wheel.]] A '''computer mouse''' is an [[input device]] which is used with a [[personal computer|computer]] or a laptop. Moving a mouse along a flat [[surface]] can move the [[cursor]] to different items on the [[Computer monitor|screen]]. ==Buttons== Items can be moved or selected by pushing the mouse [[Push-button|buttons]] (called [[wikt:click|clicking]]).<ref>{{cite web|url=http://foldoc.org/mouse|title=mouse from FOLDOC|website=foldoc.org}}</ref> Today's mouse has two buttons, the left button and right button, with a [[#Uses|scroll wheel]] in between the two. Many computer mice use [[wireless]] technology and have no wire. ==Actions== Today, this device is on almost every [[desktop computer]]. The main function is to move the [[Pointer (computing)|mouse pointer]] on the screen. The mouse pointer can be used for many actions on the computer. Clicking or double clicking an [[Computer icon|icon]], [[Folder (computing)|folder]], or other object will open a document or start a program. Users can also move an icon, folder, or other object by clicking a mouse button and dragging the object with the mouse pointer. Pointing to an item on the screen means moving your mouse so the pointer appears to be touching the item. When you point to something, a small box often appears that describes the item. ==Types== There are many types of mouse: Optical mouse, wireless mouse, mechanical mouse, trackball mouse. A computer mouse is a handheld hardware [[input device]] that controls a cursor in a GUI and can move and select text, icons, files, and folders. For desktop computers, the mouse is placed on a flat surface such as a mouse pad or a desk and is placed in front of the computer. The mouse was originally known as the X-Y Position Indicator for a display system and was invented by [[Douglas Engelbart]] in 1963 while working at [[Xerox PARC]]. However, due to Alto's lack of success, the first widely used application of the mouse was with the [[Apple Lisa]] computer. == History == In 1964 [[Douglas Engelbart]] (1925-2013), a researcher at [[Stanford University|Stanford Research Institute]], wanted to find a way to make using computers easier. In those days, computers were large and expensive. Using them was very hard because everything had to be typed in on a [[keyboard (computer)|keyboard]]. This [[command line interface]] is still used by some people, such as [[programmer]]s, to get things done faster. After studying and designing for a long time, Engelbart succeeded in inventing an input device which he named 'XY index'. At first, it needed two [[hand]]s to use, but it was changed so that only one hand was needed to use it. This model was more like the mouse that we use today, but was made up of [[Trackball|a big ball]] that the user had to roll in different [[direction]]s to move the cursor. [[File:Apple II mouse.jpg|thumb|left|Early Apple II mouse with a single button]] The computer mouse began to be widely used when [[Xerox]] Palo Alto Research introduced a [[Graphical user interfaces and consoles|GUI]] in 1981, where the mouse was used to click things on the screen. This was also the case with the [[Apple Macintosh|Macintosh]] [[operating system]] from [[Apple Inc.|Apple]]<ref>[http://computer.howstuffworks.com/mouse.htm/printable How computer mice work]: "... the mouse hit the scene - attached to the Mac..."</ref> of [[Apple Inc]] when it came out in 1984, as well as [[Microsoft Windows]]. Windows became popular over the years, so over time computer mice became used with many computers. In 1991, the company [[Logitech]] invented the wireless mouse.<ref name="logitech-macworld">{{cite web | url=https://www.macworld.com/article/1137400/mouse40.html | title=The computer mouse turns 40 }}</ref> Unlike a normal mouse, wireless mice were connected by radio signals.<ref name="logitech-macworld" /> Newer wireless mice use [[Bluetooth]] or [[WiFi]] to connect wireless mice to computers. == Uses == On most computers, the user can move the mouse to move the cursor in the same direction.<ref>{{cite web|url=http://wordnetweb.princeton.edu/perl/webwn?s=computer+mouse|title=WordNet Search - 3.1|website=wordnetweb.princeton.edu}}</ref> To choose something that is on the screen, the user can move the cursor to it and "click" the left mouse button. The right button is used to open menus that are different depending on where the cursor is. The other mouse buttons can do different things, depending on the [[software]]. Most mice have two buttons to click. Most mice also have the "scroll wheel"—a small wheel found between the two main mouse buttons. The user can move the wheel back and forth to "scroll" through things like a [[website]] or folder. "Scrolling" means moving the words or pictures up or down on the screen, so another part of the page comes into view. The wheel can also be pressed, to click it like another button. A mouse can also be connected to and used the same way with a [[laptop]] computer, but unlike a [[desktop computer]], one does not have to be connected to use the laptop. This is because along with the keyboard, laptops have a built-in input device called a [[trackpad]] which does the same thing as a mouse. Similarly, [[tablet computer]]s have a [[touchscreen]] as an input device, but some, like the [[Microsoft Surface]] and those that use [[Android (operating system)|Android]], also work with mice. == Related pages == * [[Touchpad]] * [[Trackball]] * [[Touchscreen]] == References == {{reflist}} == Other websites == * [http://en.citizendium.org/wiki/Mouse_(computing) Mouse (computing)] Citizendium {{Tech-stub}} [[Category:Computer hardware]] [[Category:Data input]]
A computer mouse (plural mice, also mouses) is a hand-held pointing device that detects two-dimensional motion relative to a surface. This motion is typically translated into the motion of the pointer (called a cursor) on a display, which allows a smooth control of the graphical user interface of a computer. The first public demonstration of a mouse controlling a computer system was done by Doug Engelbart in 1968 as part of the Mother of All Demos. Mice originally used two separate wheels to directly track movement across a surface: one in the x-dimension and one in the Y. Later, the standard design shifted to use a ball rolling on a surface to detect motion, in turn connected to internal rollers. Most modern mice use optical movement detection with no moving parts. Though originally all mice were connected to a computer by a cable, many modern mice are cordless, relying on short-range radio communication with the connected system. In addition to moving a cursor, computer mice have one or more buttons to allow operations such as the selection of a menu item on a display. Mice often also feature other elements, such as touch surfaces and scroll wheels, which enable additional control and dimensional input. The earliest known written use of the term mouse or mice in reference to a computer pointing device is in Bill English's July 1965 publication, "Computer-Aided Display Control". This likely originated from its resemblance to the shape and size of a mouse, with the cord resembling its tail. The popularity of wireless mice without cords makes the resemblance less obvious. According to Roger Bates, a hardware designer under English, the term also came about because the cursor on the screen was for some unknown reason referred to as "CAT" and was seen by the team as if it would be chasing the new desktop device. The plural for the small rodent is always "mice" in modern usage. The plural for a computer mouse is either "mice" or "mouses" according to most dictionaries, with "mice" being more common. The first recorded plural usage is "mice"; the online Oxford Dictionaries cites a 1984 use, and earlier uses include J. C. R. Licklider's "The Computer as a Communication Device" of 1968. The trackball, a related pointing device, was invented in 1946 by Ralph Benjamin as part of a post-World War II-era fire-control radar plotting system called the Comprehensive Display System (CDS). Benjamin was then working for the British Royal Navy Scientific Service. Benjamin's project used analog computers to calculate the future position of target aircraft based on several initial input points provided by a user with a joystick. Benjamin felt that a more elegant input device was needed and invented what they called a "roller ball" for this purpose. The device was patented in 1947, but only a prototype using a metal ball rolling on two rubber-coated wheels was ever built, and the device was kept as a military secret. Another early trackball was built by Kenyon Taylor, a British electrical engineer working in collaboration with Tom Cranston and Fred Longstaff. Taylor was part of the original Ferranti Canada, working on the Royal Canadian Navy's DATAR (Digital Automated Tracking and Resolving) system in 1952. DATAR was similar in concept to Benjamin's display. The trackball used four disks to pick up motion, two each for the X and Y directions. Several rollers provided mechanical support. When the ball was rolled, the pickup discs spun and contacts on their outer rim made periodic contact with wires, producing pulses of output with each movement of the ball. By counting the pulses, the physical movement of the ball could be determined. A digital computer calculated the tracks and sent the resulting data to other ships in a task force using pulse-code modulation radio signals. This trackball used a standard Canadian five-pin bowling ball. It was not patented, since it was a secret military project. Douglas Engelbart of the Stanford Research Institute (now SRI International) has been credited in published books by Thierry Bardini, Paul Ceruzzi, Howard Rheingold, and several others as the inventor of the computer mouse. Engelbart was also recognized as such in various obituary titles after his death in July 2013. By 1963, Engelbart had already established a research lab at SRI, the Augmentation Research Center (ARC), to pursue his objective of developing both hardware and software computer technology to "augment" human intelligence. That November, while attending a conference on computer graphics in Reno, Nevada, Engelbart began to ponder how to adapt the underlying principles of the planimeter to inputting X- and Y-coordinate data. On 14 November 1963, he first recorded his thoughts in his personal notebook about something he initially called a "bug", which is a "3-point" form could have a "drop point and 2 orthogonal wheels". He wrote that the "bug" would be "easier" and "more natural" to use, and unlike a stylus, it would stay still when let go, which meant it would be "much better for coordination with the keyboard". In 1964, Bill English joined ARC, where he helped Engelbart build the first mouse prototype. They christened the device the mouse as early models had a cord attached to the rear part of the device which looked like a tail, and in turn, resembled the common mouse. According to Roger Bates, a hardware designer in English, another reason for choosing this name was because the cursor on the screen was also referred to as "CAT" at this time. As noted above, this "mouse" was first mentioned in print in a July 1965 report, on which English was the lead author. On 9 December 1968, Engelbart publicly demonstrated the mouse at what would come to be known as The Mother of All Demos. Engelbart never received any royalties for it, as his employer SRI held the patent, which expired before the mouse became widely used in personal computers. In any event, the invention of the mouse was just a small part of Engelbart's much larger project of augmenting human intellect. Several other experimental pointing-devices developed for Engelbart's oN-Line System (NLS) exploited different body movements – for example, head-mounted devices attached to the chin or nose – but ultimately the mouse won out because of its speed and convenience. The first mouse, a bulky device (pictured) used two potentiometers perpendicular to each other and connected to wheels: the rotation of each wheel translated into motion along one axis. At the time of the "Mother of All Demos", Engelbart's group had been using their second-generation, 3-button mouse for about a year. On 2 October 1968, three years after Engelbart's prototype but more than two months before his public demo, a mouse device named Rollkugelsteuerung (German for "Trackball control") was shown in a sales brochure by the German company AEG-Telefunken as an optional input device for the SIG 100 vector graphics terminal, part of the system around their process computer TR 86 and the TR 440 [de] main frame. Based on an even earlier trackball device, the mouse device had been developed by the company in 1966 in what had been a parallel and independent discovery. As the name suggests and unlike Engelbart's mouse, the Telefunken model already had a ball (diameter 40 mm, weight 40 g) and two mechanical 4-bit rotational position transducers with Gray code-like states, allowing easy movement in any direction. The bits remained stable for at least two successive states to relax debouncing requirements. This arrangement was chosen so that the data could also be transmitted to the TR 86 front-end process computer and over longer distance telex lines with c. 50 baud. Weighing 465 grams (16.4 oz), the device with a total height of about 7 cm (2.8 in) came in a c. 12 cm (4.7 in) diameter hemispherical injection-molded thermoplastic casing featuring one central push button. As noted above, the device was based on an earlier trackball-like device (also named Rollkugel) that was embedded into radar flight control desks. This trackball had been originally developed by a team led by Rainer Mallebrein [de] at Telefunken Konstanz for the German Bundesanstalt für Flugsicherung [de] (Federal Air Traffic Control). It was part of the corresponding workstation system SAP 300 and the terminal SIG 3001, which had been designed and developed since 1963. Development for the TR 440 main frame began in 1965. This led to the development of the TR 86 process computer system with its SIG 100-86 terminal. Inspired by a discussion with a university customer, Mallebrein came up with the idea of "reversing" the existing Rollkugel trackball into a moveable mouse-like device in 1966, so that customers did not have to be bothered with mounting holes for the earlier trackball device. The device was finished in early 1968, and together with light pens and trackballs, it was commercially offered as an optional input device for their system starting later that year. Not all customers opted to buy the device, which added costs of DM 1,500 per piece to the already up to 20-million DM deal for the main frame, of which only a total of 46 systems were sold or leased. They were installed at more than 20 German universities including RWTH Aachen, Technical University Berlin, University of Stuttgart and Konstanz. Several Rollkugel mice installed at the Leibniz Supercomputing Centre in Munich in 1972 are well preserved in a museum, two others survived in a museum at Stuttgart University, two in Hamburg, the one from Aachen at the Computer History Museum in the US, and yet another sample was recently donated to the Heinz Nixdorf MuseumsForum (HNF) in Paderborn. Anecdotal reports claim that Telefunken's attempt to patent the device was rejected by the German Patent Office due to lack of inventiveness. For the air traffic control system, the Mallebrein team had already developed a precursor to touch screens in form of an ultrasonic-curtain-based pointing device in front of the display. In 1970, they developed a device named "Touchinput-Einrichtung" ("touch input device") based on a conductively coated glass screen. The Xerox Alto was one of the first computers designed for individual use in 1973 and is regarded as the first modern computer to use a mouse. Inspired by PARC's Alto, the Lilith, a computer which had been developed by a team around Niklaus Wirth at ETH Zürich between 1978 and 1980, provided a mouse as well. The third marketed version of an integrated mouse shipped as a part of a computer and intended for personal computer navigation came with the Xerox 8010 Star in 1981. By 1982, the Xerox 8010 was probably the best-known computer with a mouse. The Sun-1 also came with a mouse, and the forthcoming Apple Lisa was rumored to use one, but the peripheral remained obscure; Jack Hawley of The Mouse House reported that one buyer for a large organization believed at first that his company sold lab mice. Hawley, who manufactured mice for Xerox, stated that "Practically, I have the market all to myself right now"; a Hawley mouse cost $415. In 1982, Logitech introduced the P4 Mouse at the Comdex trade show in Las Vegas, its first hardware mouse. That same year Microsoft made the decision to make the MS-DOS program Microsoft Word mouse-compatible, and developed the first PC-compatible mouse. Microsoft's mouse shipped in 1983, thus beginning the Microsoft Hardware division of the company. However, the mouse remained relatively obscure until the appearance of the Macintosh 128K (which included an updated version of the single-button Lisa Mouse) in 1984, and of the Amiga 1000 and the Atari ST in 1985. A mouse typically controls the motion of a pointer in two dimensions in a graphical user interface (GUI). The mouse turns movements of the hand backward and forward, left and right into equivalent electronic signals that in turn are used to move the pointer. The relative movements of the mouse on the surface are applied to the position of the pointer on the screen, which signals the point where actions of the user take place, so hand movements are replicated by the pointer. Clicking or pointing (stopping movement while the cursor is within the bounds of an area) can select files, programs or actions from a list of names, or (in graphical interfaces) through small images called "icons" and other elements. For example, a text file might be represented by a picture of a paper notebook and clicking while the cursor points at this icon might cause a text editing program to open the file in a window. Different ways of operating the mouse cause specific things to happen in the GUI: The Concept of Gestural Interfaces Gestural interfaces have become an integral part of modern computing, allowing users to interact with their devices in a more intuitive and natural way. In addition to traditional pointing-and-clicking actions, users can now employ gestural inputs to issue commands or perform specific actions. These stylized motions of the mouse cursor, known as "gestures", have the potential to enhance user experience and streamline workflow. Mouse Gestures in Action To illustrate the concept of gestural interfaces, let's consider a drawing program as an example. In this scenario, a user can employ a gesture to delete a shape on the canvas. By rapidly moving the mouse cursor in an "x" motion over the shape, the user can trigger the command to delete the selected shape. This gesture-based interaction enables users to perform actions quickly and efficiently without relying solely on traditional input methods. Challenges and Benefits of Gestural Interfaces While gestural interfaces offer a more immersive and interactive user experience, they also present challenges. One of the primary difficulties lies in the requirement of finer motor control from users. Gestures demand precise movements, which can be more challenging for individuals with limited dexterity or those who are new to this mode of interaction. However, despite these challenges, gestural interfaces have gained popularity due to their ability to simplify complex tasks and improve efficiency. Several gestural conventions have become widely adopted, making them more accessible to users. One such convention is the drag and drop gesture, which has become pervasive across various applications and platforms. The Drag and Drop Gesture The drag and drop gesture is a fundamental gestural convention that enables users to manipulate objects on the screen seamlessly. It involves a series of actions performed by the user: Pressing the mouse button while the cursor hovers over an interface object. Moving the cursor to a different location while holding the button down. Releasing the mouse button to complete the action. This gesture allows users to transfer or rearrange objects effortlessly. For instance, a user can drag and drop a picture representing a file onto an image of a trash can, indicating the intention to delete the file. This intuitive and visual approach to interaction has become synonymous with organizing digital content and simplifying file management tasks. Standard Semantic Gestures In addition to the drag and drop gesture, several other semantic gestures have emerged as standard conventions within the gestural interface paradigm. These gestures serve specific purposes and contribute to a more intuitive user experience. Some of the notable semantic gestures include: Crossing-based goal: This gesture involves crossing a specific boundary or threshold on the screen to trigger an action or complete a task. For example, swiping across the screen to unlock a device or confirm a selection. Menu traversal: Menu traversal gestures facilitate navigation through hierarchical menus or options. Users can perform gestures such as swiping or scrolling to explore different menu levels or activate specific commands. Pointing: Pointing gestures involve positioning the mouse cursor over an object or element to interact with it. This fundamental gesture enables users to select, click, or access contextual menus. Mouseover (pointing or hovering): Mouseover gestures occur when the cursor is positioned over an object without clicking. This action often triggers a visual change or displays additional information about the object, providing users with real-time feedback. These standard semantic gestures, along with the drag and drop convention, form the building blocks of gestural interfaces, allowing users to interact with digital content using intuitive and natural movements. At the end of 20th century Digitizer mouse with Magnifying glass was used with AutoCAD for digitizations of blue-prints. Other uses of the mouse's input occur commonly in special application domains. In interactive three-dimensional graphics, the mouse's motion often translates directly into changes in the virtual objects' or camera's orientation. For example, in the first-person shooter genre of games (see below), players usually employ the mouse to control the direction in which the virtual player's "head" faces: moving the mouse up will cause the player to look up, revealing the view above the player's head. A related function makes an image of an object rotate so that all sides can be examined. 3D design and animation software often modally chord many different combinations to allow objects and cameras to be rotated and moved through space with the few axes of movement mice can detect. When mice have more than one button, the software may assign different functions to each button. Often, the primary (leftmost in a right-handed configuration) button on the mouse will select items, and the secondary (rightmost in a right-handed) button will bring up a menu of alternative actions applicable to that item. For example, on platforms with more than one button, the Mozilla web browser will follow a link in response to a primary button click, will bring up a contextual menu of alternative actions for that link in response to a secondary-button click, and will often open the link in a new tab or window in response to a click with the tertiary (middle) mouse button. The German company Telefunken published on their early ball mouse on 2 October 1968. Telefunken's mouse was sold as optional equipment for their computer systems. Bill English, builder of Engelbart's original mouse, created a ball mouse in 1972 while working for Xerox PARC. The ball mouse replaced the external wheels with a single ball that could rotate in any direction. It came as part of the hardware package of the Xerox Alto computer. Perpendicular chopper wheels housed inside the mouse's body chopped beams of light on the way to light sensors, thus detecting in their turn the motion of the ball. This variant of the mouse resembled an inverted trackball and became the predominant form used with personal computers throughout the 1980s and 1990s. The Xerox PARC group also settled on the modern technique of using both hands to type on a full-size keyboard and grabbing the mouse when required. The ball mouse has two freely rotating rollers. These are located 90 degrees apart. One roller detects the forward-backward motion of the mouse and the other the left-right motion. Opposite the two rollers is a third one (white, in the photo, at 45 degrees) that is spring-loaded to push the ball against the other two rollers. Each roller is on the same shaft as an encoder wheel that has slotted edges; the slots interrupt infrared light beams to generate electrical pulses that represent wheel movement. Each wheel's disc has a pair of light beams, located so that a given beam becomes interrupted or again starts to pass light freely when the other beam of the pair is about halfway between changes. Simple logic circuits interpret the relative timing to indicate which direction the wheel is rotating. This incremental rotary encoder scheme is sometimes called quadrature encoding of the wheel rotation, as the two optical sensors produce signals that are in approximately quadrature phase. The mouse sends these signals to the computer system via the mouse cable, directly as logic signals in very old mice such as the Xerox mice, and via a data-formatting IC in modern mice. The driver software in the system converts the signals into motion of the mouse cursor along X and Y axes on the computer screen. The ball is mostly steel, with a precision spherical rubber surface. The weight of the ball, given an appropriate working surface under the mouse, provides a reliable grip so the mouse's movement is transmitted accurately. Ball mice and wheel mice were manufactured for Xerox by Jack Hawley, doing business as The Mouse House in Berkeley, California, starting in 1975. Based on another invention by Jack Hawley, proprietor of the Mouse House, Honeywell produced another type of mechanical mouse. Instead of a ball, it had two wheels rotating at off axes. Key Tronic later produced a similar product. Modern computer mice took form at the École Polytechnique Fédérale de Lausanne (EPFL) under the inspiration of Professor Jean-Daniel Nicoud and at the hands of engineer and watchmaker André Guignard. This new design incorporated a single hard rubber mouseball and three buttons, and remained a common design until the mainstream adoption of the scroll-wheel mouse during the 1990s. In 1985, René Sommer added a microprocessor to Nicoud's and Guignard's design. Through this innovation, Sommer is credited with inventing a significant component of the mouse, which made it more "intelligent"; though optical mice from Mouse Systems had incorporated microprocessors by 1984. Another type of mechanical mouse, the "analog mouse" (now generally regarded as obsolete), uses potentiometers rather than encoder wheels, and is typically designed to be plug compatible with an analog joystick. The "Color Mouse", originally marketed by RadioShack for their Color Computer (but also usable on MS-DOS machines equipped with analog joystick ports, provided the software accepted joystick input) was the best-known example. Early optical mice relied entirely on one or more light-emitting diodes (LEDs) and an imaging array of photodiodes to detect movement relative to the underlying surface, eschewing the internal moving parts a mechanical mouse uses in addition to its optics. A laser mouse is an optical mouse that uses coherent (laser) light. The earliest optical mice detected movement on pre-printed mousepad surfaces, whereas the modern LED optical mouse works on most opaque diffuse surfaces; it is usually unable to detect movement on specular surfaces like polished stone. Laser diodes provide good resolution and precision, improving performance on opaque specular surfaces. Later, more surface-independent optical mice use an optoelectronic sensor (essentially, a tiny low-resolution video camera) to take successive images of the surface on which the mouse operates. Battery powered, wireless optical mice flash the LED intermittently to save power, and only glow steadily when movement is detected. Often called "air mice" since they do not require a surface to operate, inertial mice use a tuning fork or other accelerometer (US Patent 4787051) to detect rotary movement for every axis supported. The most common models (manufactured by Logitech and Gyration) work using 2 degrees of rotational freedom and are insensitive to spatial translation. The user requires only small wrist rotations to move the cursor, reducing user fatigue or "gorilla arm". Usually cordless, they often have a switch to deactivate the movement circuitry between use, allowing the user freedom of movement without affecting the cursor position. A patent for an inertial mouse claims that such mice consume less power than optically based mice, and offer increased sensitivity, reduced weight and increased ease-of-use. In combination with a wireless keyboard an inertial mouse can offer alternative ergonomic arrangements which do not require a flat work surface, potentially alleviating some types of repetitive motion injuries related to workstation posture. A 3D mouse is a computer input device for viewport interaction with at least three degrees of freedom (DoF), e.g. in 3D computer graphics software for manipulating virtual objects, navigating in the viewport, defining camera paths, posing, and desktop motion capture. 3D mice can also be used as spatial controllers for video game interaction, e.g. SpaceOrb 360. To perform such different tasks the used transfer function and the device stiffness are essential for efficient interaction. The virtual motion is connected to the 3D mouse control handle via a transfer function. Position control means that the virtual position and orientation is proportional to the mouse handle's deflection whereas velocity control means that translation and rotation velocity of the controlled object is proportional to the handle deflection. A further essential property of a transfer function is its interaction metaphor: Ware and Osborne performed an experiment investigating these metaphors whereby it was shown that there is no single best metaphor. For manipulation tasks, the object-in-hand metaphor was superior, whereas for navigation tasks the camera-in-hand metaphor was superior. Zhai used and the following three categories for device stiffness: Logitech 3D Mouse (1990) was the first ultrasonic mouse and is an example of an isotonic 3D mouse having six degrees of freedom (6DoF). Isotonic devices have also been developed with less than 6DoF, e.g. the Inspector at Technical University of Denmark (5DoF input). Other examples of isotonic 3D mice are motion controllers, i.e. is a type of game controller that typically uses accelerometers to track motion. Motion tracking systems are also used for motion capture e.g. in the film industry, although that these tracking systems are not 3D mice in a strict sense, because motion capture only means recording 3D motion and not 3D interaction. Early 3D mice for velocity control were almost ideally isometric, e.g. SpaceBall 1003, 2003, 3003, and a device developed at Deutsches Zentrum für Luft und Raumfahrt (DLR), cf. US patent US4589810A. At DLR an elastic 6DoF sensor was developed that was used in Logitech's SpaceMouse and in the products of 3DConnexion. SpaceBall 4000 FLX has a maximum deflection of approximately 3 mm (0.12 in) at a maximum force of approximately 10N, that is, a stiffness of approximately 33 N/cm (19 lbf/in). SpaceMouse has a maximum deflection of 1.5 mm (0.059 in) at a maximum force of 4.4 N (0.99 lbf), that is, a stiffness of approximately 30 N/cm (17 lbf/in). Taking this development further, the softly elastic Sundinlabs SpaceCat was developed. SpaceCat has a maximum translational deflection of approximately 15 mm (0.59 in) and maximum rotational deflection of approximately 30° at a maximum force less than 2N, that is, a stiffness of approximately 1.3 N/cm (0.74 lbf/in). With SpaceCat Sundin and Fjeld reviewed five comparative experiments performed with different device stiffness and transfer functions and performed a further study comparing 6DoF softly elastic position control with 6DoF stiffly elastic velocity control in a positioning task. They concluded that for positioning tasks position control is to be preferred over velocity control. They could further conjecture the following two types of preferred 3D mouse usage: 3DConnexion's 3D mice have been commercially successful over decades. They are used in combination with the conventional mouse for CAD. The Space Mouse is used to orient the target object or change the viewpoint with the non-dominant hand, whereas the dominant hand operates the computer mouse for conventional CAD GUI operation. This is a kind of space-multiplexed input where the 6 DoF input device acts as a graspable user interface that is always connected to the view port. In November 2010 a German Company called Axsotic introduced a new concept of 3D mouse called 3D Spheric Mouse. This new concept of a true six degree-of-freedom input device uses a ball to rotate isometrically in 3 axes and an elastic polymer anchored tetrahedron inspired suspension for translating the ball without any limitations. A contactless sensor design uses a magnetic sensor array for sensing three aches translation and two optical mouse sensors for three aches rotation. The special tetrahedron suspension allows a user to rotate the ball with the fingers while input translations with the hand-wrist motion. With force feedback the device stiffness can dynamically be adapted to the task just performed by the user, e.g. performing positioning tasks with less stiffness than navigation tasks. In 2000, Logitech introduced a "tactile mouse" known as the "iFeel Mouse" developed by Immersion Corporation that contained a small actuator to enable the mouse to generate simulated physical sensations. Such a mouse can augment user-interfaces with haptic feedback, such as giving feedback when crossing a window boundary. To surf the internet by touch-enabled mouse was first developed in 1996 and first implemented commercially by the Wingman Force Feedback Mouse. It requires the user to be able to feel depth or hardness; this ability was realized with the first electrorheological tactile mice but never marketed. Tablet digitizers are sometimes used with accessories called pucks, devices which rely on absolute positioning, but can be configured for sufficiently mouse-like relative tracking that they are sometimes marketed as mice. As the name suggests, this type of mouse is intended to provide optimum comfort and avoid injuries such as carpal tunnel syndrome, arthritis, and other repetitive strain injuries. It is designed to fit natural hand position and movements, to reduce discomfort. When holding a typical mouse, the ulna and radius bones on the arm are crossed. Some designs attempt to place the palm more vertically, so the bones take more natural parallel position. Increasing mouse height and angling the mouse topcase can improve wrist posture without negatively affecting performance. Some limit wrist movement, encouraging arm movement instead, that may be less precise but more optimal from the health point of view. A mouse may be angled from the thumb downward to the opposite side – this is known to reduce wrist pronation. However such optimizations make the mouse right or left hand specific, making more problematic to change the tired hand. Time has criticized manufacturers for offering few or no left-handed ergonomic mice: "Oftentimes I felt like I was dealing with someone who'd never actually met a left-handed person before." Another solution is a pointing bar device. The so-called roller bar mouse is positioned snugly in front of the keyboard, thus allowing bi-manual accessibility. These mice are specifically designed for use in computer games. They typically employ a wider array of controls and buttons and have designs that differ radically from traditional mice. They may also have decorative monochrome or programmable RGB LED lighting. The additional buttons can often be used for changing the sensitivity of the mouse or they can be assigned (programmed) to macros (i.e., for opening a program or for use instead of a key combination). It is also common for game mice, especially those designed for use in real-time strategy games such as StarCraft, or in multiplayer online battle arena games such as League of Legends to have a relatively high sensitivity, measured in dots per inch (DPI), which can be as high as 25,600. DPI and CPI are the same values that refer to the mouse's sensitivity. DPI is a misnomer used in the gaming world, and many manufacturers use it to refer to CPI, counts per inch. Some advanced mice from gaming manufacturers also allow users to adjust the weight of the mouse by adding or subtracting weights to allow for easier control. Ergonomic quality is also an important factor in gaming mouse, as extended gameplay times may render further use of the mouse to be uncomfortable. Some mice have been designed to have adjustable features such as removable and/or elongated palm rests, horizontally adjustable thumb rests and pinky rests. Some mice may include several different rests with their products to ensure comfort for a wider range of target consumers. Gaming mice are held by gamers in three styles of grip: To transmit their input, typical cabled mice use a thin electrical cord terminating in a standard connector, such as RS-232C, PS/2, ADB, or USB. Cordless mice instead transmit data via infrared radiation (see IrDA) or radio (including Bluetooth), although many such cordless interfaces are themselves connected through the aforementioned wired serial buses. While the electrical interface and the format of the data transmitted by commonly available mice is currently standardized on USB, in the past it varied between different manufacturers. A bus mouse used a dedicated interface card for connection to an IBM PC or compatible computer. Mouse use in DOS applications became more common after the introduction of the Microsoft Mouse, largely because Microsoft provided an open standard for communication between applications and mouse driver software. Thus, any application written to use the Microsoft standard could use a mouse with a driver that implements the same API, even if the mouse hardware itself was incompatible with Microsoft's. This driver provides the state of the buttons and the distance the mouse has moved in units that its documentation calls "mickeys". In the 1970s, the Xerox Alto mouse, and in the 1980s the Xerox optical mouse, used a quadrature-encoded X and Y interface. This two-bit encoding per dimension had the property that only one bit of the two would change at a time, like a Gray code or Johnson counter, so that the transitions would not be misinterpreted when asynchronously sampled. The earliest mass-market mice, such as on the original Macintosh, Amiga, and Atari ST mice used a D-subminiature 9-pin connector to send the quadrature-encoded X and Y axis signals directly, plus one pin per mouse button. The mouse was a simple optomechanical device, and the decoding circuitry was all in the main computer. The DE-9 connectors were designed to be electrically compatible with the joysticks popular on numerous 8-bit systems, such as the Commodore 64 and the Atari 2600. Although the ports could be used for both purposes, the signals must be interpreted differently. As a result, plugging a mouse into a joystick port causes the "joystick" to continuously move in some direction, even if the mouse stays still, whereas plugging a joystick into a mouse port causes the "mouse" to only be able to move a single pixel in each direction. Because the IBM PC did not have a quadrature decoder built in, early PC mice used the RS-232C serial port to communicate encoded mouse movements, as well as provide power to the mouse's circuits. The Mouse Systems Corporation version used a five-byte protocol and supported three buttons. The Microsoft version used a three-byte protocol and supported two buttons. Due to the incompatibility between the two protocols, some manufacturers sold serial mice with a mode switch: "PC" for MSC mode, "MS" for Microsoft mode. In 1986 Apple first implemented the Apple Desktop Bus allowing the daisy chaining of up to 16 devices, including mice and other devices on the same bus with no configuration whatsoever. Featuring only a single data pin, the bus used a purely polled approach to device communications and survived as the standard on mainstream models (including a number of non-Apple workstations) until 1998 when Apple's iMac line of computers joined the industry-wide switch to using USB. Beginning with the Bronze Keyboard PowerBook G3 in May 1999, Apple dropped the external ADB port in favor of USB, but retained an internal ADB connection in the PowerBook G4 for communication with its built-in keyboard and trackpad until early 2005. With the arrival of the IBM PS/2 personal-computer series in 1987, IBM introduced the eponymous PS/2 port for mice and keyboards, which other manufacturers rapidly adopted. The most visible change was the use of a round 6-pin mini-DIN, in lieu of the former 5-pin MIDI style full sized DIN 41524 connector. In default mode (called stream mode) a PS/2 mouse communicates motion, and the state of each button, by means of 3-byte packets. For any motion, button press or button release event, a PS/2 mouse sends, over a bi-directional serial port, a sequence of three bytes, with the following format: Here, XS and YS represent the sign bits of the movement vectors, XV and YV indicate an overflow in the respective vector component, and LB, MB and RB indicate the status of the left, middle and right mouse buttons (1 = pressed). PS/2 mice also understand several commands for reset and self-test, switching between different operating modes, and changing the resolution of the reported motion vectors. A Microsoft IntelliMouse relies on an extension of the PS/2 protocol: the ImPS/2 or IMPS/2 protocol (the abbreviation combines the concepts of "IntelliMouse" and "PS/2"). It initially operates in standard PS/2 format, for backward compatibility. After the host sends a special command sequence, it switches to an extended format in which a fourth byte carries information about wheel movements. The IntelliMouse Explorer works analogously, with the difference that its 4-byte packets also allow for two additional buttons (for a total of five). Mouse vendors also use other extended formats, often without providing public documentation. The Typhoon mouse uses 6-byte packets which can appear as a sequence of two standard 3-byte packets, such that an ordinary PS/2 driver can handle them. For 3-D (or 6-degree-of-freedom) input, vendors have made many extensions both to the hardware and to software. In the late 1990s, Logitech created ultrasound based tracking which gave 3D input to a few millimeters accuracy, which worked well as an input device but failed as a profitable product. In 2008, Motion4U introduced its "OptiBurst" system using IR tracking for use as a Maya (graphics software) plugin. Almost all wired mice today use USB and the USB human interface device class for communication. Cordless or wireless mice transmit data via radio. Some mice connect to the computer through Bluetooth or Wi-Fi, while others use a receiver that plugs into the computer, for example through a USB port. Many mice that use a USB receiver have a storage compartment for it inside the mouse. Some "nano receivers" are designed to be small enough to remain plugged into a laptop during transport, while still being large enough to easily remove. MS-DOS and Windows 1.0 support connecting a mouse such as a Microsoft Mouse via multiple interfaces: BallPoint, Bus (InPort), Serial port or PS/2. Windows 98 added built-in support for USB Human Interface Device class (USB HID), with native vertical scrolling support. Windows 2000 and Windows Me expanded this built-in support to 5-button mice. Windows XP Service Pack 2 introduced a Bluetooth stack, allowing Bluetooth mice to be used without any USB receivers. Windows Vista added native support for horizontal scrolling and standardized wheel movement granularity for finer scrolling. Windows 8 introduced BLE (Bluetooth Low Energy) mouse/HID support. Some systems allow two or more mice to be used at once as input devices. Late-1980s era home computers such as the Amiga used this to allow computer games with two players interacting on the same computer (Lemmings and The Settlers for example). The same idea is sometimes used in collaborative software, e.g. to simulate a whiteboard that multiple users can draw on without passing a single mouse around. Microsoft Windows, since Windows 98, has supported multiple simultaneous pointing devices. Because Windows only provides a single screen cursor, using more than one device at the same time requires cooperation of users or applications designed for multiple input devices. Multiple mice are often used in multi-user gaming in addition to specially designed devices that provide several input interfaces. Windows also has full support for multiple input/mouse configurations for multi-user environments. Starting with Windows XP, Microsoft introduced an SDK for developing applications that allow multiple input devices to be used at the same time with independent cursors and independent input points. However, it no longer appears to be available. The introduction of Windows Vista and Microsoft Surface (now known as Microsoft PixelSense) introduced a new set of input APIs that were adopted into Windows 7, allowing for 50 points/cursors, all controlled by independent users. The new input points provide traditional mouse input; however, they were designed with other input technologies like touch and image in mind. They inherently offer 3D coordinates along with pressure, size, tilt, angle, mask, and even an image bitmap to see and recognize the input point/object on the screen. As of 2009, Linux distributions and other operating systems that use X.Org, such as OpenSolaris and FreeBSD, support 255 cursors/input points through Multi-Pointer X. However, currently no window managers support Multi-Pointer X leaving it relegated to custom software usage. There have also been propositions of having a single operator use two mice simultaneously as a more sophisticated means of controlling various graphics and multimedia applications. Mouse buttons are microswitches which can be pressed to select or interact with an element of a graphical user interface, producing a distinctive clicking sound. Since around the late 1990s, the three-button scrollmouse has become the de facto standard. Users most commonly employ the second button to invoke a contextual menu in the computer's software user interface, which contains options specifically tailored to the interface element over which the mouse cursor currently sits. By default, the primary mouse button sits located on the left-hand side of the mouse, for the benefit of right-handed users; left-handed users can usually reverse this configuration via software. Nearly all mice now have an integrated input primarily intended for scrolling on top, usually a single-axis digital wheel or rocker switch which can also be depressed to act as a third button. Though less common, many mice instead have two-axis inputs such as a tiltable wheel, trackball, or touchpad. Those with a trackball may be designed to stay stationary, using the trackball instead of moving the mouse. Mickeys per second is a unit of measurement for the speed and movement direction of a computer mouse, where direction is often expressed as "horizontal" versus "vertical" mickey count. However, speed can also refer to the ratio between how many pixels the cursor moves on the screen and how far the mouse moves on the mouse pad, which may be expressed as pixels per mickey, pixels per inch, or pixels per centimeter. The computer industry often measures mouse sensitivity in terms of counts per inch (CPI), commonly expressed as dots per inch (DPI) – the number of steps the mouse will report when it moves one inch. In early mice, this specification was called pulses per inch (ppi). The mickey originally referred to one of these counts, or one resolvable step of motion. If the default mouse-tracking condition involves moving the cursor by one screen-pixel or dot on-screen per reported step, then the CPI does equate to DPI: dots of cursor motion per inch of mouse motion. The CPI or DPI as reported by manufacturers depends on how they make the mouse; the higher the CPI, the faster the cursor moves with mouse movement. However, software can adjust the mouse sensitivity, making the cursor move faster or slower than its CPI. As of 2007, software can change the speed of the cursor dynamically, taking into account the mouse's absolute speed and the movement from the last stop-point. In most software, an example being the Windows platforms, this setting is named "speed", referring to "cursor precision". However, some operating systems name this setting "acceleration", the typical Apple OS designation. This term is incorrect. Mouse acceleration in most mouse software refers to the change in speed of the cursor over time while the mouse movement is constant. For simple software, when the mouse starts to move, the software will count the number of "counts" or "mickeys" received from the mouse and will move the cursor across the screen by that number of pixels (or multiplied by a rate factor, typically less than 1). The cursor will move slowly on the screen, with good precision. When the movement of the mouse passes the value set for some threshold, the software will start to move the cursor faster, with a greater rate factor. Usually, the user can set the value of the second rate factor by changing the "acceleration" setting. Operating systems sometimes apply acceleration, referred to as "ballistics", to the motion reported by the mouse. For example, versions of Windows prior to Windows XP doubled reported values above a configurable threshold, and then optionally doubled them again above a second configurable threshold. These doublings applied separately in the X and Y directions, resulting in very nonlinear response. Engelbart's original mouse did not require a mousepad; the mouse had two large wheels which could roll on virtually any surface. However, most subsequent mechanical mice starting with the steel roller ball mouse have required a mousepad for optimal performance. The mousepad, the most common mouse accessory, appears most commonly in conjunction with mechanical mice, because to roll smoothly the ball requires more friction than common desk surfaces usually provide. So-called "hard mousepads" for gamers or optical/laser mice also exist. Most optical and laser mice do not require a pad, the notable exception being early optical mice which relied on a grid on the pad to detect movement (e.g. Mouse Systems). Whether to use a hard or soft mousepad with an optical mouse is largely a matter of personal preference. One exception occurs when the desk surface creates problems for the optical or laser tracking, for example, a transparent or reflective surface, such as glass. Some mice also come with small "pads" attached to the bottom surface, also called mouse feet or mouse skates, that help the user slide the mouse smoothly across surfaces. Around 1981, Xerox included mice with its Xerox Star, based on the mouse used in the 1970s on the Alto computer at Xerox PARC. Sun Microsystems, Symbolics, Lisp Machines Inc., and Tektronix also shipped workstations with mice, starting in about 1981. Later, inspired by the Star, Apple Computer released the Apple Lisa, which also used a mouse. However, none of these products achieved large-scale success. Only with the release of the Apple Macintosh in 1984 did the mouse see widespread use. The Macintosh design, commercially successful and technically influential, led many other vendors to begin producing mice or including them with their other computer products (by 1986, Atari ST, Amiga, Windows 1.0, GEOS for the Commodore 64, and the Apple IIGS). The widespread adoption of graphical user interfaces in the software of the 1980s and 1990s made mice all but indispensable for controlling computers. In November 2008, Logitech built their billionth mouse. The device often functions as an interface for PC-based computer games and sometimes for video game consoles. The Classic Mac OS Desk Accessory Puzzle in 1984 was the first game designed specifically for a mouse. FPSs naturally lend themselves to separate and simultaneous control of the player's movement and aim, and on computers this has traditionally been achieved with a combination of keyboard and mouse. Players use the X-axis of the mouse for looking (or turning) left and right, and the Y-axis for looking up and down; the keyboard is used for movement and supplemental inputs. Many shooting genre players prefer a mouse over a gamepad analog stick because the wide range of motion offered by a mouse allows for faster and more varied control. Although an analog stick allows the player more granular control, it is poor for certain movements, as the player's input is relayed based on a vector of both the stick's direction and magnitude. Thus, a small but fast movement (known as "flick-shotting") using a gamepad requires the player to quickly move the stick from its rest position to the edge and back again in quick succession, a difficult maneuver. In addition the stick also has a finite magnitude; if the player is currently using the stick to move at a non-zero velocity their ability to increase the rate of movement of the camera is further limited based on the position their displaced stick was already at before executing the maneuver. The effect of this is that a mouse is well suited not only to small, precise movements but also to large, quick movements and immediate, responsive movements; all of which are important in shooter gaming. This advantage also extends in varying degrees to similar game styles such as third-person shooters. Some incorrectly ported games or game engines have acceleration and interpolation curves which unintentionally produce excessive, irregular, or even negative acceleration when used with a mouse instead of their native platform's non-mouse default input device. Depending on how deeply hardcoded this misbehavior is, internal user patches or external 3rd-party software may be able to fix it. Individual game engines will also have their own sensitivities. This often restricts one from taking a game's existing sensitivity, transferring it to another, and acquiring the same 360 rotational measurements. A sensitivity converter is required in order to translate rotational movements properly. Due to their similarity to the WIMP desktop metaphor interface for which mice were originally designed, and to their own tabletop game origins, computer strategy games are most commonly played with mice. In particular, real-time strategy and MOBA games usually require the use of a mouse. The left button usually controls primary fire. If the game supports multiple fire modes, the right button often provides secondary fire from the selected weapon. Games with only a single fire mode will generally map secondary fire to aim down the weapon sights. In some games, the right button may also invoke accessories for a particular weapon, such as allowing access to the scope of a sniper rifle or allowing the mounting of a bayonet or silencer. Players can use a scroll wheel for changing weapons (or for controlling scope-zoom magnification, in older games). On most first person shooter games, programming may also assign more functions to additional buttons on mice with more than three controls. A keyboard usually controls movement (for example, WASD for moving forward, left, backward, and right, respectively) and other functions such as changing posture. Since the mouse serves for aiming, a mouse that tracks movement accurately and with less lag (latency) will give a player an advantage over players with less accurate or slower mice. In some cases the right mouse button may be used to move the player forward, either in lieu of, or in conjunction with the typical WASD configuration. Many games provide players with the option of mapping their own choice of a key or button to a certain control. An early technique of players, circle strafing, saw a player continuously strafing while aiming and shooting at an opponent by walking in circle around the opponent with the opponent at the center of the circle. Players could achieve this by holding down a key for strafing while continuously aiming the mouse toward the opponent. Games using mice for input are so popular that many manufacturers make mice specifically for gaming. Such mice may feature adjustable weights, high-resolution optical or laser components, additional buttons, ergonomic shape, and other features such as adjustable CPI. Mouse Bungees are typically used with gaming mice because it eliminates the annoyance of the cable. Many games, such as first- or third-person shooters, have a setting named "invert mouse" or similar (not to be confused with "button inversion", sometimes performed by left-handed users) which allows the user to look downward by moving the mouse forward and upward by moving the mouse backward (the opposite of non-inverted movement). This control system resembles that of aircraft control sticks, where pulling back causes pitch up and pushing forward causes pitch down; computer joysticks also typically emulate this control-configuration. After id Software's commercial hit of Doom, which did not support vertical aiming, competitor Bungie's Marathon became the first first-person shooter to support using the mouse to aim up and down. Games using the Build engine had an option to invert the Y-axis. The "invert" feature actually made the mouse behave in a manner that users now regard as non-inverted (by default, moving mouse forward resulted in looking down). Soon after, id Software released Quake, which introduced the invert feature as users now know it. In 1988, the VTech Socrates educational video game console featured a wireless mouse with an attached mouse pad as an optional controller used for some games. In the early 1990s, the Super Nintendo Entertainment System video game system featured a mouse in addition to its controllers. A mouse was also released for the Nintendo 64, although it was only released in Japan. The 1992 game Mario Paint in particular used the mouse's capabilities, as did its Japanese-only successor Mario Artist on the N64 for its 64DD disk drive peripheral in 1999. Sega released official mice for their Genesis/Mega Drive, Saturn and Dreamcast consoles. NEC sold official mice for its PC Engine and PC-FX consoles. Sony released an official mouse product for the PlayStation console, included one along with the Linux for PlayStation 2 kit, as well as allowing owners to use virtually any USB mouse with the PS2, PS3, and PS4. Nintendo's Wii also had this feature implemented in a later software update, and this support was retained on its successor, the Wii U. Microsoft's Xbox line of game consoles (which used operaring systems based on modified versions of Windows NT) also had universal-wide mouse support using USB.
{{sources|date=September 2013}} '''Metallurgy''' is the study of [[metal]]s. [[Alloy]]s (mixtures of metals) are also studied. There are two main branches of Metallurgy. They are physical and chemical metallurgy. A Metallurgist is individual who researches, manufactures and develops metal items that range from car parts to [[Semiconductor|semiconductors]]. Metallurgists and Metallurgical Engineers often work for consulting [[engineering]] firms; mining and manufacturing companies; and in government and universities. Many metallurgists hold a Bachelor of Engineering or similar 4- year degree. '''Common Job Tasks''' * Study the properties metals and alloys through scientific experiments * Plan and design machines and processes to concentrate, extract, refine and process [[Metal|metals]].<ref>{{Cite web|last=|first=|date=July 14, 2020|title=Metallurgical and materials engineers|url=https://www.workbc.ca/careers/2142|url-status=live|archive-url=https://web.archive.org/web/20220523101937/https://www.workbc.ca/careers/2142|archive-date=May 23, 2022|access-date=November 26, 2020|website=WorkBC}}</ref> *Work in a manner that is environmentally and economically conscious. == Chemical Metallurgy == Chemical Metallurgy is the study of extracting and refining metals through chemical reactions. This field also commonly involves [[corrosion]] prevention and surface treatment of metals. Scientific concepts such as [[thermodynamics]], the study of [[heat]] flow and [[kinetics]], the study of the speed of reactions are very central to chemical metallurgy. Metallurgists are very interested in the chemical reactions that occur with different elements. Chemical reactions are carried out on a large scale to obtain a useful product. For example, the production of [[steel]] in a [[blast furnace]] involves turning different [[Iron]] and oxide compounds into pure Iron.<ref>{{Cite web|date=2015-01-18|title=23.3: Metallurgy of Iron and Steel|url=https://chem.libretexts.org/Bookshelves/General_Chemistry/Map%3A_General_Chemistry_(Petrucci_et_al.)/23%3A_The_Transition_Elements/23.3%3A_Metallurgy_of_Iron_and_Steel|access-date=2020-11-27|website=Chemistry LibreTexts|language=en|archive-date=2021-05-10|archive-url=https://web.archive.org/web/20210510213442/https://chem.libretexts.org/Bookshelves/General_Chemistry/Map%3A_General_Chemistry_(Petrucci_et_al.)/23%3A_The_Transition_Elements/23.3%3A_Metallurgy_of_Iron_and_Steel|url-status=live}}</ref> <chem>3Fe2O4 + CO ->2FeO4 + CO2 </chem> <chem>Fe3O4 + CO -> 3FeO +CO2 </chem> <chem>FeO + CO ->Fe + CO2 </chem> == Physical Metallurgy == Physical metallurgy is the science of making useful properties out of metals. This field focus on the physical property of metals and alloys such as strength and hardness. Properties of metals might be electrical, magnetic and physical in nature. These properties can be changed by alloying or heat treatment. There are many ways to make things from metals. For example, large parts may be created by pouring liquid metal into a large cast. This is called [[casting]]. <ref>{{Cite book|title=Die Casting Metallurgy {{!}} ScienceDirect|url=https://www.sciencedirect.com/book/9780408107174/die-casting-metallurgy|access-date=2020-11-27|isbn=9780408107174 |language=en|archive-date=2020-12-05|archive-url=https://web.archive.org/web/20201205225239/https://www.sciencedirect.com/book/9780408107174/die-casting-metallurgy|url-status=live |last1=Kaye |first1=Alan |last2=Street |first2=Arthur |year=1982 |publisher=Butterworth Scientific }}</ref> Important factors in physical metallurgy include time and temperature in which the metal is made and cooled. Metals undergo different phase changes, the particular position and type of crystal structure inside the metal will determine if a metal is brittle or flexible. <ref>{{Cite book|title=Materials Science and Engineering: An Introduction Hardcover – Dec 30 2009|last=Callister|first=William|publisher=Wiley|year=2009|isbn=978-0470419977|location=|pages=}}</ref> {{Further|Heat treating}} == Smelting == [[Ore]]s are rocks that have a [[chemical compound]] of a metal with [[oxygen]], called an [[oxide]]. Separating the oxygen from the metal is called [[smelting]]. This is done with [[chemistry]] or [[electricity]], usually at very high temperatures. This is the first step in metallurgy. A rock containing enough metal to be profitable is called ''ore''. {{Further|Smelting}} == Alloys == An alloy is a combination of a metal with one or more elements. The metal mixture must be a solid to be called as an alloy. A common way to combine metals into an alloy is by melting them, mixing and then allowing them to solidify by cooling at room temperature. Alloys have been known since 50,000 BCE and have been continuously studied and improved. Common Example of alloys include: * Chromium alloys: Good at withstanding high temperature and have good resistance to corrosion. Common composition include a mixture of chromium and Nickel * Cast Iron: Often used in making pots and pans. Cast Iron is a mixture of Iron and around 1-3% Carbon.<ref>{{Cite book|url=http://archive.org/details/elementsmetallur00fcam|title=Elements of metallurgy and engineering alloys|last=Campbell|first=F. C. (Flake C. )|date=2008|publisher=Materials Park, Ohio : ASM International|others=Library Genesis|isbn=978-0-87170-867-0}}</ref> It is often made using a blast furnace. *Bronze: Copper with 12-15% Tin and minor additions of other metals such as manganese, aluminum and zinc. Bronze is a common metal found in archeological sites.<ref>{{Cite journal|last1=Radivojević|first1=Miljana|last2=Roberts|first2=Benjamin W.|last3=Pernicka|first3=Ernst|last4=Stos-Gale|first4=Zofia|last5=Martinón-Torres|first5=Marcos|last6=Rehren|first6=Thilo|last7=Bray|first7=Peter|last8=Brandherm|first8=Dirk|last9=Ling|first9=Johan|date=2019-06-01|title=The Provenance, Use, and Circulation of Metals in the European Bronze Age: The State of Debate|url=https://doi.org/10.1007/s10814-018-9123-9|journal=Journal of Archaeological Research|language=en|volume=27|issue=2|pages=131–185|doi=10.1007/s10814-018-9123-9|s2cid=149947751|issn=1573-7756|access-date=2020-11-27|archive-date=2022-05-23|archive-url=https://web.archive.org/web/20220523101936/https://link.springer.com/article/10.1007/s10814-018-9123-9|url-status=live}}</ref> {{Further|Alloy}} == Metal parts == Another part of metallurgy is making parts from metals. These parts must be made so they will not break when they are used. Metallurgists look to make metal that meets the needs of the process it will be used in. Sometimes the metal must be lightweight or low in density. Other times it must be tough (not easily broken). The metallurgist must follow directions when making the part to know what metal to use. the process of metal selection is known as the materials selection process. Metallurgists use mathematical equations to determine what are the most important physical features of the metal. Common desired characteristics include ductility, toughness, corrosion resistance, and strength. Different metals and their associated features are often graphed on Ashby plots.<ref>{{Cite journal|last=Skakov|first=Yu.A.|date=May 2000|title=The Formation Sequence of Intermediate Phases in Mechanical Alloying of Binary Systems|url=http://dx.doi.org/10.4028/www.scientific.net/msf.343-346.597|journal=Materials Science Forum|volume=343-346|pages=597–602|doi=10.4028/www.scientific.net/msf.343-346.597|s2cid=135534957|issn=1662-9752|access-date=2020-11-27|archive-date=2022-05-23|archive-url=https://web.archive.org/web/20220523101935/https://www.scientific.net/MSF.343-346.597|url-status=live}}</ref> A metal starts as a block, called an [[ingot]]. Metallurgists must know how to make a metal part from an ingot. Parts are made from ingots different ways. When a big [[hammer]] is used, it is called [[forge|forging]]. To make thin metals, a metal is put between two rolls and moved, called [[rolling]]. [[File:EB1911 - Forging - Fig. 1.png|thumb|An example of forging by a chisel shaped hammer |center]] Making metal hot makes it easier to change the shape of a metal part. For this reason, most metal parts are made using hot metal. This is [[hot work]]. Two metal parts can be put together with much heat. This is called [[welding]]. Iron is easy to weld. == Tools == Metallurgists use many tools to determine the characterizes of metal using a technique called metallography. One of the most useful tools is the [[Scanning electron microscope|Scanning Electron Microscope]]. To use an SEM a metal sample is polished and hit with electrons to produce a clear image of the metal. This permits for the metal microstructure to be seen. A microstructure is what the metal looks like when zoomed in very closely.<ref>{{Cite web|title=Crystal Microstructure - an overview {{!}} ScienceDirect Topics|url=https://www.sciencedirect.com/topics/chemistry/crystal-microstructure|access-date=2020-11-27|website=www.sciencedirect.com|archive-date=2022-05-23|archive-url=https://web.archive.org/web/20220523101936/https://www.sciencedirect.com/topics/chemistry/crystal-microstructure|url-status=live}}</ref> There are many types of microstructure such as pearlite and ferrite. Metals are sometimes pulled until they break using a tensiometer. This is the [[tension test]]. Information such as plastic and elastic deformation can be determined from these tests. It is also common for metallurgists to use computer simulations to predict heat flow or the cost of making a particular product. Common software used in metallurgy include ABAQUS<ref>{{Cite web|title=ABAQUS Student Edition {{!}} 3DEXPERIENCE Edu|url=https://edu.3ds.com/en/software/abaqus-student-edition|access-date=2021-02-17|website=edu.3ds.com|language=en|archive-date=2021-03-03|archive-url=https://web.archive.org/web/20210303032417/https://edu.3ds.com/en/software/abaqus-student-edition|url-status=live}}</ref> and ANSYS.<ref>{{Cite web|title=Engineering Simulation & 3D Design Software {{!}} Ansys|url=https://www.ansys.com/|access-date=2021-02-17|website=www.ansys.com|archive-date=2021-02-17|archive-url=https://web.archive.org/web/20210217015915/https://www.ansys.com/|url-status=live}}</ref> ==References== <references /> [[File:Microstructure of a stainless steel.jpg|center|thumb|320x320px|Microstructure of Stainless Steel ]] {{authority control}}{{Chem-stub}} [[Category:Metallurgy| ]] [[Category:Materials science]] [[Category:Chemistry]]
Metallurgy is a domain of materials science and engineering that studies the physical and chemical behavior of metallic elements, their inter-metallic compounds, and their mixtures, which are known as alloys. Metallurgy encompasses both the science and the technology of metals, including the production of metals and the engineering of metal components used in products for both consumers and manufacturers. Metallurgy is distinct from the craft of metalworking. Metalworking relies on metallurgy in a similar manner to how medicine relies on medical science for technical advancement. A specialist practitioner of metallurgy is known as a metallurgist. The science of metallurgy is further subdivided into two broad categories: chemical metallurgy and physical metallurgy. Chemical metallurgy is chiefly concerned with the reduction and oxidation of metals, and the chemical performance of metals. Subjects of study in chemical metallurgy include mineral processing, the extraction of metals, thermodynamics, electrochemistry, and chemical degradation (corrosion). In contrast, physical metallurgy focuses on the mechanical properties of metals, the physical properties of metals, and the physical performance of metals. Topics studied in physical metallurgy include crystallography, material characterization, mechanical metallurgy, phase transformations, and failure mechanisms. Historically, metallurgy has predominately focused on the production of metals. Metal production begins with the processing of ores to extract the metal, and includes the mixture of metals to make alloys. Metal alloys are often a blend of at least two different metallic elements. However, non-metallic elements are often added to alloys in order to achieve properties suitable for an application. The study of metal production is subdivided into ferrous metallurgy (also known as black metallurgy) and non-ferrous metallurgy, also known as colored metallurgy. Ferrous metallurgy involves processes and alloys based on iron, while non-ferrous metallurgy involves processes and alloys based on other metals. The production of ferrous metals accounts for 95% of world metal production. Modern metallurgists work in both emerging and traditional areas as part of an interdisciplinary team alongside material scientists and other engineers. Some traditional areas include mineral processing, metal production, heat treatment, failure analysis, and the joining of metals (including welding, brazing, and soldering). Emerging areas for metallurgists include nanotechnology, superconductors, composites, biomedical materials, electronic materials (semiconductors) and surface engineering. Many applications, practices, and devices associated or involved in metallurgy were established in ancient China, such as the innovation of the blast furnace, cast iron, hydraulic-powered trip hammers, and double acting piston bellows. Metallurgy derives from the Ancient Greek μεταλλουργός, metallourgós, "worker in metal", from μέταλλον, métallon, "mine, metal" + ἔργον, érgon, "work" The word was originally an alchemist's term for the extraction of metals from minerals, the ending -urgy signifying a process, especially manufacturing: it was discussed in this sense in the 1797 Encyclopædia Britannica. In the late 19th century, metallurgy's definition was extended to the more general scientific study of metals, alloys, and related processes. In English, the /mɛˈtælərdʒi/ pronunciation is the more common one in the United Kingdom. The /ˈmɛtəlɜːrdʒi/ pronunciation is the more common one in the United States US and is the first-listed variant in various American dictionaries, including Merriam-Webster Collegiate and American Heritage. The earliest recorded metal employed by humans appears to be gold, which can be found free or "native". Small amounts of natural gold have been found in Spanish caves dating to the late Paleolithic period, 40,000 BC. Silver, copper, tin and meteoric iron can also be found in native form, allowing a limited amount of metalworking in early cultures. Certain metals, such as tin, lead, and copper can be recovered from their ores by simply heating the rocks in a fire or blast furnace in a process known as smelting. The first evidence of this extractive metallurgy, dating from the 5th and 6th millennia BC, has been found at archaeological sites in Majdanpek, Jarmovac and Pločnik, in present-day Serbia. The earliest evidence of copper smelting is found at the Belovode site near Pločnik. The site produced a copper axe from 5,500 BC, belonging to the Vinča culture. The earliest use of lead was in the late neolithic settlements of Yarim Tepe and Arpachiyah in present-day Iraq. The artifacts suggest that lead smelting predated copper smelting. Copper smelting is also documented at this site at about the same time period (soon after 6,000 BC), although the use of lead seems to precede copper smelting. Early metallurgy is also documented at the nearby site of Tell Maghzaliyah, which seems to be dated even earlier, and completely lacks that pottery. The Balkans were the site of major Neolithic cultures, including Butmir, Vinča, Varna, Karanovo, and Hamangia. The Varna Necropolis, Bulgaria, is a burial site in the western industrial zone of Varna, approximately 4 km from the city centre, internationally considered one of the key archaeological sites in world prehistory. The oldest gold treasure in the world, dating from 4,600 BC to 4,200 BC, was discovered at the site. The gold piece dating from 4,500 BC, recently founded in Durankulak, near Varna is another important example. Other signs of early metals are found from the third millennium BC in [Palmela]], Portugal, Los Millares, Spain, and Stonehenge, United Kingdom. The precise beginnings, however, have not be clearly ascertained and new discoveries are both continuous and ongoing. In Tamilnadu, in approximately 1900 BC, ancient iron smelting sites were functioning in Tamil Nadu. In the Near East, about 3,500 BC, it was discovered that by combining copper and tin, a superior metal could be made, an alloy called bronze. This represented a major technological shift known as the Bronze Age. The extraction of iron from its ore into a workable metal is much more difficult than for copper or tin. The process appears to have been invented by the Hittites in about 1200 BC, beginning the Iron Age. The secret of extracting and working iron was a key factor in the success of the Philistines. Historical developments in ferrous metallurgy can be found in a wide variety of past cultures and civilizations. This includes the ancient and medieval kingdoms and empires of the Middle East and Near East, ancient Iran, ancient Egypt, ancient Nubia, and Anatolia in present-day Turkey, Ancient Nok, Carthage, the Greeks and Romans of ancient Europe, medieval Europe, ancient and medieval China, ancient and medieval India, ancient and medieval Japan, amongst others. Many applications, practices, and devices associated or involved in metallurgy were established in ancient China, such as the innovation of the blast furnace, cast iron, hydraulic-powered trip hammers, and double acting piston bellows. A 16th century book by Georg Agricola, De re metallica, describes the highly developed and complex processes of mining metal ores, metal extraction, and metallurgy of the time. Agricola has been described as the "father of metallurgy". Extractive metallurgy is the practice of removing valuable metals from an ore and refining the extracted raw metals into a purer form. In order to convert a metal oxide or sulphide to a purer metal, the ore must be reduced physically, chemically, or electrolytically. Extractive metallurgists are interested in three primary streams: feed, concentrate (metal oxide/sulphide) and tailings (waste). After mining, large pieces of the ore feed are broken through crushing or grinding in order to obtain particles small enough, where each particle is either mostly valuable or mostly waste. Concentrating the particles of value in a form supporting separation enables the desired metal to be removed from waste products. Mining may not be necessary, if the ore body and physical environment are conducive to leaching. Leaching dissolves minerals in an ore body and results in an enriched solution. The solution is collected and processed to extract valuable metals. Ore bodies often contain more than one valuable metal. Tailings of a previous process may be used as a feed in another process to extract a secondary product from the original ore. Additionally, a concentrate may contain more than one valuable metal. That concentrate would then be processed to separate the valuable metals into individual constituents. Much effort has been placed on understanding iron–carbon alloy system, which includes steels and cast irons. Plain carbon steels (those that contain essentially only carbon as an alloying element) are used in low-cost, high-strength applications, where neither weight nor corrosion are a major concern. Cast irons, including ductile iron, are also part of the iron-carbon system. Iron-Manganese-Chromium alloys (Hadfield-type steels) are also used in non-magnetic applications such as directional drilling. Other engineering metals include aluminium, chromium, copper, magnesium, nickel, titanium, zinc, and silicon. These metals are most often used as alloys with the noted exception of silicon, which is not a metal. Other forms include: In production engineering, metallurgy is concerned with the production of metallic components for use in consumer or engineering products. This involves production of alloys, shaping, heat treatment and surface treatment of product. The task of the metallurgist is to achieve balance between material properties, such as cost, weight, strength, toughness, hardness, corrosion, fatigue resistance and performance in temperature extremes. To achieve this goal, the operating environment must be carefully considered. Determining the hardness of the metal using the Rockwell, Vickers, and Brinell hardness scales is a commonly used practice that helps better understand the metal's elasticity and plasticity for different applications and production processes. In a saltwater environment, most ferrous metals and some non-ferrous alloys corrode quickly. Metals exposed to cold or cryogenic conditions may undergo a ductile to brittle transition and lose their toughness, becoming more brittle and prone to cracking. Metals under continual cyclic loading can suffer from metal fatigue. Metals under constant stress at elevated temperatures can creep. Cold-working processes, in which the product's shape is altered by rolling, fabrication or other processes, while the product is cold, can increase the strength of the product by a process called work hardening. Work hardening creates microscopic defects in the metal, which resist further changes of shape. Metals can be heat-treated to alter the properties of strength, ductility, toughness, hardness and resistance to corrosion. Common heat treatment processes include annealing, precipitation strengthening, quenching, and tempering: Often, mechanical and thermal treatments are combined in what are known as thermo-mechanical treatments for better properties and more efficient processing of materials. These processes are common to high-alloy special steels, superalloys and titanium alloys. Electroplating is a chemical surface-treatment technique. It involves bonding a thin layer of another metal such as gold, silver, chromium or zinc to the surface of the product. This is done by selecting the coating material electrolyte solution, which is the material that is going to coat the workpiece (gold, silver, zinc). There needs to be two electrodes of different materials: one the same material as the coating material and one that is receiving the coating material. Two electrodes are electrically charged and the coating material is stuck to the work piece. It is used to reduce corrosion as well as to improve the product's aesthetic appearance. It is also used to make inexpensive metals look like the more expensive ones (gold, silver). Shot peening is a cold working process used to finish metal parts. In the process of shot peening, small round shot is blasted against the surface of the part to be finished. This process is used to prolong the product life of the part, prevent stress corrosion failures, and also prevent fatigue. The shot leaves small dimples on the surface like a peen hammer does, which cause compression stress under the dimple. As the shot media strikes the material over and over, it forms many overlapping dimples throughout the piece being treated. The compression stress in the surface of the material strengthens the part and makes it more resistant to fatigue failure, stress failures, corrosion failure, and cracking. Thermal spraying techniques are another popular finishing option, and often have better high temperature properties than electroplated coatings. Thermal spraying, also known as a spray welding process, is an industrial coating process that consists of a heat source (flame or other) and a coating material that can be in a powder or wire form, which is melted then sprayed on the surface of the material being treated at a high velocity. The spray treating process is known by many different names such as HVOF (High Velocity Oxygen Fuel), plasma spray, flame spray, arc spray and metalizing. Electroless deposition (ED) or electroless plating is defined as the autocatalytic process through which metals and metal alloys are deposited onto nonconductive surfaces.These nonconductive surfaces include plastics, ceramics, and glass etc., which can then become decorative, anti-corrosive, and conductive depending on their final functions. Electroless deposition is a chemical processes that create metal coatings on various materials by autocatalytic chemical reduction of metal cations in a liquid bath. Metallurgists study the microscopic and macroscopic structure of metals using metallography, a technique invented by Henry Clifton Sorby. In metallography, an alloy of interest is ground flat and polished to a mirror finish. The sample can then be etched to reveal the microstructure and macrostructure of the metal. The sample is then examined in an optical or electron microscope, and the image contrast provides details on the composition, mechanical properties, and processing history. Crystallography, often using diffraction of x-rays or electrons, is another valuable tool available to the modern metallurgist. Crystallography allows identification of unknown materials and reveals the crystal structure of the sample. Quantitative crystallography can be used to calculate the amount of phases present as well as the degree of strain to which a sample has been subjected.
[[File:Americanairlines.arp.750pix.jpg|thumb|275px|right|[[American Airlines]] [[Boeing 767]], a fixed-wing aircraft. To [[take off]] it must reach a speed of about {{convert|150|kn}}]] A '''fixed-wing aircraft''' is a kind of [[aircraft]]. An aircraft is a [[machine]] that can [[Flight|fly]], but is heavier than [[air]]. Fixed-wing aircraft are sometimes called '''[[Airplane|airplanes]]''', '''aeroplanes''' or sometimes just "planes". All fixed-wing aircraft have [[wing]]s that use [[Movement|forward airspeed]] to generate [[lift (force)|lift]].<ref>{{cite web |url=http://www.aviation-history.com/theory/lift.htm |title=How Airplanes Fly: A Physical Description of Lift |author=David Anderson; Scott Eberhardt |website= |publisher=The Aviation History Online Museum |accessdate=26 March 2016}}</ref> [[Glider]]s are fixed-wing aircraft that do not have [[engine]]s.<ref>Susan Meredith, ''How Do Aircraft Fly?'' (New York, NY: Chelsea Clubhouse, 2010), p. 18</ref> Fixed-wing aircraft fly between many cities all over the world, carrying people and [[cargo]]. Big cities usually have an [[international]] [[airport]], where large fixed-wing [[airliner]]s operate. [[Airbus]] and [[Boeing]] are the two biggest makers of large airliners. [[Statistics]] show that riding in a plane is safer than driving in a car.<ref name="ur.umich.edu">[http://ur.umich.edu/0203/Jan20_03/18.shtml Flying still safer than driving, even in wake of Sept. 11]{{Dead link|date=November 2022|bot=InternetArchiveBot|fix-attempted=yes}}</ref> ==History== A steam-powered fixed-wing [[unmanned aerial vehicle]] that weighed {{Cvt|9|lb}},<ref>[http://news.bbc.co.uk/1/hi/england/1591057.stm High hopes for replica plane, BBC News]</ref> was built by [[John Stringfellow]], in [[Chard]], [[Somerset]], [[England]] in 1848. It could fly by itself without needing to be dropped from high up. There were [[glider|gliders]] before this, but they had to fly by being pushed off a building or hill. The first man who flew (took off, steered, and landed) a motor-powered aircraft was [[Orville Wright]] in 1903 in Kitty Hawk, North Carolina, [[USA]].<ref name="WDL">{{cite web |url = http://www.wdl.org/en/item/11372/ |title = Telegram from Orville Wright in Kitty Hawk, North Carolina, to His Father Announcing Four Successful Flights, 1903 December 17 |website = [[World Digital Library]] |date = 1903-12-17 |accessdate = 2013-07-21 }}</ref> Advances in technology have made fixed-wing aircraft more efficient. Things like [[winglet]]s and more efficient [[turbofan]]s have helped to do this. == Parts == Most fixed-wing aircraft have certain parts in common. * The '''wings''' are the most important part because they are what makes the plane fly. The [[wing]]s create a [[Force (physics)|force]] called lift that goes against gravity which makes the plane get off the ground. When air flows around the wing (which happens when the plane moves forward) the wing pushes air down, which in turn pushes the plane up. Lift can also be explained using [[Bernoulli's principle]]; since wings are designed so that air goes faster on the top of the wing than on the bottom, the higher pressure on the bottom will push the plane up. Some wings have panels on the back called flaps. Flaps can be extended to make the wing bigger. This makes the wing make more lift, but also more [[Drag (physics)|drag]]. This lets the plane fly slower without [[stall]]ing, so flaps are often used during takeoff and landing. * The '''tail''' (also called the [[vertical stabilizer]]) is a large flat panel normally at the back of the plane. This helps the plane fly straight. Some planes have more than one tail, like the [[Antonov An-225]]. Some planes also have no tail, like the [[Northrop Grumman B-2 Spirit]]. * '''Engines''' push the plane forward. This makes air flow over the wings, which makes the plane fly. Some planes have [[propeller]]s that the [[engine]] turns. Others have [[jet engine]]s. Some fixed-wing aircraft, such as [[glider]]s, don't have engines. Instead they use gravity to give them forward speed. * The '''fuselage''' is the part of the plane that holds the people and [[cargo]]. It is normally shaped like a [[cylinder]]. The [[fuselage]] might have windows or doors. * The '''landing gear''' or [[undercarriage]] is what supports the plane on the ground. It can be wheels if the plane lands on the ground, or [[float (nautical)|float]]s if the plane lands in the water. If the plane lands in snow, the landing gear usually has [[wikt:ski|ski]]s. Many planes can retract (bring inside the fuselage) their landing gear to make less [[Drag (physics)|drag]]. == Uses == === Transport === Fixed-wing aircraft have long been used as [[airliner]]<nowiki/>s for moving [[passenger]]<nowiki/>s from place to place. [[Cargo aircraft]] carry [[cargo]] across seas and long distances, and passenger aircraft also carry some cargo. === War === [[File:Su-27 low pass.jpg|thumb|''Russian Knights'' in a low pass, [[Moscow]].]] The first aircraft for war bombing was used in [[Libya]] in 1911 by [[Italy]] against the [[Ottoman Empire]]. Some fixed-wing aircraft are used by [[air force]]s to [[defense (military)|defend]] countries. These may be [[fighter aircraft]], using [[machine gun|guns]] or [[missile]]s for combat with other aircraft. They may be [[bomber]]s, dropping [[bomb]]s on ground targets. Fixed-wing aircraft allow people to travel longer distances, and faster than [[ship]]s or [[railway|trains]]. Aircraft can fly from New York to London in about 7 hours. It would take one week or more on a ship. Militaries use airplanes to carry their soldiers quickly from place to place. Airplanes are also used by the military to see many things on the ground easily. This is called [[surveillance]] or [[reconnaissance]]. Often, fixed-wing aircraft will take [[photograph]]s as well, which can be used for military planning later. ==Image gallery== <gallery widths="px" heights="px"> File:80 Blue Sukhoi Su-17 UM-3 Russian Airforce (7724104334).jpg|Military airplanes in [[Russia]] File:Edw-nasa-747-041103-01.jpg|An airplane used for research by [[NASA]] File:An Aerial Fight Art.IWMART654.jpg|Early airplanes in a battle in [[France]] File:Cessna 208B Grand Caravan AN1189964.jpg|A private airplane </gallery> == References == {{commonscat|Fixed-wing aircraft}} {{reflist}} {{aircraft}} [[Category:Aircraft]] [[Category:Aerospace engineering]]
A fixed-wing aircraft is a heavier-than-air flying machine, such as an airplane, which is capable of flight using wings that generate lift caused by the aircraft's forward airspeed and the shape of the wings. Fixed-wing aircraft are distinct from rotary-wing aircraft (in which the wings form a rotor mounted on a spinning shaft or "mast"), and ornithopters (in which the wings flap in a manner similar to that of a bird). The wings of a fixed-wing aircraft are not necessarily rigid; kites, hang gliders, variable-sweep wing aircraft and airplanes that use wing morphing are all examples of fixed-wing aircraft. Gliding fixed-wing aircraft, including free-flying gliders of various kinds and tethered kites, can use moving air to gain altitude. Powered fixed-wing aircraft (airplanes) that gain forward thrust from an engine include powered paragliders, powered hang gliders and some ground effect vehicles. Most fixed-wing aircraft are flown by a pilot on board the craft, but some are specifically designed to be unmanned and controlled either remotely or autonomously (using onboard computers). Kites were used approximately 2,800 years ago in China, where materials ideal for kite building were readily available. Some authors hold that leaf kites were being flown much earlier in what is now Sulawesi, based on their interpretation of cave paintings on Muna Island off Sulawesi. By at least 549 AD paper kites were being flown, as it was recorded in that year a paper kite was used as a message for a rescue mission. Ancient and medieval Chinese sources list other uses of kites for measuring distances, testing the wind, lifting men, signaling, and communication for military operations. Stories of kites were brought to Europe by Marco Polo towards the end of the 13th century, and kites were brought back by sailors from Japan and Malaysia in the 16th and 17th centuries. Although they were initially regarded as mere curiosities, by the 18th and 19th centuries kites were being used as vehicles for scientific research. Around 400 BC in Greece, Archytas was reputed to have designed and built the first artificial, self-propelled flying device, a bird-shaped model propelled by a jet of what was probably steam, said to have flown some 200 m (660 ft). This machine may have been suspended for its flight. One of the earliest purported attempts with gliders was by the 11th-century monk Eilmer of Malmesbury, which ended in failure. A 17th-century account states that the 9th-century poet Abbas Ibn Firnas made a similar attempt, though no earlier sources record this event. In 1799, Sir George Cayley set forth the concept of the modern airplane as a fixed-wing flying machine with separate systems for lift, propulsion, and control. Cayley was building and flying models of fixed-wing aircraft as early as 1803, and he built a successful passenger-carrying glider in 1853. In 1856, Frenchman Jean-Marie Le Bris made the first powered flight, by having his glider "L'Albatros artificiel" pulled by a horse on a beach. In 1884, the American John J. Montgomery made controlled flights in a glider as a part of a series of gliders built between 1883 and 1886. Other aviators who made similar flights at that time were Otto Lilienthal, Percy Pilcher, and protégés of Octave Chanute. In the 1890s, Lawrence Hargrave conducted research on wing structures and developed a box kite that lifted the weight of a man. His box kite designs were widely adopted. Although he also developed a type of rotary aircraft engine, he did not create and fly a powered fixed-wing aircraft. Sir Hiram Maxim built a craft that weighed 3.5 tons, with a 110-foot (34-meter) wingspan that was powered by two 360-horsepower (270-kW) steam engines driving two propellers. In 1894, his machine was tested with overhead rails to prevent it from rising. The test showed that it had enough lift to take off. The craft was uncontrollable, which Maxim, it is presumed, realized, because he subsequently abandoned work on it. The Wright brothers' flights in 1903 with their Flyer I are recognized by the Fédération Aéronautique Internationale (FAI), the standard setting and record-keeping body for aeronautics, as "the first sustained and controlled heavier-than-air powered flight". By 1905, the Wright Flyer III was capable of fully controllable, stable flight for substantial periods. In 1906, Brazilian inventor Alberto Santos Dumont designed, built and piloted an aircraft that set the first world record recognized by the Aéro-Club de France by flying the 14 bis 220 metres (720 ft) in less than 22 seconds. The flight was certified by the FAI. The Bleriot VIII design of 1908 was an early aircraft design that had the modern monoplane tractor configuration. It had movable tail surfaces controlling both yaw and pitch, a form of roll control supplied either by wing warping or by ailerons and controlled by its pilot with a joystick and rudder bar. It was an important predecessor of his later Bleriot XI Channel-crossing aircraft of the summer of 1909. World War I served as a testbed for the use of the aircraft as a weapon. Aircraft demonstrated their potential as mobile observation platforms, then proved themselves to be machines of war capable of causing casualties to the enemy. The earliest known aerial victory with a synchronized machine gun-armed fighter aircraft occurred in 1915, by German Luftstreitkräfte Leutnant Kurt Wintgens. Fighter aces appeared; the greatest (by number of air victories) was Manfred von Richthofen. Following World War I, aircraft technology continued to develop. Alcock and Brown crossed the Atlantic non-stop for the first time in 1919. The first commercial flights took place between the United States and Canada in 1919. The so-called Golden Age of Aviation occurred between the two World Wars, during which both updated interpretations of earlier breakthroughs – as with Hugo Junkers' pioneering of all-metal airframes in 1915 leading to giant multi-engined aircraft of up to 60+ meter wingspan sizes by the early 1930s, adoption of the mostly air-cooled radial engine as a practical aircraft powerplant alongside powerful V-12 liquid-cooled aviation engines, and ever-greater instances of long-distance flight attempts – as with a Vickers Vimy in 1919, followed only months later by the U.S. Navy's NC-4 transatlantic flight; culminating in May 1927 with Charles Lindbergh's solo trans-Atlantic flight in the Spirit of St. Louis spurring ever-longer flight attempts, pioneering the way for long-distance flights of the future to become commonplace. Airplanes had a presence in all the major battles of World War II. They were an essential component of the military strategies of the period, such as the German Blitzkrieg or the American and Japanese aircraft carrier campaigns of the Pacific. Military gliders were developed and used in several campaigns, but they did not become widely used due to the high casualty rate often encountered. The Focke-Achgelis Fa 330 Bachstelze (Wagtail) rotor kite of 1942 was notable for its use by German submarines. Before and during the war, both British and German designers were developing jet engines to power airplanes. The first jet aircraft to fly, in 1939, was the German Heinkel He 178. In 1943, the first operational jet fighter, the Messerschmitt Me 262, went into service with the German Luftwaffe and later in the war the British Gloster Meteor entered service but never saw action – top airspeeds of aircraft for that era went as high as 1,130 km/h (700 mph), with the early July 1944 unofficial record flight of the German Me 163B V18 rocket fighter prototype. In October 1947, the Bell X-1 was the first aircraft to exceed the speed of sound. In 1948–49, aircraft transported supplies during the Berlin Blockade. New aircraft types, such as the B-52, were produced during the Cold War. The first jet airliner, the de Havilland Comet, was introduced in 1952, followed by the Soviet Tupolev Tu-104 in 1956. The Boeing 707, the first widely successful commercial jet, was in commercial service for more than 50 years, from 1958 to 2010. The Boeing 747 was the world's biggest passenger aircraft from 1970 until it was surpassed by the Airbus A380 in 2005. The most successful aircraft is the Douglas DC-3, a medium twin engined passenger aircraft that is still in sevice since 1936 and is still being used in small amounts for skydiving and other recreational flights. With some of the thousands of other versions being used for other purposes, like the AC-47, a Vietnam War era gunship, which is still used in some militaries today. An airplane (also known as an aeroplane or simply a plane) is a powered fixed-wing aircraft that is propelled forward by thrust from a jet engine or propeller. Planes come in a variety of sizes, shapes, and wing configurations. The broad spectrum of uses for planes includes recreation, transportation of goods and people, military, and research. A seaplane is a fixed-wing aircraft capable of taking off and landing (alighting) on water. Seaplanes that can also operate from dry land are a subclass called amphibian aircraft. These aircraft were sometimes called hydroplanes. Seaplanes and amphibians are usually divided into two categories based on their technological characteristics: floatplanes and flying boats. Many forms of glider (see below) may be modified by adding a small power plant. These include: A ground effect vehicle (GEV) is a craft that attains level flight near the surface of the earth, making use of the ground effect – an aerodynamic interaction between the wings and the earth's surface. Some GEVs are able to fly higher out of ground effect (OGE) when required – these are classed as powered fixed-wing aircraft. A glider is a heavier-than-air craft that is supported in flight by the dynamic reaction of the air against its lifting surfaces, and whose free flight does not depend on an engine. A sailplane is a fixed-wing glider designed for soaring – the ability to gain height in updrafts of air and to fly for long periods. Gliders are mainly used for recreation, but have also been used for other purposes such as aerodynamics research, warfare and recovering spacecraft. A motor glider does have an engine for extending its performance and some have engines powerful enough to take off, but the engine is not used in normal flight. As is the case with planes, there are a wide variety of glider types differing in the construction of their wings, aerodynamic efficiency, location of the pilot and controls. Perhaps the most familiar type is the toy paper plane. Large gliders are most commonly launched by a tow-plane or by a winch. Military gliders have been used in war to deliver assault troops, and specialized gliders have been used in atmospheric and aerodynamic research. Rocket-powered aircraft and spaceplanes have also made unpowered landings. Gliders and sailplanes that are used for the sport of gliding have high aerodynamic efficiency. The highest lift-to-drag ratio is 70:1, though 50:1 is more common. After launch, further energy is obtained through the skillful exploitation of rising air in the atmosphere. Flights of thousands of kilometers at average speeds over 200 km/h have been achieved. The most numerous unpowered aircraft are paper airplanes, a handmade type of glider. Like hang gliders and paragliders, they are foot-launched and are in general slower, smaller, and less expensive than sailplanes. Hang gliders most often have flexible wings given shape by a frame, though some have rigid wings. Paragliders and paper airplanes have no frames in their wings. Gliders and sailplanes can share a number of features in common with powered aircraft, including many of the same types of fuselage and wing structures. For example, the Horten H.IV was a tailless flying wing glider, and the delta wing-shaped Space Shuttle orbiter flew much like a conventional glider in the lower atmosphere. Many gliders also use similar controls and instruments as powered craft. The main application today of glider aircraft is sport and recreation. Gliders were developed from the 1920s for recreational purposes. As pilots began to understand how to use rising air, sailplane gliders were developed with a high lift-to-drag ratio. These allowed longer glides to the next source of "lift", and so increase their chances of flying long distances. This gave rise to the popular sport of gliding. Early gliders were mainly built of wood and metal but the majority of sailplanes now use composite materials incorporating glass, carbon or aramid fibers. To minimize drag, these types have a streamlined fuselage and long narrow wings having a high aspect ratio. Both single-seat and two-seat gliders are available. Initially training was done by short "hops" in primary gliders which are very basic aircraft with no cockpit and minimal instruments. Since shortly after World War II training has always been done in two-seat dual control gliders, but high performance two-seaters are also used to share the workload and the enjoyment of long flights. Originally skids were used for landing, but the majority now land on wheels, often retractable. Some gliders, known as motor gliders, are designed for unpowered flight, but can deploy piston, rotary, jet or electric engines. Gliders are classified by the FAI for competitions into glider competition classes mainly on the basis of span and flaps. A class of ultralight sailplanes, including some known as microlift gliders and some known as "airchairs", has been defined by the FAI based on a maximum weight. They are light enough to be transported easily, and can be flown without licensing in some countries. Ultralight gliders have performance similar to hang gliders, but offer some additional crash safety as the pilot can be strapped in an upright seat within a deformable structure. Landing is usually on one or two wheels which distinguishes these craft from hang gliders. Several commercial ultralight gliders have come and gone, but most current development is done by individual designers and home builders. Military gliders were used during World War II for carrying troops (glider infantry) and heavy equipment to combat zones. The gliders were towed into the air and most of the way to their target by military transport planes, e.g. C-47 Dakota, or by bombers that had been relegated to secondary activities, e.g. Short Stirling. Once released from the tow near the target, they landed as close to the target as possible. The advantage over paratroopers were that heavy equipment could be landed and that the troops were quickly assembled rather than being dispersed over a drop zone. The gliders were treated as disposable, leading to construction from common and inexpensive materials such as wood, though a few were retrieved and re-used. By the time of the Korean War, transport aircraft had also become larger and more efficient so that even light tanks could be dropped by parachute, causing gliders to fall out of favor. Even after the development of powered aircraft, gliders continued to be used for aviation research. The NASA Paresev Rogallo flexible wing was originally developed to investigate alternative methods of recovering spacecraft. Although this application was abandoned, publicity inspired hobbyists to adapt the flexible-wing airfoil for modern hang gliders. Initial research into many types of fixed-wing craft, including flying wings and lifting bodies was also carried out using unpowered prototypes. A hang glider is a glider aircraft in which the pilot is ensconced in a harness suspended from the airframe, and exercises control by shifting body weight in opposition to a control frame. Most modern hang gliders are made of an aluminum alloy or composite-framed fabric wing. Pilots have the ability to soar for hours, gain thousands of meters of altitude in thermal updrafts, perform aerobatics, and glide cross-country for hundreds of kilometers. A paraglider is a lightweight, free-flying, foot-launched glider aircraft with no rigid primary structure. The pilot sits in a harness suspended below a hollow fabric wing whose shape is formed by its suspension lines, the pressure of air entering vents in the front of the wing and the aerodynamic forces of the air flowing over the outside. Paragliding is most often a recreational activity. A paper plane is a toy aircraft (usually a glider) made out of paper or paperboard. Model glider aircraft are models of aircraft using lightweight materials such as polystyrene and balsa wood. Designs range from simple glider aircraft to accurate scale models, some of which can be very large. Glide bombs are bombs with aerodynamic surfaces to allow a gliding flightpath rather than a ballistic one. This enables the carrying aircraft to attack a heavily defended target from a distance. A kite is an aircraft tethered to a fixed point so that the wind blows over its wings. Lift is generated when air flows over the kite's wing, producing low pressure above the wing and high pressure below it, and deflecting the airflow downwards. This deflection also generates horizontal drag in the direction of the wind. The resultant force vector from the lift and drag force components is opposed by the tension of the one or more rope lines or tethers attached to the wing. Kites are mostly flown for recreational purposes, but have many other uses. Early pioneers such as the Wright Brothers and J.W. Dunne sometimes flew an aircraft as a kite in order to develop it and confirm its flight characteristics, before adding an engine and flight controls, and flying it as an airplane. Kites have been used for signaling, for delivery of munitions, and for observation, by lifting an observer above the field of battle, and by using kite aerial photography. Kites have been used for scientific purposes, such as Benjamin Franklin's famous experiment proving that lightning is electricity. Kites were the precursors to the traditional aircraft, and were instrumental in the development of early flying craft. Alexander Graham Bell experimented with very large man-lifting kites, as did the Wright brothers and Lawrence Hargrave. Kites had a historical role in lifting scientific instruments to measure atmospheric conditions for weather forecasting. Kites can be used to carry radio antennas. This method was used for the reception station of the first transatlantic transmission by Marconi. Captive balloons may be more convenient for such experiments, because kite-carried antennas require a lot of wind, which may be not always possible with heavy equipment and a ground conductor. Kites can be used to carry light effects such as lightsticks or battery powered lights. Kites can be used to pull people and vehicles downwind. Efficient foil-type kites such as power kites can also be used to sail upwind under the same principles as used by other sailing craft, provided that lateral forces on the ground or in the water are redirected as with the keels, center boards, wheels and ice blades of traditional sailing craft. In the last two decades, several kite sailing sports have become popular, such as kite buggying, kite landboarding, kite boating and kite surfing. Snow kiting has also become popular. Kite sailing opens several possibilities not available in traditional sailing: Conceptual research and development projects are being undertaken by over a hundred participants to investigate the use of kites in harnessing high altitude wind currents for electricity generation. Kite festivals are a popular form of entertainment throughout the world. They include local events, traditional festivals and major international festivals. The structural parts of a fixed-wing aircraft are called the airframe. The parts present can vary according to the aircraft's type and purpose. Early types were usually made of wood with fabric wing surfaces, When engines became available for a powered flight around a hundred years ago, their mounts were made of metal. Then as speeds increased more and more parts became metal until by the end of World War II all-metal aircraft were common. In modern times, increasing use of composite materials has been made. Typical structural parts include: The wings of a fixed-wing aircraft are static planes extending to either side of the aircraft. When the aircraft travels forwards, air flows over the wings which are shaped to create lift. Kites and some light weight gliders and airplanes have flexible wing surfaces which are stretched across a frame and made rigid by the lift forces exerted by the airflow over them. Larger aircraft have rigid wing surfaces which provide additional strength. Whether flexible or rigid, most wings have a strong frame to give them their shape and to transfer lift from the wing surface to the rest of the aircraft. The main structural elements are one or more spars running from root to tip, and many ribs running from the leading (front) to the trailing (rear) edge. Early airplane engines had little power and light weight was very important. Also, early aerofoil sections were very thin, and could not have strong frame installed within. So until the 1930s, most wings were too light weight to have enough strength and external bracing struts and wires were added. When the available engine power increased during the 1920s and 1930s, wings could be made heavy and strong enough that bracing was not needed anymore. This type of unbraced wing is called a cantilever wing. The number and shape of the wings vary widely on different types. A given wing plane may be full-span or divided by a central fuselage into port (left) and starboard (right) wings. Occasionally, even more, wings have been used, with the three-winged triplane achieving some fame in World War I. The four-winged quadruplane and other multiplane designs have had little success. A monoplane, which derives from the prefix, mono means one which means it has a single wing plane, a biplane has two stacked one above the other, a tandem wing has two placed one behind the other. When the available engine power increased during the 1920s and 1930s and bracing was no longer needed, the unbraced or cantilever monoplane became the most common form of powered type. The wing planform is the shape when seen from above. To be aerodynamically efficient, a wing should be straight with a long span from side to side but have a short chord (high aspect ratio). But to be structurally efficient, and hence lightweight, a wing must have a short span but still enough area to provide lift (low aspect ratio). At transonic speeds, near the speed of sound, it helps to sweep the wing backward or forwards to reduce drag from supersonic shock waves as they begin to form. The swept wing is just a straight wing swept backward or forwards. The delta wing is a triangle shape which may be used for a number of reasons. As a flexible Rogallo wing it allows a stable shape under aerodynamic forces, and so is often used for kites and other ultralight craft. As a supersonic wing, it combines high strength with low drag and so is often used for fast jets. A variable geometry wing can be changed in flight to a different shape. The variable-sweep wing transforms between an efficient straight configuration for takeoff and landing, to a low-drag swept configuration for high-speed flight. Other forms of variable planform have been flown, but none have gone beyond the research stage. A fuselage is a long, thin body, usually with tapered or rounded ends to make its shape aerodynamically smooth. Most fixed-wing aircraft have a single fuselage, often referred to as simply "the body". Others may have two or more fuselages, or the fuselage may be fitted with booms on either side of the tail to allow the extreme rear of the fuselage to be utilized. The fuselage may contain a flight crew, passengers, cargo or payload, fuel and engine(s). Pilotless aircraft (drones) do not usually have a pilot or any other flight crew or passengers on board. Gliders usually have no fuel or engines, although some variations such as motor gliders and rocket gliders have them for temporary or optional use. Pilots of manned fixed-wing aircraft usually control them from inside a cockpit, usually located at the front or top of the fuselage, equipped with controls and usually windows and instruments. Aircraft often have two or more pilots, with one in overall command (the "pilot") and one or more "co-pilots". On larger aircraft a navigator is typically also seated in the cockpit as well. Some military or specialized aircraft may have other flight crew members in the cockpit as well. In small aircraft, the passengers are typically seated behind the pilot(s) in the same cabin, although occasionally a passenger seat may be beside or even in front of the pilot. Larger passenger aircraft have a separate passenger cabin or occasionally cabins which are physically separated from the cockpit. A flying wing is a tailless aircraft which has no definite fuselage, with most of the crew, payload and equipment being housed inside the main wing structure. The flying wing configuration was studied extensively in the 1930s and 1940s, notably by Jack Northrop and Cheston L. Eshelman in the United States, and Alexander Lippisch and the Horten brothers in Germany. After the war, a number of experimental designs were based on the flying wing concept. Some general interest continued until the early 1950s, but designs did not necessarily offer a great advantage in range and presented a number of technical problems, leading to the adoption of "conventional" solutions like the Convair B-36 and the B-52 Stratofortress. Due to the practical need for a deep wing, the flying wing concept is most practical for designs in the slow-to-medium speed range, and there has been continual interest in using it as a tactical airlifter design. Interest in flying wings was renewed in the 1980s due to their potentially low radar reflection cross-sections. Stealth technology relies on shapes which only reflect radar waves in certain directions, thus making the aircraft hard to detect unless the radar receiver is at a specific position relative to the aircraft – a position that changes continuously as the aircraft moves. This approach eventually led to the Northrop B-2 Spirit stealth bomber. In this case the aerodynamic advantages of the flying wing are not the primary needs. However, modern computer-controlled fly-by-wire systems allowed for many of the aerodynamic drawbacks of the flying wing to be minimized, making for an efficient and stable long-range bomber. Blended wing body aircraft have a flattened and airfoil shaped body, which produces most of the lift to keep itself aloft, and distinct and separate wing structures, though the wings are smoothly blended in with the body. Thus blended wing bodied aircraft incorporate design features from both a futuristic fuselage and flying wing design. The purported advantages of the blended wing body approach are efficient high-lift wings and a wide airfoil-shaped body. This enables the entire craft to contribute to lift generation with the result of potentially increased fuel economy. A lifting body is a configuration in which the body itself produces lift. In contrast to a flying wing, which is a wing with minimal or no conventional fuselage, a lifting body can be thought of as a fuselage with little or no conventional wing. Whereas a flying wing seeks to maximize cruise efficiency at subsonic speeds by eliminating non-lifting surfaces, lifting bodies generally minimize the drag and structure of a wing for subsonic, supersonic, and hypersonic flight, or, spacecraft re-entry. All of these flight regimes pose challenges for proper flight stability. Lifting bodies were a major area of research in the 1960s and 1970s as a means to build a small and lightweight manned spacecraft. The US built a number of famous lifting body rocket planes to test the concept, as well as several rocket-launched re-entry vehicles that were tested over the Pacific. Interest waned as the US Air Force lost interest in the manned mission, and major development ended during the Space Shuttle design process when it became clear that the highly shaped fuselages made it difficult to fit fuel tankage. The classic aerofoil section wing is unstable in flight and difficult to control. Flexible-wing types often rely on an anchor line or the weight of a pilot hanging beneath to maintain the correct attitude. Some free-flying types use an adapted aerofoil that is stable, or other ingenious mechanisms including, most recently, electronic artificial stability. But in order to achieve trim, stability and control, most fixed-wing types have an empennage comprising a fin and rudder which act horizontally and a tailplane and elevator which act vertically. This is so common that it is known as the conventional layout. Sometimes there may be two or more fins, spaced out along the tailplane. Some types have a horizontal "canard" foreplane ahead of the main wing, instead of behind it. This foreplane may contribute to the trim, stability or control of the aircraft, or to several of these. Kites are controlled by wires running down to the ground. Typically each wire acts as a tether to the part of the kite it is attached to. Gliders and airplanes have more complex control systems, especially if they are piloted. The main controls allow the pilot to direct the aircraft in the air. Typically these are: Other common controls include: A craft may have two pilots' seats with dual controls, allowing two pilots to take turns. This is often used for training or for longer flights. The control system may allow full or partial automation of flight, such as an autopilot, a wing leveler, or a flight management system. An unmanned aircraft has no pilot but is controlled remotely or via means such as gyroscopes or other forms of autonomous control. On manned fixed-wing aircraft, instruments provide information to the pilots, including flight, engines, navigation, communications, and other aircraft systems that may be installed. The six basic instruments, sometimes referred to as the "six pack", are as follows: Other cockpit instruments might include: (Wayback Machine copy)
{{main|Aerospace engineering}} [[Image:2006 Ojiya balloon festival 011.jpg|thumb|A [[hot air balloon]] is a kind of non-powered aircraft]] An '''aircraft''' is a [[flight|flying]] [[machine]]. The word ''aircraft'' originally meant airships and balloons. It comes from the words ''air'' and ''craft'', a term from boating as were many early aviation words. There are many different kinds of aircraft. Some aircraft keep in the sky by moving air over their [[wing|wings]]. Examples are [[aeroplane|aeroplanes]], [[helicopter|helicopters]], and [[glider|gliders]]. Some aircraft keep in the sky by [[buoyancy|floating]]. Examples are [[balloon (aircraft)|balloons]] and [[airship|airships]]. Most aircraft use [[engine]] power. Examples are [[aeroplane]]s, [[helicopter]], and [[airship|airships]]. Gliders and balloons use no power. A few aircraft use [[human powered aircraft|muscle power]]. Big aeroplanes for transporting people are called ''[[airliner]]s''. Airliners are the quickest way to travel. Airliners can fly over mountains and bad weather. Airliners have complex technology to make them fly quickly, safely and for less money. A few [[fighter aircraft]] can fly at 3,200&nbsp;km/h (2,000&nbsp;mph). The person who drives an aircraft is called the [[pilot]]. == Largest aircraft == These are the largest aircrafts. They are: *The largest aircraft by [[Dimension|dimensions]] and [[volume]] is the British [[Airlander 10]].<ref>{{Cite news|last=Video|first=Telegraph|date=2016-08-17|title=World's largest aircraft the Airlander makes maiden flight in UK|language=en-GB|work=The Telegraph|url=https://www.telegraph.co.uk/news/2016/08/17/worlds-largest-aircraft-the-airlander-takes-first-flight/|access-date=2020-09-21|issn=0307-1235}}</ref> *The largest aircraft by [[weight]] and largest regularly used [[fixed-wing aircraft]] is the [[Antonov An-225 Mriya|Antonov An-225 ''Mriya'']]''.''<ref>{{Cite web|last=FlyDriveReizen.nl|title=Fly Drive Aanbiedingen|url=https://www.flydrivereizen.nl/aanbiedingen/|access-date=2020-09-21|website=FlyDriveReizen.nl|language=nl}}</ref><ref>{{Cite web|title=Aerospaceweb.org {{!}} Ask Us - Largest Plane in the World|url=http://www.aerospaceweb.org/question/design/q0188.shtml|access-date=2020-09-21|website=www.aerospaceweb.org}}</ref> *The largest military aircraft and the second largest [[airplane]] is the [[Antonov An-124 Ruslan|Antonov An-124 ''Ruslan'']]''.''<ref>{{Cite web|last=Administrator|first=NASA|date=2013-06-06|title=World's Second Largest Aircraft|url=http://www.nasa.gov/centers/glenn/multimedia/imagegallery/image_feature_028_Antonov.html|access-date=2020-09-21|website=NASA|language=en|archive-date=2021-03-08|archive-url=https://web.archive.org/web/20210308103434/https://www.nasa.gov/centers/glenn/multimedia/imagegallery/image_feature_028_Antonov.html|url-status=dead}}</ref> == Related pages == *[[Evacuation slide]] *[[Flight recorder]] ==References== {{commonscat}} {{reflist}} {{aircraft}} {{authority control}} [[Category:Aircraft| ]] [[Category:Aerospace engineering]] {{transport-stub}}
An aircraft (pl.: aircraft) is a vehicle that is able to fly by gaining support from the air. It counters the force of gravity by using either static lift or the dynamic lift of an airfoil, or, in a few cases, direct downward thrust from its engines. Common examples of aircraft include airplanes, helicopters, airships (including blimps), gliders, paramotors, and hot air balloons. The human activity that surrounds aircraft is called aviation. The science of aviation, including designing and building aircraft, is called aeronautics. Crewed aircraft are flown by an onboard pilot, whereas unmanned aerial vehicles may be remotely controlled or self-controlled by onboard computers. Aircraft may be classified by different criteria, such as lift type, aircraft propulsion (if any), usage and others. Flying model craft and stories of manned flight go back many centuries; however, the first manned ascent — and safe descent — in modern times took place by larger hot-air balloons developed in the 18th century. Each of the two World Wars led to great technical advances. Consequently, the history of aircraft can be divided into five eras: Aerostats use buoyancy to float in the air in much the same way that ships float on the water. They are characterized by one or more large cells or canopies, filled with a relatively low-density gas such as helium, hydrogen, or hot air, which is less dense than the surrounding air. When the weight of this is added to the weight of the aircraft structure, it adds up to the same weight as the air that the craft displaces. Small hot-air balloons, called sky lanterns, were first invented in ancient China prior to the 3rd century BC and used primarily in cultural celebrations, and were only the second type of aircraft to fly, the first being kites, which were first invented in ancient China over two thousand years ago (see Han Dynasty). A balloon was originally any aerostat, while the term airship was used for large, powered aircraft designs — usually fixed-wing. In 1919, Frederick Handley Page was reported as referring to "ships of the air," with smaller passenger types as "Air yachts." In the 1930s, large intercontinental flying boats were also sometimes referred to as "ships of the air" or "flying-ships". — though none had yet been built. The advent of powered balloons, called dirigible balloons, and later of rigid hulls allowing a great increase in size, began to change the way these words were used. Huge powered aerostats, characterized by a rigid outer framework and separate aerodynamic skin surrounding the gas bags, were produced, the Zeppelins being the largest and most famous. There were still no fixed-wing aircraft or non-rigid balloons large enough to be called airships, so "airship" came to be synonymous with these aircraft. Then several accidents, such as the Hindenburg disaster in 1937, led to the demise of these airships. Nowadays a "balloon" is an unpowered aerostat and an "airship" is a powered one. A powered, steerable aerostat is called a dirigible. Sometimes this term is applied only to non-rigid balloons, and sometimes dirigible balloon is regarded as the definition of an airship (which may then be rigid or non-rigid). Non-rigid dirigibles are characterized by a moderately aerodynamic gasbag with stabilizing fins at the back. These soon became known as blimps. During World War II, this shape was widely adopted for tethered balloons; in windy weather, this both reduces the strain on the tether and stabilizes the balloon. The nickname blimp was adopted along with the shape. In modern times, any small dirigible or airship is called a blimp, though a blimp may be unpowered as well as powered. Heavier-than-air aircraft, such as airplanes, must find some way to push air or gas downwards so that a reaction occurs (by Newton's laws of motion) to push the aircraft upwards. This dynamic movement through the air is the origin of the term. There are two ways to produce dynamic upthrust — aerodynamic lift, and powered lift in the form of engine thrust. Aerodynamic lift involving wings is the most common, with fixed-wing aircraft being kept in the air by the forward movement of wings, and rotorcraft by spinning wing-shaped rotors sometimes called "rotary wings." A wing is a flat, horizontal surface, usually shaped in cross-section as an aerofoil. To fly, air must flow over the wing and generate lift. A flexible wing is a wing made of fabric or thin sheet material, often stretched over a rigid frame. A kite is tethered to the ground and relies on the speed of the wind over its wings, which may be flexible or rigid, fixed, or rotary. With powered lift, the aircraft directs its engine thrust vertically downward. V/STOL aircraft, such as the Harrier jump jet and Lockheed Martin F-35B take off and land vertically using powered lift and transfer to aerodynamic lift in steady flight. A pure rocket is not usually regarded as an aerodyne because it does not depend on the air for its lift (and can even fly into space); however, many aerodynamic lift vehicles have been powered or assisted by rocket motors. Rocket-powered missiles that obtain aerodynamic lift at very high speed due to airflow over their bodies are a marginal case. The forerunner of the fixed-wing aircraft is the kite. Whereas a fixed-wing aircraft relies on its forward speed to create airflow over the wings, a kite is tethered to the ground and relies on the wind blowing over its wings to provide lift. Kites were the first kind of aircraft to fly and were invented in China around 500 BC. Much aerodynamic research was done with kites before test aircraft, wind tunnels, and computer modelling programs became available. The first heavier-than-air craft capable of controlled free-flight were gliders. A glider designed by George Cayley carried out the first true manned, controlled flight in 1853. The first powered and controllable fixed-wing aircraft (the airplane or aeroplane) was invented by Wilbur and Orville Wright. Besides the method of propulsion (if any), fixed-wing aircraft are in general characterized by their wing configuration. The most important wing characteristics are: A variable geometry aircraft can change its wing configuration during flight. A flying wing has no fuselage, though it may have small blisters or pods. The opposite of this is a lifting body, which has no wings, though it may have small stabilizing and control surfaces. Wing-in-ground-effect vehicles are generally not considered aircraft. They "fly" efficiently close to the surface of the ground or water, like conventional aircraft during takeoff. An example is the Russian ekranoplan nicknamed the "Caspian Sea Monster". Man-powered aircraft also rely on ground effect to remain airborne with minimal pilot power, but this is only because they are so underpowered—in fact, the airframe is capable of flying higher. Rotorcraft, or rotary-wing aircraft, use a spinning rotor with aerofoil cross-section blades (a rotary wing) to provide lift. Types include helicopters, autogyros, and various hybrids such as gyrodynes and compound rotorcraft. Helicopters have a rotor turned by an engine-driven shaft. The rotor pushes air downward to create lift. By tilting the rotor forward, the downward flow is tilted backward, producing thrust for forward flight. Some helicopters have more than one rotor and a few have rotors turned by gas jets at the tips. Some have a tail rotor to counteract the rotation of the main rotor, and to aid directional control. Autogyros have unpowered rotors, with a separate power plant to provide thrust. The rotor is tilted backward. As the autogyro moves forward, air blows upward across the rotor, making it spin. This spinning increases the speed of airflow over the rotor, to provide lift. Rotor kites are unpowered autogyros, which are towed to give them forward speed or tethered to a static anchor in high-wind for kited flight. Compound rotorcraft have wings that provide some or all of the lift in forward flight. They are nowadays classified as powered lift types and not as rotorcraft. Tiltrotor aircraft (such as the Bell Boeing V-22 Osprey), tiltwing, tail-sitter, and coleopter aircraft have their rotors/propellers horizontal for vertical flight and vertical for forward flight. The smallest aircraft are toys/recreational items, and nano aircraft. The largest aircraft by dimensions and volume (as of 2016) is the 302 ft (92 m) long British Airlander 10, a hybrid blimp, with helicopter and fixed-wing features, and reportedly capable of speeds up to 90 mph (140 km/h; 78 kn), and an airborne endurance of two weeks with a payload of up to 22,050 lb (10,000 kg). The largest aircraft by weight and largest regular fixed-wing aircraft ever built, as of 2016, was the Antonov An-225 Mriya. That Soviet-built (Ukrainian SSR) six-engine transport of the 1980s was 84 m (276 ft) long, with an 88 m (289 ft) wingspan. It holds the world payload record, after transporting 428,834 lb (194,516 kg) of goods, and has flown 100 t (220,000 lb) loads commercially. With a maximum loaded weight of 550–700 t (1,210,000–1,540,000 lb), it was also the heaviest aircraft built to date. It could cruise at 500 mph (800 km/h; 430 kn). The aircraft was destroyed during the Russo-Ukrainian War. The largest military airplanes are the Ukrainian Antonov An-124 Ruslan (world's second-largest airplane, also used as a civilian transport), and American Lockheed C-5 Galaxy transport, weighing, loaded, over 380 t (840,000 lb). The 8-engine, piston/propeller Hughes H-4 Hercules "Spruce Goose" — an American World War II wooden flying boat transport with a greater wingspan (94m/260ft) than any current aircraft and a tail height equal to the tallest (Airbus A380-800 at 24.1m/78ft) — flew only one short hop in the late 1940s and never flew out of ground effect. The largest civilian airplanes, apart from the above-noted An-225 and An-124, are the Airbus Beluga cargo transport derivative of the Airbus A300 jet airliner, the Boeing Dreamlifter cargo transport derivative of the Boeing 747 jet airliner/transport (the 747-200B was, at its creation in the 1960s, the heaviest aircraft ever built, with a maximum weight of over 400 t (880,000 lb)), and the double-decker Airbus A380 "super-jumbo" jet airliner (the world's largest passenger airliner). The fastest fixed-wing aircraft and fastest glider, is the Space Shuttle, which re-entered the atmosphere at nearly Mach 25 or 17,500 mph (28,200 km/h) The fastest recorded powered aircraft flight and fastest recorded aircraft flight of an air-breathing powered aircraft was of the NASA X-43A Pegasus, a scramjet-powered, hypersonic, lifting body experimental research aircraft, at Mach 9.68 or 6,755 mph (10,870 km/h) on 16 November 2004. Prior to the X-43A, the fastest recorded powered airplane flight, and still the record for the fastest manned powered airplane, was the North American X-15, rocket-powered airplane at Mach 6.7 or 7,274 km/h (4,520 mph) on 3 October 1967. The fastest manned, air-breathing powered airplane is the Lockheed SR-71 Blackbird, a U.S. reconnaissance jet fixed-wing aircraft, having reached 3,530 km/h (2,193 mph) on 28 July 1976. Gliders are heavier-than-air aircraft that do not employ propulsion once airborne. Take-off may be by launching forward and downward from a high location, or by pulling into the air on a tow-line, either by a ground-based winch or vehicle, or by a powered "tug" aircraft. For a glider to maintain its forward air speed and lift, it must descend in relation to the air (but not necessarily in relation to the ground). Many gliders can "soar", i.e., gain height from updrafts such as thermal currents. The first practical, controllable example was designed and built by the British scientist and pioneer George Cayley, whom many recognise as the first aeronautical engineer. Common examples of gliders are sailplanes, hang gliders and paragliders. Balloons drift with the wind, though normally the pilot can control the altitude, either by heating the air or by releasing ballast, giving some directional control (since the wind direction changes with altitude). A wing-shaped hybrid balloon can glide directionally when rising or falling; but a spherically shaped balloon does not have such directional control. Kites are aircraft that are tethered to the ground or other object (fixed or mobile) that maintains tension in the tether or kite line; they rely on virtual or real wind blowing over and under them to generate lift and drag. Kytoons are balloon-kite hybrids that are shaped and tethered to obtain kiting deflections, and can be lighter-than-air, neutrally buoyant, or heavier-than-air. Powered aircraft have one or more onboard sources of mechanical power, typically aircraft engines although rubber and manpower have also been used. Most aircraft engines are either lightweight reciprocating engines or gas turbines. Engine fuel is stored in tanks, usually in the wings but larger aircraft also have additional fuel tanks in the fuselage. Propeller aircraft use one or more propellers (airscrews) to create thrust in a forward direction. The propeller is usually mounted in front of the power source in tractor configuration but can be mounted behind in pusher configuration. Variations of propeller layout include contra-rotating propellers and ducted fans. Many kinds of power plant have been used to drive propellers. Early airships used man power or steam engines. The more practical internal combustion piston engine was used for virtually all fixed-wing aircraft until World War II and is still used in many smaller aircraft. Some types use turbine engines to drive a propeller in the form of a turboprop or propfan. Human-powered flight has been achieved, but has not become a practical means of transport. Unmanned aircraft and models have also used power sources such as electric motors and rubber bands. Jet aircraft use airbreathing jet engines, which take in air, burn fuel with it in a combustion chamber, and accelerate the exhaust rearwards to provide thrust. Different jet engine configurations include the turbojet and turbofan, sometimes with the addition of an afterburner. Those with no rotating turbomachinery include the pulsejet and ramjet. These mechanically simple engines produce no thrust when stationary, so the aircraft must be launched to flying speed using a catapult, like the V-1 flying bomb, or a rocket, for example. Other engine types include the motorjet and the dual-cycle Pratt & Whitney J58. Compared to engines using propellers, jet engines can provide much higher thrust, higher speeds and, above about 40,000 ft (12,000 m), greater efficiency. They are also much more fuel-efficient than rockets. As a consequence nearly all large, high-speed or high-altitude aircraft use jet engines. Some rotorcraft, such as helicopters, have a powered rotary wing or rotor, where the rotor disc can be angled slightly forward so that a proportion of its lift is directed forwards. The rotor may, like a propeller, be powered by a variety of methods such as a piston engine or turbine. Experiments have also used jet nozzles at the rotor blade tips. Aircraft are designed according to many factors such as customer and manufacturer demand, safety protocols and physical and economic constraints. For many types of aircraft the design process is regulated by national airworthiness authorities. The key parts of an aircraft are generally divided into three categories: The approach to structural design varies widely between different types of aircraft. Some, such as paragliders, comprise only flexible materials that act in tension and rely on aerodynamic pressure to hold their shape. A balloon similarly relies on internal gas pressure, but may have a rigid basket or gondola slung below it to carry its payload. Early aircraft, including airships, often employed flexible doped aircraft fabric covering to give a reasonably smooth aeroshell stretched over a rigid frame. Later aircraft employed semi-monocoque techniques, where the skin of the aircraft is stiff enough to share much of the flight loads. In a true monocoque design there is no internal structure left. The key structural parts of an aircraft depend on what type it is. Lighter-than-air types are characterised by one or more gasbags, typically with a supporting structure of flexible cables or a rigid framework called its hull. Other elements such as engines or a gondola may also be attached to the supporting structure. Heavier-than-air types are characterised by one or more wings and a central fuselage. The fuselage typically also carries a tail or empennage for stability and control, and an undercarriage for takeoff and landing. Engines may be located on the fuselage or wings. On a fixed-wing aircraft the wings are rigidly attached to the fuselage, while on a rotorcraft the wings are attached to a rotating vertical shaft. Smaller designs sometimes use flexible materials for part or all of the structure, held in place either by a rigid frame or by air pressure. The fixed parts of the structure comprise the airframe. The source of motive power for an aircraft is normally called the powerplant, and includes engine or motor, propeller or rotor, (if any), jet nozzles and thrust reversers (if any), and accessories essential to the functioning of the engine or motor (e.g.: starter, ignition system, intake system, exhaust system, fuel system, lubrication system, engine cooling system, and engine controls). Powered aircraft are typically powered by internal combustion engines (piston or turbine) burning fossil fuels -- typically gasoline (avgas) or jet fuel. A very few are powered by rocket power, ramjet propulsion, or by electric motors, or by internal combustion engines of other types, or using other fuels. A very few have been powered, for short flights, by human muscle energy (e.g.: Gossamer Condor). The avionics comprise any electronic aircraft flight control systems and related equipment, including electronic cockpit instrumentation, navigation, radar, monitoring, and communications systems. The flight envelope of an aircraft refers to its approved design capabilities in terms of airspeed, load factor and altitude. The term can also refer to other assessments of aircraft performance such as maneuverability. When an aircraft is abused, for instance by diving it at too-high a speed, it is said to be flown outside the envelope, something considered foolhardy since it has been taken beyond the design limits which have been established by the manufacturer. Going beyond the envelope may have a known outcome such as flutter or entry to a non-recoverable spin (possible reasons for the boundary). The range is the distance an aircraft can fly between takeoff and landing, as limited by the time it can remain airborne. For a powered aircraft the time limit is determined by the fuel load and rate of consumption. For an unpowered aircraft, the maximum flight time is limited by factors such as weather conditions and pilot endurance. Many aircraft types are restricted to daylight hours, while balloons are limited by their supply of lifting gas. The range can be seen as the average ground speed multiplied by the maximum time in the air. The Airbus A350-900ULR is now the longest range airliner. Flight dynamics is the science of air vehicle orientation and control in three dimensions. The three critical flight dynamics parameters are the angles of rotation around three axes which pass through the vehicle's center of gravity, known as pitch, roll, and yaw. Flight dynamics is concerned with the stability and control of an aircraft's rotation about each of these axes. An aircraft that is unstable tends to diverge from its intended flight path and so is difficult to fly. A very stable aircraft tends to stay on its flight path and is difficult to maneuver. Therefore, it is important for any design to achieve the desired degree of stability. Since the widespread use of digital computers, it is increasingly common for designs to be inherently unstable and rely on computerised control systems to provide artificial stability. A fixed wing is typically unstable in pitch, roll, and yaw. Pitch and yaw stabilities of conventional fixed wing designs require horizontal and vertical stabilisers, which act similarly to the feathers on an arrow. These stabilizing surfaces allow equilibrium of aerodynamic forces and to stabilise the flight dynamics of pitch and yaw. They are usually mounted on the tail section (empennage), although in the canard layout, the main aft wing replaces the canard foreplane as pitch stabilizer. Tandem wing and tailless aircraft rely on the same general rule to achieve stability, the aft surface being the stabilising one. A rotary wing is typically unstable in yaw, requiring a vertical stabiliser. A balloon is typically very stable in pitch and roll due to the way the payload is slung underneath the center of lift. Flight control surfaces enable the pilot to control an aircraft's flight attitude and are usually part of the wing or mounted on, or integral with, the associated stabilizing surface. Their development was a critical advance in the history of aircraft, which had until that point been uncontrollable in flight. Aerospace engineers develop control systems for a vehicle's orientation (attitude) about its center of mass. The control systems include actuators, which exert forces in various directions, and generate rotational forces or moments about the aerodynamic center of the aircraft, and thus rotate the aircraft in pitch, roll, or yaw. For example, a pitching moment is a vertical force applied at a distance forward or aft from the aerodynamic center of the aircraft, causing the aircraft to pitch up or down. Control systems are also sometimes used to increase or decrease drag, for example to slow the aircraft to a safe speed for landing. The two main aerodynamic forces acting on any aircraft are lift supporting it in the air and drag opposing its motion. Control surfaces or other techniques may also be used to affect these forces directly, without inducing any rotation. Aircraft permit long distance, high speed travel and may be a more fuel efficient mode of transportation in some circumstances. Aircraft have environmental and climate impacts beyond fuel efficiency considerations, however. They are also relatively noisy compared to other forms of travel and high altitude aircraft generate contrails, which experimental evidence suggests may alter weather patterns. Aircraft are produced in several different types optimized for various uses; military aircraft, which includes not just combat types but many types of supporting aircraft, and civil aircraft, which include all non-military types, experimental and model. A military aircraft is any aircraft that is operated by a legal or insurrectionary armed service of any type. Military aircraft can be either combat or non-combat: Most military aircraft are powered heavier-than-air types. Other types, such as gliders and balloons, have also been used as military aircraft; for example, balloons were used for observation during the American Civil War and World War I, and military gliders were used during World War II to land troops. Civil aircraft divide into commercial and general types, however there are some overlaps. Commercial aircraft include types designed for scheduled and charter airline flights, carrying passengers, mail and other cargo. The larger passenger-carrying types are the airliners, the largest of which are wide-body aircraft. Some of the smaller types are also used in general aviation, and some of the larger types are used as VIP aircraft. General aviation is a catch-all covering other kinds of private (where the pilot is not paid for time or expenses) and commercial use, and involving a wide range of aircraft types such as business jets (bizjets), trainers, homebuilt, gliders, warbirds and hot air balloons to name a few. The vast majority of aircraft today are general aviation types. An experimental aircraft is one that has not been fully proven in flight, or that carries a Special Airworthiness Certificate, called an Experimental Certificate in United States parlance. This often implies that the aircraft is testing new aerospace technologies, though the term also refers to amateur-built and kit-built aircraft, many of which are based on proven designs. A model aircraft is a small unmanned type made to fly for fun, for static display, for aerodynamic research or for other purposes. A scale model is a replica of some larger design.
[[File:2009-0605-Victoria-Harbor-PAN.JPG|thumb|300px|Victoria's Inner Harbour with The Empress hotel at the left and government buildings at the center]] '''Victoria''' is the [[capital city]] of the [[province]] of [[British Columbia]], [[Canada]]. It is named after [[Queen Victoria]]. Over 350,000 people live in the Victoria area. Many people work for the [[government]] or in [[tourism]]. It is warmer than other Canadian cities in the winter, so many Canadians move there when they retire. Victoria is on [[Vancouver Island]]. Most people travel there by boat (ferry) from [[Vancouver]], or from [[Washington]] State. Victoria has three ferry docks and an airport, Victoria International Airport. Victoria has one of the mildest climates in Canada, with gardens blooming year-round. Some tourist attractions are the Butchart Gardens and the Royal British Columbia [[Museum]]. Some of the best-known buildings are the government buildings and the Empress Hotel, and both are by the Inner Harbour. The first people to live in the area were the Coast Salish native people. People from England built Fort Victoria in 1843. Victoria became a city in 1862. 18,000 students study at the [[University]] of Victoria. Other people study at Camosun [[College]], Royal Roads University, University Canada West, Victoria College of Art, and the Canadian College of Performing Arts. Students from all over the world study at Pearson College, near Victoria. The Canadian navy has ships in a part of the city called Esquimalt. Some of the parks include Beacon Hill Park, East Sooke Park, Mount Work Park, and Mount Douglas Park. There are many flowers in Victoria; it is called the "city of gardens." It is a lot drier than places nearby, and has little snow. Some well-known people from Victoria are: * [[Jamie Benn]] (hockey) * [[Emily Carr]] (art) * [[Steve Nash]] (basketball) * [[David Foster]] (music) * [[Nelly Furtado]] (singer) * [[Kyle Germaine Johnson]] (Pro Biker) == Regions of Victoria == The Victoria area is made up of 13 parts. Four of them make up Greater Victoria: * Victoria * [[Saanich]] * Esquimalt * Oak Bay Three of them make up the Saanich [[Peninsula]]: * Sidney * Central Saanich * North Saanich The others make up the Western Communities: * Colwood * Langford * View Royal * Metchosin * Sooke * Highlands {{Canada capitals}} {{multistub|Geo|Canada}} {{authority control}} [[Category:Victoria, British Columbia| ]]
Victoria is the capital city of the Canadian province of British Columbia, on the southern tip of Vancouver Island off Canada's Pacific coast. The city has a population of 91,867, and the Greater Victoria area has a population of 397,237. The city of Victoria is the seventh most densely populated city in Canada with 4,406 inhabitants per square kilometre (11,410/sq mi). Victoria is the southernmost major city in Western Canada and is about 100 km (62 mi) southwest from British Columbia's largest city of Vancouver on the mainland. The city is about 100 km (62 mi) from Seattle by airplane, seaplane, ferry, or the Victoria Clipper passenger-only ferry, and 40 km (25 mi) from Port Angeles, Washington, by ferry Coho across the Strait of Juan de Fuca. Named for Queen Victoria, the city is one of the oldest in the Pacific Northwest, with British settlement beginning in 1843. The city has retained a large number of its historic buildings, in particular its two most famous landmarks, the Parliament Buildings (finished in 1897 and home of the Legislative Assembly of British Columbia) and the Empress Hotel (opened in 1908). The city's Chinatown is the second oldest in North America, after that of San Francisco. The region's Coast Salish First Nations peoples established communities in the area long before European settlement, which had large populations at the time of European exploration. Known as "the Garden City", Victoria is an attractive city and a popular tourism destination and has a regional technology sector that has risen to be its largest revenue-generating private industry. Victoria is in the top 20 world cities for quality of life, according to Numbeo. Prior to the arrival of European navigators in the late 1700s, the Greater Victoria area was home to several communities of Coast Salish peoples, including the Lək̓ʷəŋən (Lekwungen) and W̱SÁNEĆ (Saanich) peoples. The Spanish and British took up the exploration of the northwest coast, beginning with the visits of Juan Pérez in 1774, and of James Cook in 1778. Although the Victoria area of the Strait of Juan de Fuca was not explored until 1790, Spanish sailors visited Esquimalt Harbour (just west of Victoria proper) in 1790, 1791, and 1792. In 1841 James Douglas was charged with the duty of setting up a trading post on the southern tip of Vancouver Island. Upon the recommendation by George Simpson a new more northerly post should be built in case Fort Vancouver fell into American hands (see Oregon boundary dispute). Douglas founded Fort Victoria on the site of present-day Victoria in anticipation of the outcome of the Oregon Treaty in 1846, extending the British North America/United States border along the 49th parallel from the Rockies to the Strait of Georgia. Erected in 1843 as a Hudson's Bay Company trading post on a site originally called Camosack meaning "rush of water". Known briefly as "Fort Albert", the settlement was renamed Fort Victoria in November 1843, in honour of Queen Victoria. The Songhees established a village across the harbour from the fort. The Songhees' village was later moved north of Esquimalt in 1911.The crown colony was established in 1849. Between the years 1850–1854 a series of treaty agreements known as the Douglas Treaties were made with indigenous communities to purchase certain plots of land in exchange for goods. These agreements contributed to a town being laid out on the site and made the capital of the colony, though controversy has followed about the ethical negotiation and upholding of rights by the colonial government. The superintendent of the fort, Chief Factor James Douglas was made the second governor of the Vancouver Island Colony (Richard Blanshard was first governor, Arthur Edward Kennedy was third and last governor), and would be the leading figure in the early development of the city until his retirement in 1864. When news of the discovery of gold on the British Columbia mainland reached San Francisco in 1858, Victoria became the port, supply base, and outfitting centre for miners on their way to the Fraser Canyon gold fields, mushrooming from a population of 300 to over 5000 within a few days. Victoria was incorporated as a city in 1862. In 1862 Victoria was the epicentre of the 1862 Pacific Northwest smallpox epidemic which devastated First Nations, killing about two-thirds of all natives in British Columbia. In 1865, the North Pacific home of the Royal Navy was established in Esquimalt and today is Canada's Pacific coast naval base. In 1866 when the island was politically united with the mainland, Victoria was designated the capital of the new united colony instead of New Westminster – an unpopular move on the Mainland – and became the provincial capital when British Columbia joined the Canadian Confederation in 1871. In the latter half of the 19th century, the Port of Victoria became one of North America's largest importers of opium, serving the opium trade from Hong Kong and distribution into North America. Opium trade was legal and unregulated until 1865, when the legislature issued licences and levied duties on its import and sale. The opium trade was banned in 1908. In 1886, with the completion of the Canadian Pacific Railway terminus on Burrard Inlet, Victoria's position as the commercial centre of British Columbia was irrevocably lost to the city of Vancouver. The city subsequently began cultivating an image of genteel civility within its natural setting, aided by the impressions of visitors such as Rudyard Kipling, the opening of the popular Butchart Gardens in 1904 and the construction of the Empress Hotel by the Canadian Pacific Railway in 1908. Robert Dunsmuir, a leading industrialist whose interests included coal mines and a railway on Vancouver Island, constructed Craigdarroch Castle in the Rockland area, near the official residence of the province's Lieutenant Governor. His son James Dunsmuir became Premier and subsequently Lieutenant Governor of the province and built his own grand residence at Hatley Park (used for several decades as Royal Roads Military College, now civilian Royal Roads University) in the present City of Colwood. A real-estate and development boom ended just before World War I, leaving Victoria with a large stock of Edwardian public, commercial and residential buildings that have greatly contributed to the city's character. With the economic crash and an abundance of unmarried men, Victoria became an excellent location for military recruiting. Two militia infantry battalions, the 88th Victoria Fusiliers and the 50th Gordon Highlanders, formed in the immediate pre-war period. Victoria was the home of Sir Arthur Currie. He had been a high-school teacher and real-estate agent prior to the war and was the Commanding Officer of the Gordon Highlanders in the summer of 1914. Before the end of the war he commanded the Canadian Corps. A number of municipalities surrounding Victoria were incorporated during this period, including the Township of Esquimalt, the District of Oak Bay, and several municipalities on the Saanich Peninsula. Since World War II the Victoria area has seen relatively steady growth, becoming home to two major universities. Since the 1980s the western suburbs have been incorporated as new municipalities, such as Colwood and Langford, which are known collectively as the Western Communities. Greater Victoria periodically experiences calls for the amalgamation of the thirteen municipal governments within the Capital Regional District. The opponents of amalgamation state that separate governance affords residents a greater deal of local autonomy. The proponents of amalgamation argue it would reduce duplication of services, while allowing for more efficient use of resources and the ability to better handle broad, regional issues and long-term planning. The landscape of Victoria was formed by volcanism followed by water in various forms. Pleistocene glaciation put the area under a thick ice cover, the weight of which depressed the land below present sea level. These glaciers also deposited stony sandy loam till. As they retreated, their melt water left thick deposits of sand and gravel. Marine clay settled on what would later become dry land. Post-glacial rebound exposed the present-day terrain to air, raising beach and mud deposits well above sea level. The resulting soils are highly variable in texture, and abrupt textural changes are common. In general, clays are most likely to be encountered in the northern part of town and in depressions. The southern part has coarse-textured subsoils and loamy topsoils. Sandy loams and loamy sands are common in the eastern part adjoining Oak Bay. Victoria's soils are relatively unleached and less acidic than soils elsewhere on the British Columbia Coast. Their thick dark topsoils denote a high level of fertility which made them valuable for farming prior to urbanization. Depending on the classification used, Victoria either has a warm-summer Mediterranean or oceanic climate (Köppen: Csb, Trewartha: Do); with fresh, dry, sunny summers, and cool, cloudy, rainy winters. Victoria is farther north than many "cold-winter" cities, such as Ottawa, Quebec City, and Minneapolis. However, westerly winds and Pacific Ocean currents keep Victoria's winter temperatures substantially higher, with an average January temperature of 5.0 °C (41.0 °F) compared to Ottawa, the nation's capital, with −10.2 °C (13.6 °F). At the Victoria Gonzales weather station, daily temperatures rise above 30 °C (86 °F) on average less than one day per year and fall below 0 °C (32 °F) on average only ten nights per year. Victoria has recorded completely freeze-free winter seasons four times (in 1925–26, 1939–40, 1999–2000, and 2002–03). 1999 is the only calendar year on record without a single occurrence of frost. During this time the city went 718 days without freezing, starting on 23 December 1998 and ending 10 December 2000. The second longest frost-free period was a 686-day stretch covering 1925 and 1926, marking the first and last time the city has gone the entire season without dropping below 1 °C (34 °F). During the winter, the average daily high and low temperatures are 8 and 4 °C (46 and 39 °F), respectively. The summer months are also relatively mild, with an average high temperature of 20 °C (68 °F) and low of 11 °C (52 °F), although inland areas often experience warmer daytime highs. The highest temperature ever recorded at Victoria Gonzales was 39.8 °C (103.6 °F) on 28 June 2021; The coldest temperature on record is −15.6 °C (3.9 °F) on 29 December 1968. The average annual temperature varies from a high of 11.4 °C (52.5 °F) set in 2004 to a low of 8.6 °C (47.5 °F) set in 1916. Due to the rain shadow effect of the nearby Olympic Mountains, Victoria is the driest location on the British Columbia coast and one of the driest in the region. Average precipitation amounts in the Greater Victoria area range from 608 mm (23.9 in) at the Gonzales observatory in the City of Victoria to 1,124 mm (44.3 in) in nearby Langford. The Victoria Airport, 25 km (16 mi) north of the city, receives about 45% more precipitation than the city proper. Regional average precipitation amounts range from as low as 406 mm (16.0 in) on the north shore of the Olympic Peninsula to 3,505 mm (138.0 in) in Port Renfrew just 80 km (50 mi) away on the more exposed southwest coast of Vancouver Island. Vancouver measures 1,589 mm (62.6 in) annually and Seattle is at 952 mm (37.5 in). One feature of Victoria's climate is its distinct dry and rainy seasons. Nearly two-thirds of the annual precipitation falls during the four wettest months, November to February. Precipitation in December, the wettest month (109 mm [4.3 in]) is nearly eight times as high as in July, the driest month (14 mm [0.55 in]). Victoria experiences the driest summers in Canada (outside of the extreme northern reaches of the Northwest Territories and Nunavut). Victoria averages just 26 cm (10 in) of snow annually, about half that of Vancouver. Roughly one third of winters see virtually no snow, with less than 5 cm (2.0 in) falling during the entire season. When snow does fall, it rarely lasts long on the ground. Victoria averages just two or three days per year with at least 5 cm (2.0 in) of snow on the ground. Every few decades Victoria receives very large snowfalls including the record breaking 100 cm (39 in) of snow that fell in December 1996. That amount places Victoria 3rd for biggest snowfall among major cities in Canada. With 2,193 hours of bright sunshine annually during the last available measurement period, Victoria is effectively tied with Cranbrook as the sunniest city in British Columbia. In July 2013, Victoria received 432.8 hours of bright sunshine, which is the most sunshine ever recorded in any month in British Columbia history. Victoria's equable climate has also added to its reputation as the "City of Gardens". The city takes pride in the many flowers that bloom during the winter and early spring, including crocuses, daffodils, early-blooming rhododendrons, cherry and plum trees. Every February there is an annual "flower count" in what for the rest of the country and most of the province is still the dead of winter. Due to its mild climate, Victoria and its surrounding area (southeastern Vancouver Island, Gulf Islands, and parts of the Lower Mainland and Sunshine Coast) are also home to many rare, native plants found nowhere else in Canada, including Quercus garryana (Garry oak), Arctostaphylos columbiana (hairy manzanita), and Canada's only broad-leaf evergreen tree, Arbutus menziesii (Pacific madrone). Many of these species exist here, at the northern end of their range, and are found as far south as southern California and parts of Mexico. Non-native plants grown in Victoria include the cold-hardy palm Trachycarpus fortunei, which can be found in gardens and public areas of Victoria. One of these Trachycarpus palms stands in front of City Hall. In the 2021 Census of Population conducted by Statistics Canada, Victoria had a population of 91,867 living in 49,222 of its 53,070 total private dwellings, a change of 7.1% from its 2016 population of 85,792. With a land area of 19.45 km (7.51 sq mi), it had a population density of 4,723.2/km (12,233.1/sq mi) in 2021. Victoria is one of the most gender diverse cities in Canada, with approximately 0.75% of residents identifying as transgender or non-binary in the 2021 Statistics Canada Census of Population. At the census metropolitan area (CMA) level in the 2021 census, the Victoria CMA had a population of 397,237 living in 176,676 of its 186,674 total private dwellings, a change of 8% from its 2016 population of 367,770. With a land area of 695.29 km (268.45 sq mi), it had a population density of 571.3/km (1,479.7/sq mi) in 2021. Victoria is known for its disproportionately large retiree population. Some 23.4 percent of the population of Victoria and its surrounding area are over 65 years of age, which is higher than the overall Canadian distribution of over 65 year-olds in the population (19%). A historically popular cliché refers to Victoria as the home of "the newly wed and nearly dead". According to the 2021 census, the majority of the population of Victoria described themselves as irreligious (63.4%). Over 25% of Victoria residents are Christian, with the second largest religious group being Muslim (1.9%). A similar proportion of residents are Buddhist (1.4%) or Jewish (1.1%). Hinduism, Sikhism and Indigenous Spirituality make up under 1% of other groups. The following is a list of neighbourhoods in the City of Victoria, as defined by the city planning department. For a list of neighbourhoods in other area municipalities, see Greater Victoria, or the individual entries for those municipalities. Informal neighbourhoods include: A point-in-time homeless count was conducted by volunteers between March 11 and March 12, 2020, that counted at least 1,523 homeless that night. The homeless count is considered an underestimate due to the hidden homeless that may be couch surfing or have found somewhere to stay that is not on the street or homeless shelters. The first homeless count was conducted in January 2005 by the Victoria Cool Aid Society and counted a homeless population of approximately 700 individuals. Like many west coast cities in North America the homeless population is both concentrated in specific areas (parts of Pandora avenue in Victoria) and is often outside due to milder climates that make homelessness more visible year-round. The 2020 point-in-time homeless count found 35% respondents identified as being Indigenous. This is over representative in the homeless population as only 4.7% of the overall population of Victoria identify as Indigenous. During the COVID-19 pandemic, many homeless people sheltered in camping tents within the city's parks and some roadside greenspaces, including in Beacon Hill Park. In March 2021, city council reinstated a bylaw prohibiting daytime camping in parks, and with support from the provincial government, pledged to find indoor accommodation for all those camping in parks. Homeless campers from parks and other public spaces were housed temporarily in motels, the Save-on-Foods arena and a tiny home village on a portion of the Royal Athletic Park's parking lot. The city's chief industries are technology, tourism, education, federal and provincial government administration and services. Other nearby employers include the Canadian Forces (the Township of Esquimalt is the home of the Pacific headquarters of the Royal Canadian Navy), and the University of Victoria (in the municipalities of Oak Bay and Saanich) and Camosun College in Saanich (which have over 33,000 faculty, staff and students combined). Other sectors of the Greater Victoria area economy include: investment and banking, online book publishing, various public and private schools, food products manufacturing, light aircraft manufacturing (in North Saanich), technology products, various high tech firms in pharmaceuticals and computers, engineering, architecture and telecommunications. There are three major shopping malls in the City of Victoria, including the Bay Centre, Hillside Shopping Centre, and Mayfair Shopping Centre. Mayfair, one of the first major shopping centres in Victoria, first opened as an outdoor strip mall on 16 October 1963 with 27 stores. It was built on the site of a former brickyard in the Maywood district, a then-semi-rural area in the northern part of Victoria. Woodward's was Mayfair's original department store anchor upon the mall's opening. Mayfair was enclosed and renovated into an indoor mall in 1974. The mall underwent three later expansions in 1984 (with the addition of Consumers Distributing), 1985 (expansion of the mall food court) and a major expansion in 1990 that saw the addition of more retail space. The Bay (now Hudson's Bay) replaced Woodward's as Mayfair's department store anchor in 1993 following Hudson's Bay Company's acquisition of the Woodward's chain. The mall was more recently renovated in 2019. Mayfair now offers over 100 stores and services including Hudson's Bay. It has 42,197.8 m (454,213 sq ft) of retail space and it also provides customers with rooftop parking. Advanced technology is Victoria's largest revenue-producing private industry with $3.15 billion in annual revenues generated by more than 880 tech companies that have over 15,000 direct employees. The annual economic impact of the sector is estimated at more than $4.03 billion per year. With three post-secondary institutions in Saanich, eight federal research labs in the region, and Canada's Pacific Navy Base in Esquimalt, Victoria relies heavily upon the neighbouring communities for economic activity and as employment hubs. The region has many of the elements required for a strong technology sector, including Canada's highest household internet usage. Over a hundred technology, software and engineering companies have an office in Victoria. Victoria is a major tourism destination with over 3.5 million overnight visitors per year who add more than a billion dollars to the local economy. As well, over 500,000 daytime visitors arrive via cruise ships which dock at Ogden Point near the city's Inner Harbour. Many whale watching tour companies operate from this harbour due to the whales often present near its coast. The city is also close to Canadian Forces Base Esquimalt, the Canadian Navy's primary Pacific Ocean naval base. Downtown Victoria also serves as Greater Victoria's regional downtown, where many night clubs, theatres, restaurants and pubs are clustered, and where many regional public events occur. Canada Day fireworks displays, Symphony Splash, and many other music festivals and cultural events draw tens of thousands of Greater Victorians and visitors to the downtown core. The Rifflandia and Electronic Music Festival are other music events that draw crowds to the downtown core. Victoria relies upon neighbouring communities for many recreational opportunities including ice rinks in Oak Bay and Saanich. Victoria has one small public pool (Crystal Pool) and many residents use larger and newer pool facilities in Oak Bay, and Saanich (Commonwealth Pool and Gordon Head Pool). The city and metro region has hosted high-profile sports events including the 1994 Commonwealth Games which hosted track events at the Saanich-Oak Bay based University of Victoria and the Saanich Commonwealth Pool, the 2009 Scotties Tournament of Hearts, the 2005 Ford World Men's Curling Championship tournament, and 2006 Skate Canada. Victoria co-hosted the 2007 FIFA U-20 World Cup at Royal Athletic Park, and is the venue for the Bastion Square Grand Prix Criterium road cycling race. The city is also a destination for conventions, meetings, and conferences, including a 2007 North Atlantic Treaty Organization military chief of staff meeting held at the Hotel Grand Pacific. Every year, the Swiftsure International Yacht Race attracts boaters from around the world to participate in the boat race in the waters off of Vancouver Island, and the Victoria Dragon Boat Festival brings over 90 teams from around North America. The Tall Ships Festival brings sailing ships to the city harbour. Victoria also hosts the start of the Vic-Maui Yacht Race, the longest offshore sailboat race on the West Coast. The Port of Victoria consists of three parts, the Outer Harbour, used by deep sea vessels, the Inner and Upper Harbours, used by coastal and industrial traffic. It is protected by a breakwater with a deep and wide opening. The port is a working harbour, tourist attraction and cruise destination. Esquimalt Harbour is also a well-protected harbour with a large graving dock and shipbuilding and repair facilities. The Victoria Symphony, led by Christian Kluxen, performs at the Royal Theatre and the Farquhar Auditorium of the Saanich-Oak Bay sited University of Victoria from September to May. Every BC Day weekend, the Symphony mounts Symphony Splash, an outdoor event that includes a performance by the orchestra sitting on a barge in Victoria's Inner Harbour. Streets in the local area are closed, as each year approximately 40,000 people attend a variety of concerts and events throughout the day. The event culminates with the Symphony's evening concert, with Tchaikovsky's 1812 Overture as the grand finale, complete with cannon fire from Royal Canadian Sea Cadet Gunners from HMCS QUADRA, a pealing carillon and a fireworks display to honour BC Day. Pacific Opera Victoria, Victoria Operatic Society, Victoria Philharmonic Choir, Canadian Pacific Ballet and Ballet Victoria stage two or three productions each year at the Macpherson or Royal Theatres. The Bastion Theatre, a professional dramatic company, functioned in Victoria through the 1970s and 1980s and performed high quality dramatic productions but ultimately declared bankruptcy in 1988. Reborn as The New Bastion Theatre in 1990 the company struggled for two more years before closing operations in 1992. The Belfry Theatre started in 1974 as the Springridge Cultural Centre in 1974. The venue was renamed the Belfry Theatre in 1976 as the company began producing its own shows. The Belfry's mandate is to produce contemporary plays with an emphasis on new Canadian plays. Other regional theatre venues include: Phoenix Theatre student theatre at the University of Victoria; The Roxy Theatre, home of the Blue Bridge Repertory Theatre company; Kaleidoscope Theatre and Intrepid Theatre, producers of the Victoria Fringe Theatre Festival and The Uno Festival of Solo Performance. The only Canadian Forces Primary Reserve brass/reed band on Vancouver Island is in Victoria. The 5th (British Columbia) Field Regiment, Royal Canadian Artillery Band traces its roots back to 1864, making it the oldest, continually operational military band west of Thunder Bay, Ontario. Its mandate is to support the island's military community by performing at military dinners, parades and ceremonies, and other events. The band performs weekly in August at Fort Rodd Hill National Historic Site where the Regiment started manning the guns of the fort in 1896, and also performs every year at the Cameron Bandshell at Beacon Hill Park. The annual multi-day Rifflandia Music Festival is one of Canada's largest modern rock and pop music festivals. A number of well-known musicians and bands are from the Victoria area, including Nelly Furtado, David Foster, The Moffatts, Frog Eyes, Johnny Vallis, Jets Overhead, Bryce Soderberg, Armchair Cynics, Nomeansno, The New Colors, Wolf Parade, The Racoons, Tal Bachman, Dayglo Abortions, Hot Hot Heat, Aidan Knight and Noah Becker. Due to the proximity to Vancouver and a 6% distance location tax credit, Victoria is used as a filming location for many films, television series, and television movies. Some of these films include X2, X-Men: The Last Stand, In the Land of Women, White Chicks, Scary Movie, Final Destination, Excess Baggage and Bird on a Wire. Television series such as Smallville, The Dead Zone and Poltergeist: The Legacy were also filmed there. Canadian director Atom Egoyan was raised in neighbouring Saanich. Actors Cameron Bright (Ultraviolet, X-Men: The Last Stand, Thank You for Smoking, New Moon) and Ryan Robbins (Stargate Atlantis, Battlestar Galactica, Sanctuary) were born in Victoria. Actor Cory Monteith from the television series Glee was raised in Victoria. Actor, artist, and athlete Duncan Regehr of Star Trek: Deep Space Nine was raised in the region. Artist, art magazine publisher and jazz saxophonist Noah Becker of Whitehot Magazine has been a long time Victoria resident. Nobel laureate Alice Munro lived in Victoria during the years when she published her first story collections and co-founded Munro's Books. Victoria resident Stanley Evans has written a series of mysteries featuring a Coast Salish character, Silas Seaweed, who works as an investigator with the Victoria Police Department. Other Victoria writers include Kit Pearson, Esi Edugyan, Robert Wiersema, W. D. Valgardson, Elizabeth Louisa Moresby, Madeline Sonik, Jack Hodgins, Dave Duncan, Bill Gaston, David Gurr, Ken Steacy, Sheryl McFarlane, Carol Shields and Patrick Lane. Gayleen Froese's 2005 novel Touch is set in Victoria. The comedy troupe LoadingReadyRun is based in Victoria. In the heart of downtown are the British Columbia Parliament Buildings, The Empress Hotel, Victoria Police Department Station Museum, the gothic Christ Church Cathedral, and the Royal British Columbia Museum/IMAX National Geographic Theatre, with large exhibits on local Aboriginal peoples, natural history, and modern history, along with travelling international exhibits. In addition, the heart of downtown also has the Maritime Museum of British Columbia, Emily Carr House, Victoria Bug Zoo, and Market Square. The oldest (and most intact) Chinatown in Canada is within downtown and includes the Chinese Public School built in 1909, and some cultural items and pictures displayed at the Pandora avenue entrances to Market Square. The Art Gallery of Greater Victoria is close to downtown in the Rockland neighbourhood several city blocks from Craigdarroch Castle built by industrialist Robert Dunsmuir and Government House, the official residence of the Lieutenant-Governor of British Columbia. Numerous other buildings of historic importance or interest are also in central Victoria, including: the 1845 St. Ann's Schoolhouse; the 1852 Helmcken House built for Victoria's first doctor; the 1863 Congregation Emanu-El, the oldest synagogue in continuous use in Canada; the 1865 Angela College built as Victoria's first Anglican Collegiate School for Girls, now housing retired nuns of the Sisters of St. Ann; the 1871 St. Ann's Academy built as a Catholic school; the 1874 Church of Our Lord, built to house a breakaway congregation from the Anglican Christ Church cathedral; the 1890 St. Andrew's Presbyterian Church; the 1890 Metropolitan Methodist Church (now the Victoria Conservatory of Music), which is publicly open for faculty, student, and guest performances, also acts as Camosun College Music Department; the 1892 St. Andrew's Cathedral; and the 1925 Crystal Gardens, originally a saltwater swimming pool, restored as a conservatory and most recently a tourist attraction called the B.C. Experience, which closed down in 2006. Downtown Victoria is a very walkable area with many midblock crosswalks, an expanding central pedestrian street, public squares, and alleys that are predominantly spaces for pedestrians. Fan Tan alley is the narrowest commercial street in North America and runs between Pandora avenue and Fisgard street in Victoria's Chinatown. Dragon alley is also located in Chinatown and is a mix of commercial and residential units, located between Fisgard and Herald streets. Theatre alley was rebuilt in a newer condo construction in Chinatown and is a narrow alley that winds between Pandora avenue and Fisgard street just west of Fan Tan alley, but it does not include direct access to any commercial businesses. Waddington alley is uniquely paved with wooden blocks located between Yates and Johnson streets. Trounce alley is a small commercial alley located between Government and Broad streets. Beacon Hill Park is the central city's main urban green space. Its area of 75 ha (190 acres) adjacent to Victoria's southern shore includes numerous playing fields, manicured gardens, exotic species of plants and animals such as wild peacocks, a petting zoo, and views of the Strait of Juan de Fuca and the Olympic Mountains in Washington across it. The sport of cricket has been played in Beacon Hill Park since the mid-19th century. Each summer, the City of Victoria presents dozens of concerts at the Cameron Band Shell in Beacon Hill Park. The extensive system of parks in Victoria also includes a few areas of natural Garry oak meadow habitat, an increasingly scarce ecosystem that once dominated the region. Private gardens that are open to the public with sometimes limited opening hours are located throughout the city and offer access at low or no cost to visitors, they include the rose garden next to the Empress hotel, the gardens on the grounds of the Lieutenant Governor of British Columbia's house (also known as Government House) on Rockland Road, and Abkahazi Garden on Fairfield Road. Dallas Road is a waterfront trail and road with 7.1 km to walk, run, bike or drive. Clover Point is its main rest area with benches, lounge chairs, picnic tables and a public washroom. The David Foster Harbour Pathway is a predominantly a pedestrian pathway that meanders around the inner harbour between the southern start at Ogden point by the cruise ship terminal and Rock Bay at its northern terminus. The pathway has some disconnected sections that are expected to be connected with redevelopments along the pathway near the Johnson street bridge. When completed the David Foster Harbour Pathway is expected to extend over 5 kilometres in length. CFB Esquimalt navy base, in the adjacent municipality of Esquimalt, has a base museum dedicated to naval and military history, in the Naden part of the base. North of the city on the Saanich Peninsula are the marine biology Shaw Ocean Discovery Centre, Butchart Gardens in Brentwood Bay, one of the biggest tourist and local resident attractions on Vancouver Island, as well as the Dominion Astrophysical Observatory, part of the National Research Council of Canada, Victoria Butterfly Gardens and Centre of the Universe planetarium. Victoria's climate, location and variety of facilities make it ideal for many recreational activities including rock climbing, hiking, kayaking, golf, water sports, informal team sports and jogging. Victoria is also known as the Cycling Capital of Canada, with hundreds of kilometres of bicycle paths, bike lanes and bike routes in the city, including the Galloping Goose Regional Trail. There are mountain biking trails at Mount Work Regional Park in the neghbour commmunity The District of Highlands, and Victoria is quickly becoming a bike tourism destination. Cycling advocacy groups including Greater Victoria Cycling Coalition (GVCC) and the Bike to Work Society have worked to improve Victoria's cycling infrastructure and facilities, and to make cycling a viable transportation alternative, attracting 5% of commuters in 2005. Greater Victoria also has a rich motorsports history, and was home to a 4/10ths mile oval race track called Western Speedway in the nearby City of Langford. Opened in 1954, Western Speedway was the oldest speedway in western Canada, and featured stock car racing, drag racing, demolition derbies and other events. Western Speedway was also home to the Victoria Auto Racing Hall of Fame and Museum. The Greater Victoria area also serves as a headquarters for Rugby Canada, based out of Starlight Stadium in Langford, as well as a headquarters for Rowing Canada, based out of Victoria City Rowing Club at Elk Lake in Saanich. The Greater Victoria Sports Hall of Fame is at the Save-on-Foods Memorial Centre, and features numerous displays and information on the sporting history of the city. The major sporting and entertainment complex, for Victoria and Vancouver Island Region, is the Save-On-Foods Memorial Centre arena. It replaced the former Victoria Memorial Arena, which was constructed by efforts of World War II veterans as a monument to fallen comrades. World War I, World War II, Korean War, and other conflict veterans are also commemorated. Fallen Canadian soldiers in past, present, and future wars and/or United Nations, NATO missions are noted, or will be noted by the main lobby monument at the Save-On-Foods Memorial Centre. The arena was the home of the ECHL (formerly known as the East Coast Hockey League) team, Victoria Salmon Kings, owned by RG Properties Limited, a real estate development firm that built the Victoria Save On Foods Memorial Centre, and Prospera Place Arena in Kelowna. The arena is the home of the Victoria Royals Western Hockey League (WHL) team that replaced the Victoria Salmon Kings (ECHL). Victoria has also been a destination for numerous high-profile international sporting events. It co-hosted the 1994 Commonwealth Games with Saanich and Oak Bay and the 2005 Ford World Men's Curling Championship. The 2007 FIFA U-20 World Cup was co-hosted by Victoria along with five other Canadian cities; (Burnaby, Toronto, Edmonton, Ottawa, Montreal). Victoria was also the first city location of the cross Canada 2010 Winter Olympics torch relay that occurred before the start of the 2010 Winter Olympics. Victoria co-hosted the 2019 World Junior Ice Hockey Championships along with Vancouver, British Columbia. Victoria was one of four host cities for the 2020 FIBA Men's Olympic Qualifying Tournaments in June 2020. The city has also been home to numerous high-profile sports teams in its history. The Victoria Cougars are perhaps the most famous sports franchise the city has known, existing as members of several professional leagues from 1911 to 1926, and again from 1949 to 1961. The Cougars won the Stanley Cup as members of the WCHL in 1925 after defeating the Montreal Canadiens three games to one in a best-of-five final. The Cougars were reincarnated in 1971 as a major junior hockey team in the Western Hockey League, before they moved to Prince George to become the Prince George Cougars. Today, the Cougars name and legacy continue in the form of the Junior 'B' team that plays in the Vancouver Island Junior Hockey League. Minor professional hockey returned to Victoria in the form of the Victoria Salmon Kings, which played in the ECHL from 2004 to 2011, and were a minor league affiliate of the Vancouver Canucks. In baseball, Victoria was once home of the Victoria Athletics of the Western International League, a Class 'A' minor league baseball affiliate of the New York Yankees. The Victoria region's newest sports team is Pacific FC of the Canadian Premier League. Pacific FC play their home matches at Starlight Stadium in Langford. Victoria has been home to many accomplished athletes that have participated in professional sports or the Olympics. Notable professional athletes include basketball Hall of Famer Steve Nash, twice Most Valuable Player in the National Basketball Association, who grew up in Victoria and played youth basketball at St. Michael's University School in Saanich and Mount Douglas Secondary School in Saanich. Furthermore, there are several current NHL hockey players from Greater Victoria, including brothers Jamie Benn and Jordie Benn of the Dallas Stars and Toronto Maple Leafs, respectively who grew up in North Saanich; Tyson Barrie of the Edmonton Oilers, and Matt Irwin of the Washington Capitals. Current Boston Red Sox pitcher Nick Pivetta was born in Victoria and played summer collegiate baseball for the Victoria HarbourCats. Former professional racing cyclist and 2012 Giro d'Italia winner, Ryder Hesjedal was born in Victoria and still calls the city home. Victoria has also been home to many Olympic Games athletes, including multi-time medalists such as Silken Laumann, Ryan Cochrane, and Simon Whitfield. Sports teams presently operating in Victoria include: Notable defunct teams that operated in Victoria include: The Jordan River Diversion Dam is Vancouver Island's main hydroelectric power station. It was built in 1911. The city's water is supplied by the Capital Regional District's Water Services Department from its Sooke Lake Reservoir. The lake is connected to a treatment plant at Japan Gulch by the 8.8 km (5.5 mi) Kapoor Tunnel. The lake water is very soft and requires no filtering. It is treated with chlorine, ammonia and ultraviolet light to control micro-organisms. Until the tunnel was completed in 1967, water flowed from the lake through the circuitous, leaky and much smaller 44 km (27 mi) Sooke Flowline. The Hartland landfill in Saanich is the waste disposal site for Greater Victoria area. Since 1985, it has been run by the Capital Regional District environmental services. It is on top of a hill, between Victoria and Sidney, at the end of Hartland Avenue (48°32′17″N 123°27′48″W / 48.53806°N 123.46333°W / 48.53806; -123.46333 (Hartland landfill)) There is a recycling centre, a sewer solid waste collection, hazardous waste collection, and an electricity generating station. This generating station now creates 1.6 megawatts of electricity, enough for 1,600 homes. The site has won international environmental awards. The CRD conducts public tours of the facility. It is predicted to be full by 2045. As of December 15, 2020 the CRD announced that core municipalities within Greater Victoria no longer discharge screened wastewater into the Strait of Juan de Fuca. The wastewater treatment facility was required to comply with federal regulations that forbid untreated discharge into waterways. The wastewater treatment project included pump stations at Clover Point and Macaulay Point in addition to the wastewater treatment plant at McLoughlin Point and the residuals treatment facility at Hartland landfill. The wastewater treatment plant serves Victoria, Esquimalt, Saanich, Oak Bay, View Royal, Langford, Colwood and the Esquimalt and Songhees First Nations. The Saanich Peninsula wastewater treatment plant serves North Saanich, Central Saanich and the Town of Sidney as well as the Victoria International Airport, the Institute of Ocean Sciences and the Tseycum and Pauquachin First Nations. This is a secondary level treatment plant which produces Class A biosolids. Victoria International Airport in North Saanich has non-stop flights to and from Toronto, Seattle, Montreal (seasonal), select seasonal sun destinations, and many cities throughout Western Canada. Multiple scheduled helicopter and seaplane flights are available daily from Victoria Inner Harbour Airport to Vancouver International Airport, Vancouver Harbour, and Seattle. Victoria is also home to the world's largest all-seaplane airline, Harbour Air. Harbour air offers flights during daylight hours at least every 30 minutes between Victoria's inner harbour and Vancouver's downtown terminal or YVR south terminal. Harbour Air also operate scenic tour flights over the Victoria harbour and gulf islands area. Due to Victoria's mild year round weather with mostly rainy winters, commuting by bicycle is more enjoyable year-round compared to many other Canadian cities. For this reason, the Greater Victoria area has the highest rate of bicycle commuting to work of any census metropolitan area in Canada. Greater Victoria also has an expanding system designed to facilitate cyclists, electrically assisted bicycles and other micromobility users via protected bike lanes on many roads, as well as separated multi-use paths for bicycles and pedestrians including the Galloping Goose Regional Trail, Lochside Regional Trail and the E&N rail trail. These multi-use trails are designed exclusively for foot traffic, cyclists, and micro-mobility users and pass through many communities in the Greater Victoria area, beginning at the downtown core and extending into areas such as Langford and Central and North Saanich. Victoria is currently finishing a 32 kilometre protected bike lane network intended for all ages and abilities (AAA). The first lane opened in Spring 2017 on Pandora Avenue, between Store Street and Cook Street in the downtown core and provides an easy cycling connection across the Johnson Street Bridge to the Galloping Goose Trail and E&N rail trail. The second protected bike lane in the network opened on Fort Street on May 27, 2018. The next two roads added to the downtown area bike network were Wharf and Humboldt streets, completed in 2019 and 2020 respectively, with Vancouver Street and Graham/Jackson streets added to the AAA bike network in 2021. The next round of streets upgraded starting in 2021 as "complete streets" with AAA cycling infrastructure included Richardson Street, Haultain Street, Government Street north of Pandora Avenue to Gorge Road, and finally Kimta Road connecting the network to the E&N rail trail. Connector routes in the Fernwood and Oaklands neighbourhoods to the Vancouver Street lanes were also constructed starting in 2021, avoiding hills and adding safer pedestrian and cyclist crossings. In 2022 the city constructed further AAA bicycle connections along Montreal street, Superior street, Government street (south, between Humboldt street and Dallas road), Fort street (between Cook street and up to the municipal border with Oak Bay), and Gorge Road (between Government street and up to the municipal border with Saanich). Go By Bike Week, previously called Bike to Work Week, is a bi-annual event held in communities throughout greater Victoria, British Columbia. It is organized by Capital Bike, a group created in 2021 by the merging of the Greater Victoria Cycling Coalition and Greater Victoria Bike to Work Society, and typically lasts one or two weeks. There is a large Spring event scheduled in late May every year, and again later during Fall typically in October. The original "Bike to Work Week" began in 1995 in Victoria and expanded to include other communities in BC through their local bicycle advocacy groups, all supported by the Bike to Work BC Society. The Bike to Work BC Society was formed in 2008 as a legal entity to run the event in other communities around BC, and was renamed the GoByBike BC Society to encourage cycling beyond the scope of commuting. The behaviour change (public health) model, relying on research conducted by both the provincial and federal governments that identified barriers to cycling and reasons for choosing cycling, was applied in the original Bike to Work Week event as a way to accomplish the goal of recruiting employees to bicycle to work. Since its inception, ridership in Go By Bike Week has steadily increased, and in 2017 over 7000 people participated in Greater Victoria. The event aims to attract new riders, promote cycling for commuting, recreation, and general transportation, and advocate for expanding safe cycling networks with prizes, activities and free cycling skills workshops. Pop-up "Celebration Stations" are set up throughout Greater Victoria, which typically feature free snacks and local coffee for cyclists, bicycle repair stands, and local cycling-related vendors and advocacy groups. The events were cancelled during the height of the COVID-19 pandemic, though individualized events were still promoted where participants could win prizes, and in-person events resumed in 2022. Other cycling advocacy initiatives in the Greater Victoria area include the Victoria chapter of Cycling Without Age, the Bike2Farm program and several recreational cycling clubs. The CRD is served by several ferries with the Lower Mainland, Gulf Islands and the United States. BC Ferries provides service between Swartz Bay, located on the northern tip of the Saanich Peninsula, to Tsawwassen on the Lower Mainland for cars, bus, trucks, pedestrians and cyclists. The Coho ferry operates as a car and pedestrian/cyclist ferry between the inner harbour of Victoria and Port Angeles, Washington. The Victoria Clipper is a pedestrian and cyclist-only (no vehicles) ferry which operates daily, year-round between downtown Seattle and the inner harbour of Victoria. The Washington State Ferries ran a ferry until 2020 for cars, pedestrians and cyclists between Friday Harbor, Orcas Island and Anacortes in Washington State from the port at Sidney, north of Victoria on the Saanich Peninsula. However, the service was shut down during the COVID-19 pandemic and did not resume. Washington State Ferries, citing crew and vessel shortages, estimates that it will not resume until at least 2030. Local public transportation is run by the Victoria Regional Transit System, which is part of BC Transit. Since 2000, double-decker buses have been introduced to the fleet, and have become an icon for the city. Rider fare payments can be made in cash, monthly bus passes, disability yearly passes, day-passes purchased from the driver or tickets purchased from a store. As of April 1, 2016 bus drivers do not provide transfers as proof of payment. Transfers were a source of disagreement and delay on the bus, due to improper transfer use, and disagreements over expired transfers or transfers used for return trips. Instead, a day-pass was added that can be purchased from the bus driver for $5, or two bus tickets (purchased from a retailer) for the equivalent of $4.50. To improve bus reliability and reduce delays, a bike and bus priority lane was opened in 2014 during peak traffic periods with fines for motorists operating in the bus/bike lane who are not turning in the same block. The dedicated bike and bus lane on Douglas street is being expanded from Downtown to near Uptown and may be changed to be restricted to only buses and bikes 24/7 rather than just during peak traffic periods depending on direction of travel. Most buses operating in the Greater Victoria area have a bike rack installed at the front of the bus that can accommodate two bicycles. Passenger rail service previously operated by Via Rail provided a single daily return trip along between Victoria – Courtenay, along the eastern coast of Vancouver Island, to the cities of Nanaimo, Courtenay, and points between. The service was discontinued along this line indefinitely on 19 March 2011, due to needed track replacement work. Prior to further inspection of the track, service along the segment between Nanaimo and Victoria was originally planned to resume on 8 April, but lack of funding has prevented any of the work from taking place and it is unclear when or if the service will resume. Local roadways are not based on a grid system as they are in Vancouver or Edmonton, and many streets do not follow a straight line from beginning to end as they wind around hills, parks, coastlines, and historic neighbourhoods, often changing names two or three times. There is no directional indication in street names that may be used in other cities with numbered roads where a street may run north–south or avenue may run east–west. The compact size of the city with few steep hills lends itself readily to smaller, fuel efficient alternatives to full size passenger cars, such as scooters, Smart Cars, motorcycles and electric bicycles. Victoria incentivizes the use of smaller modes of transport by offering smaller metered parking spaces in the downtown core specifically designated for small vehicles and motorcycles. Rush hour traffic delays along the Trans-Canada Highway to western suburban municipalities (including Langford, Colwood, Sooke and Metchosin) is commonly referred to as the "Colwood Crawl". Victoria serves as the western terminus (Mile Zero) for Canada's Trans-Canada Highway, the longest national highway in the world. The Mile Zero marker is at the southernmost point of Douglas Street where it meets Dallas Road along the waterfront. The Mile Zero location includes a statue to honour Terry Fox. Coach bus service between downtown Victoria and downtown Vancouver or the Vancouver International Airport, which includes the ferry fare is called the BC Ferries Connector run by Wilson's Transportation Limited. The coach bus travels on the ferry to Vancouver with separate trips for the bus to downtown and a bus to the Vancouver International Airport (YVR). Average travel time between the two cities is under 4 hours with an hour and half of that time spent on the ferry crossing. Bus service from Victoria to points up island is run by Island Link Bus or Tofino Bus. Both bus services depart from the Victoria bus terminal at 700 Douglas Street, behind the Fairmont Empress Hotel and offer trips to destinations further up island and the west coast of the island. The city of Victoria lies entirely within the Greater Victoria School District. Victoria High School is the only public high school located within the municipal boundaries of the city of Victoria. Opened August 7, 1876, Victoria High School is the oldest High School in North America north of San Francisco and west of Winnipeg, Manitoba. Many of the elementary schools in Victoria offer both French immersion and English programs of instruction. École Victor-Brodeur in Esquimalt serves as a dedicated school for Francophones. Greater Victoria is served by three public post secondary educational institutions, listed by student population size below: In addition, there are several private schools serving the Greater Victoria community, including the Chinese School in Chinatown, the Christ Church Cathedral School, Glenlyon-Norfolk School, Lester B. Pearson United World College of the Pacific, Victoria College of Art, a Justice Institute of British Columbia campus, Pacific Rim College, and the Sprott Shaw College that has had a campus in Victoria since 1913 Victoria is the only Canadian provincial capital without a local CBC Television station, owned-and-operated or affiliate, although it does host a small CBC Radio One (CBCV-FM) station at 780 Kings Road. The region is considered to be a part of the Vancouver television market, receiving most stations that broadcast from across the Strait of Georgia, including CBC Television, Ici Radio-Canada Télé, CTV, Global, Citytv and Omni owned-and-operated stations. The two television stations based in Victoria are CHEK-DT (an independent station) and CIVI-DT (a CTV 2 owned-and-operated station). Victoria has three sister cities: As of March 4, 2022, Victoria City Council voted to suspend the city's relationship with Khabarovsk, Russia as a result of the 2022 Russian invasion of Ukraine.
[[File:SuezCanal-EO.JPG|right|thumb|[[Satellite (artificial)|Satellite]] view of the Suez Canal]] The '''Suez Canal''' ([[Arabic language|Arabic]]: قناة السويس, ''Qanā al-Suways'', [[French language|French]]: ''Le Canal de Suez'') is a [[canal]] in [[Egypt]]. It lies west of the [[Sinai Peninsula]]. The canal is 163 km long (101 miles) and, at its narrowest point, 200 m wide (656 ft). It runs between [[Port Said]] (''Būr Sa'īd'') on the [[Mediterranean Sea]], and [[Suez]] (''al-Suways'') on the [[Red Sea]]. It was built by a [[France|French]] [[company]]. The canal was started in [[1859]] and finished in [[1869]]. The canal allows ships to travel between [[Europe]] and [[Asia]] without having to go the way around [[Africa]]. This saves time and fuel. It was built for Europeans to go to and from the [[Indian Ocean]]. == History == In 1859, Egyptian workers and European engineers, working for [[Ferdinand de Lesseps]] of the Universal Suez Ship Canal Company, began digging under a contract with the Khedive of Ottoman Egypt. It took ten years to build. The first ship passed through the canal 17 November, 1869; [[Giuseppe Verdi]] wrote the famous opera [[Aida]] for this ceremony. The canal made it possible to easily transport goods across the world. The canal also allowed Europeans to travel to [[East Africa]], and soon to take control of it. The British tried to stop the canal, fearing that it would increase French power in the Indian Ocean. Later, they bought shares in the company. The success of the Suez Canal encouraged the French to try to build the [[Panama Canal]]. But they did not finish it. The Panama Canal was finished later. The canal was a central point during the [[Six Day War]] in 1967. A UN peacekeeping force has been stationed in the [[Sinai Peninsula]] since 1974, to avoid more wars. The canal reopened in 1975. About 15,000 ships pass through the canal each year, which is about 14% of world shipping. Each ship takes up to 16 hours to cross the canal. In 2015 a central part of the canal was expanded so more ships can go through and go faster. The canal was [[2021 Suez Canal obstruction|blocked for part of late March 2021]] by the ''[[Ever Given]]'', a [[container ship]], which became stuck diagonally.<ref>{{cite news |last1=Gitlin |first1=Jonathan M. |title=The massive cargo ship that blocked the Suez Canal is now moving again |url=https://arstechnica.com/cars/2021/03/the-massive-cargo-ship-that-blocked-the-suez-canal-is-now-moving-again/ |accessdate=29 March 2021 |work=Ars Technica |date=29 March 2021 |language=en-us}}</ref> == References == {{reflist}} {{authority control}} [[Category:1869 establishments]] [[Category:19th-century establishments in Egypt]] [[Category:Canals]] [[Category:Buildings and structures in Egypt]] [[Category:Transport in Egypt]] [[Category:Mediterranean Sea]]
The Suez Canal (Egyptian Arabic: قَنَاةُ ٱلسُّوَيْسِ, Qanāt es-Suwais) is an artificial sea-level waterway in Egypt, connecting the Mediterranean Sea to the Red Sea through the Isthmus of Suez and dividing Africa and Asia (and by extension, the Sinai Peninsula from the rest of Egypt). The 193.30-kilometre-long (120.11 mi) canal is a key trade route between Europe and Asia. In 1858, Ferdinand de Lesseps formed the Suez Canal Company for the express purpose of building the canal. Construction of the canal lasted from 1859 to 1869. The canal officially opened on 17 November 1869. It offers vessels a direct route between the North Atlantic and northern Indian oceans via the Mediterranean Sea and the Red Sea, avoiding the South Atlantic and southern Indian oceans and reducing the journey distance from the Arabian Sea to London by approximately 8,900 kilometres (5,500 mi), to 10 days at 20 knots (37 km/h; 23 mph) or 8 days at 24 knots (44 km/h; 28 mph). The canal extends from the northern terminus of Port Said to the southern terminus of Port Tewfik at the city of Suez. In 2021, more than 20,600 vessels traversed the canal (an average of 56 per day). The original canal featured a single-lane waterway with passing locations in the Ballah Bypass and the Great Bitter Lake. It contained, according to Alois Negrelli's plans, no locks, with seawater flowing freely through it. In general, the water in the canal north of the Bitter Lakes flows north in winter and south in summer. South of the lakes, the current changes with the tide at Suez. The canal was the property of the Egyptian government, but European shareholders, mostly British and French, owned the concessionary company which operated it until July 1956, when President Gamal Abdel Nasser nationalised it—an event which led to the Suez Crisis of October–November 1956. The canal is operated and maintained by the state-owned Suez Canal Authority (SCA) of Egypt. Under the Convention of Constantinople, it may be used "in time of war as in time of peace, by every vessel of commerce or of war, without distinction of flag." Nevertheless, the canal has played an important military strategic role as a naval short-cut and choke point. Navies with coastlines and bases on both the Mediterranean Sea and the Red Sea (Egypt and Israel) have a particular interest in the Suez Canal. After Egypt closed the Suez Canal at the beginning of the Six-Day War on 5 June 1967, the canal remained closed for precisely eight years, reopening on 5 June 1975. The Egyptian government launched construction in 2014 to expand and widen the Ballah Bypass for 35 km (22 mi) to speed up the canal's transit time. The expansion intended to nearly double the capacity of the Suez Canal, from 49 to 97 ships per day. At a cost of LE 59.4 billion (US$9 billion), this project was funded with interest-bearing investment certificates issued exclusively to Egyptian entities and individuals. The Suez Canal Authority officially opened the new side channel in 2016. This side channel, at the northern side of the east extension of the Suez Canal, serves the East Terminal for berthing and unberthing vessels from the terminal. As the East Container Terminal is located on the Canal itself, before the construction of the new side channel it was not possible to berth or unberth vessels at the terminal while a convoy was running. Ancient west–east canals were built to facilitate travel from the Nile to the Red Sea. One smaller canal is believed to have been constructed under the auspices of Senusret II or Ramesses II. Another canal, probably incorporating a portion of the first, was constructed under the reign of Necho II (610–595 BCE), but the only fully functional canal was engineered and completed by Darius I (522–486 BCE). James Henry Breasted attributes the earliest-known attempt to construct a canal to the first cataract, near Aswan, to the Sixth Dynasty of Egypt and its completion to Senusret III (1878–1839 BCE) of the Twelfth Dynasty of Egypt. The legendary Sesostris (likely either Pharaoh Senusret II or Senusret III of the Twelfth Dynasty of Egypt) may have constructed the ancient canal, the Canal of the Pharaohs, joining the Nile with the Red Sea (1897–1839 BCE), when an irrigation channel was constructed around 1848 BCE that was navigable during the flood season, leading into a dry river valley east of the Nile River Delta named Wadi Tumilat. (It is said that in ancient times the Red Sea reached northward to the Bitter Lakes and Lake Timsah). In his Meteorology, Aristotle (384–322 BCE) wrote: One of their kings tried to make a canal to it (for it would have been of no little advantage to them for the whole region to have become navigable; Sesostris is said to have been the first of the ancient kings to try), but he found that the sea was higher than the land. So he first, and Darius afterwards, stopped making the canal, lest the sea should mix with the river water and spoil it. Strabo wrote that Sesostris started to build a canal, and Pliny the Elder (23/24–79 CE) wrote: 165. Next comes the Tyro tribe and, the harbour of the Daneoi, from which Sesostris, king of Egypt, intended to carry a ship-canal to where the Nile flows into what is known as the Delta; this is a distance of over 60 miles (100 km). Later the Persian king Darius had the same idea, and yet again Ptolemy II, who made a trench 100 feet (30 m) wide, 30 feet (9 m) deep and about 35 miles (55 km) long, as far as the Bitter Lakes. In the 20th century, the northward extension of the later Darius I canal was discovered, extending from Lake Timsah to the Ballah Lakes. This was dated to the Middle Kingdom of Egypt by extrapolating the dates of ancient sites along its course. The reliefs of the Punt expedition under Hatshepsut, 1470 BCE, depict seagoing vessels carrying the expeditionary force returning from Punt. This suggests that a navigable link existed between the Red Sea and the Nile. Recent excavations in Wadi Gawasis may indicate that Egypt's maritime trade started from the Red Sea and did not require a canal. Evidence seems to indicate its existence by the 13th century BCE during the time of Ramesses II. Remnants of an ancient west–east canal through the ancient Egyptian cities of Bubastis, Pi-Ramesses, and Pithom were discovered by Napoleon Bonaparte and his engineers and cartographers in 1799. According to the Histories of the Greek historian Herodotus, about 600 BCE, Necho II undertook to dig a west–east canal through the Wadi Tumilat between Bubastis and Heroopolis, and perhaps continued it to the Heroopolite Gulf and the Red Sea. Regardless, Necho is reported as having never completed his project. Herodotus was told that 120,000 men perished in this undertaking, but this figure is doubtless exaggerated. According to Pliny the Elder, Necho's extension to the canal was about 92 kilometres (57 statute miles), equal to the total distance between Bubastis and the Great Bitter Lake, allowing for winding through valleys. The length that Herodotus tells, of over 1,000 stadia (i.e., over 183 kilometres or 114 miles), must be understood to include the entire distance between the Nile and the Red Sea at that time. With Necho's death, work was discontinued. Herodotus tells that the reason the project was abandoned was because of a warning received from an oracle that others would benefit from its successful completion. Necho's war with Nebuchadnezzar II most probably prevented the canal's continuation. Necho's project was completed by Darius I of Persia, who ruled over Ancient Egypt after it had been conquered by his predecessor Cambyses II. It may be that by Darius's time a natural waterway passage which had existed between the Heroopolite Gulf and the Red Sea in the vicinity of the Egyptian town of Shaluf (alt. Chalouf or Shaloof), located just south of the Great Bitter Lake, had become so blocked with silt that Darius needed to clear it out so as to allow navigation once again. According to Herodotus, Darius's canal was wide enough that two triremes could pass each other with oars extended, and required four days to traverse. Darius commemorated his achievement with a number of granite stelae that he set up on the Nile bank, including one near Kabret, and a further one a few kilometres north of Suez. Darius the Great's Suez Inscriptions read: King Darius says: I am a Persian; setting out from Persia I conquered Egypt. I ordered to dig this canal from the river that is called Nile and flows in Egypt, to the sea that begins in Persia. Therefore, when this canal had been dug as I had ordered, ships went from Egypt through this canal to Persia, as I had intended The canal left the Nile at Bubastis. An inscription on a pillar at Pithom records that in 270 or 269 BCE, it was again reopened, by Ptolemy II Philadelphus. In Arsinoe, Ptolemy constructed a navigable lock, with sluices, at the Heroopolite Gulf of the Red Sea, which allowed the passage of vessels but prevented salt water from the Red Sea from mingling with the fresh water in the canal. In the second half of the 19th century, French cartographers discovered the remnants of an ancient north–south canal past the east side of Lake Timsah and ending near the north end of the Great Bitter Lake. This proved to be the canal made by Darius I, as his stele commemorating its construction was found at the site. (This ancient, second canal may have followed a course along the shoreline of the Red Sea when it once extended north to Lake Timsah.) The Red Sea is believed by some historians to have gradually receded over the centuries, its coastline slowly moving southward away from Lake Timsah and the Great Bitter Lake. Coupled with persistent accumulations of Nile silt, maintenance and repair of Ptolemy's canal became increasingly cumbersome over each passing century. Two hundred years after the construction of Ptolemy's canal, Cleopatra seems to have had no west–east waterway passage, because the Pelusiac branch of the Nile, which fed Ptolemy's west–east canal, had by that time dwindled, being choked with silt. In support of this contention one can note that in 31 BCE, during a reversal of fortune in Mark Antony's and Cleopatra's war against Octavian, she attempted to escape Egypt with her fleet by raising the ships out of the Mediterranean and dragging them across the isthmus of Suez to the Red Sea. Then, according to Plutarch, the Arabs of Petra attacked and burned the first wave of these ships and Cleopatra abandoned the effort. (Modern historians, however, maintain that her ships were burned by the enemy forces of Malichus I.) By the 8th century, a navigable canal existed between Old Cairo and the Red Sea, but accounts vary as to who ordered its construction—either Trajan or 'Amr ibn al-'As, or Umar. This canal was reportedly linked to the River Nile at Old Cairo and ended near modern Suez. A geography treatise De Mensura Orbis Terrae written by the Irish monk Dicuil (born late 8th century) reports a conversation with another monk, Fidelis, who had sailed on the canal from the Nile to the Red Sea during a pilgrimage to the Holy Land in the first half of the 8th century The Abbasid Caliph al-Mansur is said to have ordered this canal closed in 767 to prevent supplies from reaching Arabian detractors. Al-Hakim bi-Amr Allah is claimed to have repaired the Cairo to Red Sea passageway, but only briefly, circa 1000 CE, as it soon "became choked with sand". However, parts of this canal still continued to fill in during the Nile's annual inundations. The successful 1488 navigation of southern Africa by Bartolomeu Dias opened a direct maritime trading route to India and the Spice Islands, and forever changed the balance of Mediterranean trade. One of the most prominent losers in the new order, as former middlemen, was the former spice trading center of Venice. Venetian leaders, driven to desperation, contemplated digging a waterway between the Red Sea and the Nile—anticipating the Suez Canal by almost 400 years—to bring the luxury trade flooding to their doors again. But this remained a dream. Despite entering negotiations with Egypt's ruling Mamelukes, the Venetian plan to build the canal was quickly put to rest by the Ottoman conquest of Egypt in 1517, led by Sultan Selim I. During the 16th century, the Ottoman Grand Vizier Sokollu Mehmed Pasha attempted to construct a canal connecting the Red Sea and the Mediterranean. This was motivated by a desire to connect Constantinople to the pilgrimage and trade routes of the Indian Ocean, as well as by strategic concerns—as the European presence in the Indian Ocean was growing, Ottoman mercantile and strategic interests were increasingly challenged, and the Sublime Porte was increasingly pressed to assert its position. A navigable canal would allow the Ottoman Navy to connect its Red Sea, Black Sea, and Mediterranean fleets. However, this project was deemed too expensive, and was never completed. During the French campaign in Egypt and Syria in late 1798, Napoleon expressed interest in finding the remnants of an ancient waterway passage. This culminated in a cadre of archaeologists, scientists, cartographers and engineers scouring northern Egypt. Their findings, recorded in the Description de l'Égypte, include detailed maps that depict the discovery of an ancient canal extending northward from the Red Sea and then westward toward the Nile. Later, Napoleon, who became the French Emperor in 1804, contemplated the construction of a north–south canal to connect the Mediterranean with the Red Sea. But the plan was abandoned because it incorrectly concluded that the waterway would require locks to operate, the construction of which would be costly and time-consuming. The belief in the need for locks was based on the erroneous belief that the Red Sea was 8.5 m (28 ft) higher than the Mediterranean. This was the result of using fragmentary survey measurements taken in wartime during Napoleon's Egyptian Expedition. As late as 1861, the unnavigable ancient route discovered by Napoleon from Bubastis to the Red Sea still channelled water as far east as Kassassin. Despite the construction challenges that could have been the result of the alleged difference in sea levels, the idea of finding a shorter route to the east remained alive. In 1830, General Francis Chesney submitted a report to the British government that stated that there was no difference in elevation and that the Suez Canal was feasible, but his report received no further attention. Lieutenant Waghorn established his "Overland Route", which transported post and passengers to India via Egypt. Linant de Bellefonds, a French explorer of Egypt, became chief engineer of Egypt's Public Works. In addition to his normal duties, he surveyed the Isthmus of Suez and made plans for the Suez Canal. French Saint-Simonianists showed an interest in the canal and in 1833, Barthélemy Prosper Enfantin tried to draw Muhammad Ali's attention to the canal but was unsuccessful. Alois Negrelli, the Italian-Austrian railroad pioneer, became interested in the idea in 1836. In 1846, Prosper Enfantin's Société d'Études du Canal de Suez invited a number of experts, among them Robert Stephenson, Negrelli and Paul-Adrien Bourdaloue to study the feasibility of the Suez Canal (with the assistance of Linant de Bellefonds). Bourdaloue's survey of the isthmus was the first generally accepted evidence that there was no practical difference in altitude between the two seas. Britain, however, feared that a canal open to everyone might interfere with its India trade and therefore preferred a connection by train from Alexandria via Cairo to Suez, which Stephenson eventually built. In 1854 and 1856, Ferdinand de Lesseps obtained a concession from Sa'id Pasha, the Khedive of Egypt and Sudan, to create a company to construct a canal open to ships of all nations. The company was to operate the canal for 99 years from its opening. De Lesseps had used his friendly relationship with Sa'id, which he had developed while he was a French diplomat in the 1830s. As stipulated in the concessions, de Lesseps convened the International Commission for the piercing of the isthmus of Suez (Commission Internationale pour le percement de l'isthme de Suez) consisting of 13 experts from seven countries, among them John Robinson McClean, later President of the Institution of Civil Engineers in London, and again Negrelli, to examine the plans developed by Linant de Bellefonds, and to advise on the feasibility of and the best route for the canal. After surveys and analyses in Egypt and discussions in Paris on various aspects of the canal, where many of Negrelli's ideas prevailed, the commission produced a unanimous report in December 1856 containing a detailed description of the canal complete with plans and profiles. The Suez Canal Company (Compagnie universelle du canal maritime de Suez) came into being on 15 December 1858. The British government had opposed the project from the outset to its completion. The British, who controlled both the Cape route and the Overland route to India and the Far East, favored the status quo, given that a canal might disrupt their commercial and maritime supremacy. Lord Palmerston, the project's most unwavering foe, confessed in the mid-1850s the real motive behind his opposition: that Britain's commercial and maritime relations would be overthrown by the opening of a new route, open to all nations, and thus deprive his country of its present exclusive advantages. As one of the diplomatic moves against the project when it nevertheless went ahead, it disapproved of the use of "forced labour" for construction of the canal. Involuntary labour on the project ceased, and the viceroy condemned the corvée, halting the project. International opinion was initially skeptical, and shares of the Suez Canal Company did not sell well overseas. Britain, Austria, and Russia did not buy a significant number of shares. With assistance from the Cattaui banking family, and their relationship with James de Rothschild of the French House of Rothschild bonds and shares were successfully promoted in France and other parts of Europe. All French shares were quickly sold in France. A contemporary British skeptic claimed "One thing is sure... our local merchant community doesn't pay practical attention at all to this grand work, and it is legitimate to doubt that the canal's receipts... could ever be sufficient to recover its maintenance fee. It will never become a large ship's accessible way in any case." Work started on the shore of the future Port Said on 25 April 1859. The excavation took some 10 years, with forced labour (corvée) being employed until 1864 to dig out the canal. Some sources estimate that over 30,000 people were working on the canal at any given period, that more than 1.5 million people from various countries were employed, and that tens of thousands of labourers died, many of them from cholera and similar epidemics. Estimates of the number of deaths vary widely with Gamal Abdel Nasser citing 120,000 deaths upon nationalisation of the canal in a 26 July 1956 speech and the company's chief medical officer reporting no higher than 2.49 deaths per thousand in 1866. Doubling these estimates with a generous assumption of 50,000 working staff per year over 11 years would put a conservative estimate at fewer than 3,000 deaths. More closely relying on the limited reported data of the time, the number would be fewer than 1,000. From its inauguration, till 1925, the Suez Canal Company built a series of company towns along the canal to serve its operation. They included ports and their facilities as well as housing for employees segregated by race or nationality. These were Port Said (1869) and Port Fuad (1925) at the canal's northern entrance by the Mediterranenan, Ismailia (1862) near the middle and north of Lake Timsah, and Port Twefik (1867) at the canal's southern entrance on the Red Sea. The canal opened under French control in November 1869. The opening ceremonies began at Port Said on the evening of 15 November, with illuminations, fireworks, and a banquet on the yacht of the Khedive Isma'il Pasha of Egypt and Sudan. The royal guests arrived the following morning: the Emperor Franz Joseph I, the French Empress Eugenie in the Imperial yacht L'Aigle, the Crown Prince of Prussia, and Prince Louis of Hesse. Other international guests included the American natural historian H. W. Harkness. In the afternoon there were blessings of the canal with both Muslim and Christian ceremonies, a temporary mosque and church having been built side by side on the beach. In the evening there were more illuminations and fireworks. On the morning of 17 November, a procession of ships entered the canal, headed by the L'Aigle. Among the ships following was HMS Newport, captained by George Nares, which surveyed the canal on behalf of the Admiralty a few months later. The Newport was involved in an incident that demonstrated some of the problems with the canal. There were suggestions that the depth of parts of the canal at the time of the inauguration were not as great as promised, and that the deepest part of the channel was not always clear, leading to a risk of grounding. The first day of the passage ended at Lake Timsah, 76 kilometres (41 nmi) south of Port Said. The French ship Péluse anchored close to the entrance, then swung around and grounded, the ship and its hawser blocking the way into the lake. The following ships had to anchor in the canal itself until the Péluse was hauled clear the next morning, making it difficult for them to join that night's celebration in Ismailia. Except for the Newport: Nares sent out a boat to carry out soundings, and was able to manoeuver around the Péluse to enter the lake and anchor there for the night. Ismailia was the scene of more celebrations the following day, including a military "march past", illuminations and fireworks, and a ball at the Governor's Palace. The convoy set off again on the morning of 19 November, for the remainder of the trip to Suez. After Suez, many of the participants headed for Cairo, and then to the Pyramids, where a new road had been built for the occasion. An Anchor Line ship, the S.S. Dido, became the first to pass through the Canal from South to North. Although numerous technical, political, and financial problems had been overcome, the final cost was more than double the original estimate. The Khedive, in particular, was able to overcome initial reservations held by both British and French creditors by enlisting the help of the Sursock family, whose deep connections proved invaluable in securing much international support for the project. After the opening, the Suez Canal Company was in financial difficulties. The remaining works were completed only in 1871, and traffic was below expectations in the first two years. De Lesseps therefore tried to increase revenues by interpreting the kind of net ton referred to in the second concession (tonneau de capacité) as meaning a ship's cargo capacity and not only the theoretical net tonnage of the "Moorsom System" introduced in Britain by the Merchant Shipping Act in 1854. The ensuing commercial and diplomatic activities resulted in the International Commission of Constantinople establishing a specific kind of net tonnage and settling the question of tariffs in its protocol of 18 December 1873. This was the origin of the Suez Canal Net Tonnage and the Suez Canal Special Tonnage Certificate, both of which are still in use today. The canal had an immediate and dramatic effect on world trade. Combined with the American transcontinental railroad completed six months earlier, it allowed the world to be circled in record time. It played an important role in increasing European colonization of Africa. The construction of the canal was one of the reasons for the Panic of 1873 in Great Britain, because goods from the Far East had, until then, been carried in sailing vessels around the Cape of Good Hope and stored in British warehouses. An inability to pay his bank debts led Said Pasha's successor, Isma'il Pasha, in 1875 to sell his 44% share in the canal for £4,000,000 ($19.2 million), equivalent to £432 million to £456 million ($540 million to $570 million) in 2019, to the government of the United Kingdom. French shareholders still held the majority. Local unrest caused the British to invade in 1882 and take full control, although nominally Egypt remained part of the Ottoman Empire. The British representative from 1883 to 1907 was Evelyn Baring, 1st Earl of Cromer, who reorganized and modernized the government and suppressed rebellions and corruption, thereby facilitating increased traffic on the canal. The European Mediterranean countries in particular benefited economically from the Suez Canal, as they now had much faster connections to Asia and East Africa than the North and West European maritime trading nations such as Great Britain, the Netherlands or Germany. The biggest beneficiary in the Mediterranean was Austria-Hungary, which had participated in the planning and construction of the canal. The largest Austrian maritime trading company, Österreichischer Lloyd, experienced rapid expansion after the canal was completed, as did the port city of Trieste, then an Austrian possession. The company was a partner in the Compagnie Universelle du Canal de Suez, whose vice-president was the Lloyd co-founder Pasquale Revoltella. In 1900, a dredging trial was held by the Suez Canal Company to determine which ship would assist in the widening and deepening of the canal. One of the ships trialed in the dredging was The Hercules, a ship owned by the Queensland Government in Australia. The Hercules dredged deposits of granite and limestone, but it was determined at the end of the trial that the Hercules would not be used for the dredging of the Suez Canal. The ship was then returned to Brisbane, Australia in January 1901. The Convention of Constantinople in 1888 declared the canal a neutral zone under the protection of the British, who had occupied Egypt and Sudan at the request of Khedive Tewfiq to suppress the Urabi Revolt against his rule. The revolt went on from 1879 to 1882. The British defended the strategically important passage against a major Ottoman attack in 1915, during the First World War. Under the Anglo-Egyptian Treaty of 1936, the UK retained control over the canal. With outbreak of World War II the canal was again strategically important; Italo-German attempts to capture it were repulsed during the North Africa Campaign, which ensured the canal remained closed to Axis shipping. In 1951 Egypt repudiated the 1936 treaty with Great Britain. In October 1954 the UK tentatively agreed to remove its troops from the Canal Zone. Because of Egyptian overtures towards the Soviet Union, both the United Kingdom and the United States withdrew their pledge to financially support construction of the Aswan Dam. Egyptian President Gamal Abdel Nasser responded by nationalising the canal on 26 July 1956 and transferring it to the Suez Canal Authority, intending to finance the dam project using revenue from the canal. On the same day that the canal was nationalised Nasser also closed the Straits of Tiran to all Israeli ships. This led to the Suez Crisis in which the UK, France, and Israel invaded Egypt. According to the pre-agreed war plans under the Protocol of Sèvres, Israel invaded the Sinai Peninsula on 29 October, forcing Egypt to engage them militarily, and allowing the Anglo-French partnership to declare the resultant fighting a threat to stability in the Middle East and enter the war – officially to separate the two forces but in reality to regain the Canal and bring down the Nasser government. To save the British from what he thought was a disastrous action and to stop the war from a possible escalation, Canadian Secretary of State for External Affairs Lester B. Pearson proposed the creation of the first United Nations peacekeeping force to ensure access to the canal for all and an Israeli withdrawal from the Sinai Peninsula. On 4 November 1956, a majority at the United Nations voted for Pearson's peacekeeping resolution, which mandated the UN peacekeepers to stay in Sinai unless both Egypt and Israel agreed to their withdrawal. The United States backed this proposal by putting pressure on the British government through the selling of sterling, which would cause it to depreciate. Britain then called a ceasefire, and later agreed to withdraw its troops by the end of the year. Pearson was later awarded the Nobel Peace Prize. As a result of damage and ships sunk under orders from Nasser the canal was closed until April 1957, when it was cleared with UN assistance. A UN force (UNEF) was established to maintain the free navigability of the canal, and peace in the Sinai Peninsula. Prior to the 1967 war, Egypt had repeatedly closed the canal to Israeli shipping as a defensive measure, refusing to open the canal in 1949 and during the 1956 Suez Crisis. Egypt did so despite UN Security Council resolutions from 1949 and 1951 urging it not to, claiming that hostilities had ended with the 1949 armistice agreement. On 16 May 1967, when Nasser ordered the UNEF peacekeeping forces out of the Sinai Peninsula, including the Suez Canal area, Egyptian troops were sent into Sinai to take their place. Israel protested Nasser's order to close the Straits of Tiran to Israeli trade on 21 May, the same year. This halted Israeli shipping between the port of Eilat and the Red Sea. After the 1967 Six-Day War, Israeli forces occupied the Sinai Peninsula, including the entire east bank of the Suez Canal. In the following years the tensions between Egypt and Israel intensified and from March 1969 until August 1970, a war of attrition took place as the then Egyptian president, Gamal Abdel Nasser, tried to retake the territories occupied by Israel during the conflict. The fighting ceased after the death of Nasser in 1970. After this conflict there were no changes in the distribution of territory, but the underlying tensions persisted. Unwilling to allow the Israelis to use the canal, Egypt immediately imposed a blockade which closed the canal to all shipping immediately after the beginning of the Six-Day War. The canal remained blocked for eight years. There was no anticipation of this event and consequently fifteen cargo ships, known as the "Yellow Fleet", were trapped in the canal, and remained there until its reopening in 1975. On 6 October 1973, during the Yom Kippur War, the canal was the scene of the Operation Badr, in which the Egyptian military crossed the Suez Canal into Israeli-occupied Sinai Peninsula. Much wreckage from this conflict remains visible along the canal's edges. Arab oil exporters, sympathetic to Egypt, pushed OPEC to raise the price of crude oil by around 17 per cent and eventually imposed an embargo against the United States and other Israeli allies. After the Yom Kippur War, the United States initiated Operation Nimbus Moon. The amphibious assault ship USS Inchon (LPH-12) was sent to the Canal, carrying 12 RH-53D minesweeping helicopters of Helicopter Mine Countermeasures Squadron 12. These partly cleared the canal between May and December 1974. It was relieved by the LST USS Barnstable County (LST1197). The British Royal Navy initiated Operation Rheostat and Task Group 65.2 provided for Operation Rheostat One (six months in 1974), the minehunters HMS Maxton, HMS Bossington, and HMS Wilton, the Fleet Clearance Diving Team (FCDT) and HMS Abdiel, a practice minelayer/MCMV support ship; and for Operation Rheostat Two (six months in 1975) the minehunters HMS Hubberston and HMS Sheraton, and HMS Abdiel. When the Canal Clearance Operations were completed, the canal and its lakes were considered 99% clear of mines. The canal was then reopened by Egyptian President Anwar Sadat aboard an Egyptian destroyer, which led the first convoy northbound to Port Said in 1975, at his side stood the Iranian Crown Prince Reza Pahlavi. The UNEF mandate expired in 1979. Despite the efforts of the United States, Israel, Egypt, and others to obtain an extension of the UN role in observing the peace between Israel and Egypt, as called for under the Egypt–Israel peace treaty of 1979, the mandate could not be extended because of the veto by the Soviet Union in the UN Security Council, at the request of Syria. Accordingly, negotiations for a new observer force in the Sinai produced the Multinational Force and Observers (MFO), stationed in Sinai in 1981 in coordination with a phased Israeli withdrawal. The MFO remains active under agreements between the United States, Israel, Egypt, and other nations. In 2014, months after taking office as President of Egypt, Abdel Fattah el-Sisi ordered the expansion of the Ballah Bypass from 61 metres (200 ft) wide to 312 metres (1,024 ft) wide for 35 kilometres (22 mi). The project was called the New Suez Canal, as it allows ships to transit the canal in both directions simultaneously. The project cost more than LE 59.4 billion (US$9 billion) and was completed within a year. Sisi declared the expanded channel open to business in a ceremony on 6 August 2015. On 23 March 2021, at around 05:40 UTC (07:40 local time), the Suez Canal was blocked in both directions by the ultra-large Evergreen G-class container ship Ever Given. The ship, operated by Evergreen Marine, was en route from Malaysia to the Netherlands when it ran aground after strong winds allegedly blew the ship off course. Upon running aground, Ever Given turned sideways, completely blocking the canal. Although part of the length of the canal is paralleled by an older narrower channel which can be used to bypass obstructions, this incident occurred south of that area, in a section of the canal where there is only one channel. The site was located at 30°00′57″N 32°34′45″E / 30.01574°N 32.57918°E / 30.01574; 32.57918. When the incident began, many economists and trade experts commented on the effects of the obstruction if not resolved quickly, citing how important the Suez was to global trade; the incident was likely to drastically affect the global economy because of the trapped goods scheduled to go through the canal. Among those goods, oil shipments were the most affected in the immediate aftermath, due to a significant number still blocked with no other way to reach their destination. Referring to the European and American market, a few maritime experts have disputed the prediction of a drastic effect on trade, saying this "really isn't a substantial transit route for crude" according to Marshall Steeves, energy markets analyst at IHS Markit, and "there are existing stocks" according to Camille Egloff of Boston Consulting Group and alternative sources of supply, noting that traffic only slowed down and that it might only have impacted sectors with existing shortages (such as the semiconductor industry). The International Chamber of Shipping (ICS) estimates that up to $3 billion worth of cargo passes through the Suez Canal every day. It was said the blockage would have an impact on cargo schedules around the world. Shipping companies were also considering whether to divert their ships along the much longer route around the Cape of Good Hope. The first container ship to do so was Ever Given's sister ship, Ever Greet. The ship was re-floated on 29 March. Within a few hours, cargo traffic resumed, slowly resolving the backlog of around 450 ships. The first ship to successfully pass through the canal after the Ever Given's recovery was the YM Wish, a Hong Kong-based cargo ship. On 2 April 2021, Usama Rabie, chairman of the Suez Canal Authority of Egypt, said that the damage caused by the blockage of the canal could reach about $1 billion. Rabie also revealed that after the Suez Canal resumed navigation, as of noon on 31 March 285 cargo ships had passed through the canal smoothly. He said that the remaining 175 freighters waiting to pass through the canal would all pass by 2 April. After the incident, the Egyptian government announced that they would be widening the narrower parts of the canal. On 9 September 2021, the canal was briefly blocked again by the MV Coral Crystal. However, this ship was freed within 15 minutes, presenting minimal disruption to other convoys. On 25 May 2023, another Hong Kong-flagged cargo ship, this time the MV Xin Hai Tong 23 was grounded near the southern end of the canal, but it was refloated by tugboats in less than a day. When built, the canal was 164 km (102 mi) long and 8 m (26 ft) deep. After several enlargements, it is 193.30 km (120+1⁄8 mi) long, 24 m (79 ft) deep and 205 m (673 ft) wide. It consists of the northern access channel of 22 km (14 mi), the canal itself of 162.25 km (100+7⁄8 mi) and the southern access channel of 9 km (5+1⁄2 mi). The so-called New Suez Canal, functional since 6 August 2015, currently has a new parallel canal in the middle part, with its length over 35 kilometres (22 mi). The current parameters of the Suez Canal, including both individual canals of the parallel section are: depth 23 to 24 m (75 to 79 ft) and width at least 205 to 225 m (673 to 738 ft) (that width measured at 11 m (36 ft) of depth). The canal allows passage of ships up to 20 m (66 ft) draft or 240,000 deadweight tons and up to a height of 68 m (223 ft) above water level and a maximum beam of 77.5 m (254 ft) under certain conditions. The canal can handle more traffic and larger ships than the Panama Canal, as Suezmax dimensions are greater than both Panamax and New Panamax. Some supertankers are too large to traverse the canal. Others can offload part of their cargo onto a canal-owned ship to reduce their draft, transit, and reload at the other end of the canal. On 15 April 2021 Egyptian authorities announced that they would widen the southern section of the Suez Canal to improve the efficiency of the canal. The plan mainly covers about 30 kilometres (19 mi) from Suez to the Great Bitter Lake. It will be widened by 40 metres (130 ft)and the maximum depth will be increased from about 20 metres (66 ft) to about 22 metres (72 ft). Ships approaching the canal from the sea are expected to radio the harbour when they are within 15 nautical miles (28 kilometres) of the Fairway Buoy near Port Said. The canal has no locks because of the flat terrain, and the minor sea level difference between each end is inconsequential for shipping. As the canal has no sea surge gates, the ports at the ends would be subject to the sudden impact of tsunamis from the Mediterranean Sea and Red Sea, according to a 2012 article in the Journal of Coastal Research. There is one shipping lane with passing areas in Ballah-Bypass near El Qantara and in the Great Bitter Lake. On a typical day, three convoys transit the canal, two southbound and one northbound. The passage takes between 11 and 16 hours at a speed of around 8 knots (15 km/h; 9 mph). The low speed helps prevent erosion of the banks by ships' wakes. By 1955, about two-thirds of Europe's oil passed through the canal. Around 8% of world sea trade is carried via the canal. In 2008, 21,415 vessels passed through the canal and the receipts totalled $5.381 billion, with an average cost per ship of $251,000. New Rules of Navigation came into force on 1 January 2008, passed by the board of directors of the Suez Canal Authority (SCA) to organise vessels' transit. The most important amendments include allowing vessels with 19-metre (62 ft) draught to pass, increasing the allowed breadth from 32 to 40 metres (105 to 131 ft) (following improvement operations), and imposing a fine on vessels using pilots from outside the SCA inside the canal boundaries without permission. The amendments allow vessels loaded with dangerous cargo (such as radioactive or flammable materials) to pass if they conform with the latest amendments provided by international conventions. The SCA has the right to determine the number of tugs required to assist warships traversing the canal, to achieve the highest degree of safety during transit. Before August 2015, the canal was too narrow for free two-way traffic, so ships had to pass in convoys and use bypasses. The bypasses were 78 km (48 mi) out of 193 km (120 mi) (40%). From north to south, they are Port Said bypass (entrances) 36.5 km (23 mi), Ballah bypass & anchorage 9 km (6 mi), Timsah bypass 5 km (3 mi), and the Deversoir bypass (northern end of the Great Bitter Lake) 27.5 kilometres (17 mi). The bypasses were completed in 1980. Typically, it would take a ship 12 to 16 hours to transit the canal. The canal's 24-hour capacity was about 76 standard ships. In August 2014, Egypt chose a consortium that includes the Egyptian army and global engineering firm Dar Al-Handasah to develop an international industrial and logistics hub in the Suez Canal area, and began the construction of a new canal section from 60 to 95 km (37 to 59 mi) combined with expansion and deep digging of the other 37 km (23 mi) of the canal. This will allow navigation in both directions simultaneously in the 72-kilometre-long (45 mi) central section of the canal. These extensions were formally opened on 6 August 2015 by President Al-Sisi. Since the canal does not cater to unregulated two-way traffic, all ships transit in convoys on regular times, scheduled on a 24-hour basis. Each day, a single northbound convoy starts at 04:00 from Suez. At dual lane sections, the convoy uses the eastern route. Synchronised with this convoy's passage is the southbound convoy. It starts at 03:30 from Port Said and so passes the Northbound convoy in the two-lane section. From north to south, the crossings are: A railway on the west bank runs parallel to the canal for its entire length. The five pontoon bridges were opened between 2016 and 2019. They are designed to be movable, and can be completely rotated against the banks of the canal to allow shipping through, or else individual sections can be moved to create a narrower channel. Six new tunnels for cars and trains are also planned across the canal. Currently the Ahmed Hamdi is the only tunnel connecting Suez to the Sinai. Economically, after its completion, the Suez Canal benefited primarily the sea trading powers of the Mediterranean countries, which now had much faster connections to the Near and Far East than the North and West European sea trading nations such as Great Britain or Germany. The main Habsburg trading port of Trieste with its direct connections to Central Europe experienced a meteoric rise at that time. The time saved in the 19th century for an assumed steamship trip to Bombay from Brindisi and Trieste was 37 days, from Genoa 32, from Marseille 31, from Bordeaux, Liverpool, London, Amsterdam and Hamburg 24 days. At that time, it was also necessary to consider whether the goods to be transported could bear the costly canal tariff. This led to a rapid growth of Mediterranean ports with their land routes to Central and Eastern Europe. According to today's information from the shipping companies, the route from Singapore to Rotterdam through the Suez Canal will be shortened by 6,000 kilometres (3,700 mi) and thus by nine days compared to the route around Africa. As a result, liner services between Asia and Europe save 44 per cent CO2 (carbon dioxide) thanks to this shorter route. The Suez Canal has a correspondingly important role in the connection between East Africa and the Mediterranean region. In the 20th century, trade through the Suez Canal came to a standstill several times, due to the two world wars and the Suez Canal crisis. Many trade flows were also shifted away from the Mediterranean ports towards Northern European terminals, such as Hamburg and Rotterdam. Only after the end of the Cold War, the growth in European economic integration, the consideration of CO2 emission and the Chinese Silk Road Initiative, are Mediterranean ports such as Piraeus and Trieste again at the focus of growth and investment. The Suez Canal set a new record with annual revenue of $9.4 billion in USD for the fiscal year that ended 30 June 2023. Before the canal's opening in 1869, goods were sometimes offloaded from ships and carried overland between the Mediterranean and the Red Sea. The main alternative is around Cape Agulhas, the southernmost point of Africa, commonly referred to as the Cape of Good Hope route. This was the only sea route before the canal was constructed, and when the canal was closed. It is still the only route for ships that are too large for the canal. In the early 21st century, the Suez Canal has suffered from diminished traffic due to piracy in Somalia, with many shipping companies choosing to take the long route instead. Between 2008 and 2010, it is estimated that the canal lost 10% of traffic due to the threat of piracy, and another 10% due to the financial crisis. An oil tanker going from Saudi Arabia to the United States has 4,345 km (2,700 mi) farther to go when taking the route south of Africa rather than the canal. In recent years, the shrinking Arctic sea ice has made the Northern Sea Route feasible for commercial cargo ships between Europe and East Asia during a six-to-eight-week window in the summer months, shortening the voyage by thousands of kilometres compared to that through the Suez Canal. According to polar climate researchers, as the extent of the Arctic summer ice pack recedes the route will become passable without the help of icebreakers for a greater period each summer. The Bremen-based Beluga Group claimed in 2009 to be the first Western company to attempt using the Northern Sea Route without assistance from icebreakers, cutting 6,400 kilometres (4,000 mi) off the journey between Ulsan, Korea and Rotterdam, the Netherlands. Sailing ships, such as the windjammers in the heyday of the Great Grain Race between Australia and Europe during the 1930s, often preferred the Cape Horn route when going to Europe, due to prevalent wind directions, even though it is slightly longer from Sydney to Europe this way than past Cape Agulhas. In February 2012, Israel announced its intention to construct a railway between the Mediterranean and Eilat through the Negev desert to compete with the canal. By 2019, the project had been put on indefinite hold. The opening of the canal created the first salt-water passage between the Mediterranean Sea and the Red Sea. Although the Red Sea is about 1.2 m (4 ft) higher than the eastern Mediterranean, the current between the Mediterranean and the middle of the canal at the Bitter Lakes flows north in winter and south in summer. The current south of the Bitter Lakes is tidal, varying with the tide at Suez. The Bitter Lakes, which were hypersaline natural lakes, blocked the migration of Red Sea species into the Mediterranean for many decades, but as the salinity of the lakes gradually equalised with that of the Red Sea the barrier to migration was removed, and plants and animals from the Red Sea have begun to colonise the eastern Mediterranean. The Red Sea is generally saltier and less nutrient-rich than the Mediterranean, so that Erythrean species will often do well in the 'milder' eastern Mediterranean environment. To the contrary very few Mediterranean species have been able to settle in the 'harsher' conditions of the Red Sea. The dominant, south to north, migratory passage across the canal is often called Lessepsian migration (after Ferdinand de Lesseps) or "Erythrean invasion". The recent construction by the Egyptian government of a major canal extension – allowing for two-way traffic in the central section of the canal and finally implemented in 2015 – raised concerns from marine biologists, who fear that it will enhance the arrival of Red Sea species in the Mediterranean. Exotic species from the Indo-Pacific Ocean and introduced into the Mediterranean via the canal since the 1880s have become a significant component of the Mediterranean ecosystem. They already impact its ecology, endangering some local and endemic species. Since the piercing of the canal, over a thousand species from the Red Sea—plankton, seaweeds, invertebrates, fishes—have been recorded in the Mediterranean, and many others will clearly follow. The resulting change in biodiversity is without precedent in human memory and is accelerating: a long-term cross-Basin survey engaged by the Mediterranean Science Commission recently documented that in the first twenty years of this century more exotic fish species from the Indian Ocean had reached the Mediterranean than during the entire 20th century. Historically, the construction of the canal was preceded by cutting a small fresh-water canal called Sweet Water Canal from the Nile delta along Wadi Tumilat to the future canal, with a southern branch to Suez and a northern branch to Port Said. Completed in 1863, these brought fresh water to a previously arid area, initially for canal construction, and subsequently facilitating growth of agriculture and settlements along the canal. However the Aswan High Dam construction across the Nile, which started operating in 1968, much reduced the inflow of freshwater and cut all natural nutrient-rich silt entering the eastern Mediterranean at the Nile Delta. The Suez Canal Economic Zone, sometimes shortened to SCZONE, describes the set of locations neighbouring the canal where customs rates have been reduced to zero in order to attract investment. The zone comprises over 461 km (178 sq mi) within the governorates of Port Said, Ismailia and Suez. Projects in the zone are collectively described as the Suez Canal Area Development Project (SCADP). The plan focuses on development of East Port Said and the port of Ain Sokhna, and hopes to extend to four more ports at West Port Said, El-Adabiya, Arish and El Tor. The zone incorporates the four "Qualifying Industrial Zones" at Port Said, Ismailia and Suez, a 1996 American initiative to encourage economic ties between Israel and its neighbours.
[[File:USS Bainbridge (CGN-25) underway in the Suez Canal on 27 February 1992.jpg|thumb|An [[United States|American]] [[warship]] in the [[Suez Canal]].]][[File:Bateau traversant le canal de corinthe.jpg|thumb|Some canals require blasting via the rock.]]A '''canal''' is a [[waterway]] made by [[Human|humans]]. For many [[Century|centuries]], canals have been built as a way of transporting heavy goods in barges or [[Boat|boats]]. Canals often connect [[lake]]s, [[River|rivers]], or [[Ocean|oceans]]. The [[Panama Canal]] is a famous canal that connects the [[Atlantic Ocean]] with the [[Pacific Ocean]]. Many canals are reinforced with [[clay]] or [[concrete]] on the sides and have [[Lock (water transport)|locks]]. Other canals are used for [[irrigation]] or [[hydropower]]. == Related pages == * [[mwod:canal|Merriam webster dictionary]] ==References== {{Commons category|Canals}} {{-}} {{Transport-stub}} [[Category:Canals| ]]
Canals or artificial waterways are waterways or engineered channels built for drainage management (e.g. flood control and irrigation) or for conveyancing water transport vehicles (e.g. water taxi). They carry free, calm surface flow under atmospheric pressure, and can be thought of as artificial rivers. In most cases, a canal has a series of dams and locks that create reservoirs of low speed current flow. These reservoirs are referred to as slack water levels, often just called levels. A canal can be called a navigation canal when it parallels a natural river and shares part of the latter's discharges and drainage basin, and leverages its resources by building dams and locks to increase and lengthen its stretches of slack water levels while staying in its valley. A canal can cut across a drainage divide atop a ridge, generally requiring an external water source above the highest elevation. The best-known example of such a canal is the Panama Canal. Many canals have been built at elevations, above valleys and other waterways. Canals with sources of water at a higher level can deliver water to a destination such as a city where water is needed. The Roman Empire's aqueducts were such water supply canals. The term was once used to describe linear features seen on the surface of Mars, Martian canals, an optical illusion. A navigation is a series of channels that run roughly parallel to the valley and stream bed of an unimproved river. A navigation always shares the drainage basin of the river. A vessel uses the calm parts of the river itself as well as improvements, traversing the same changes in height. A true canal is a channel that cuts across a drainage divide, making a navigable channel connecting two different drainage basins. Both navigations and canals use engineered structures to improve navigation: Since they cut across drainage divides, canals are more difficult to construct and often need additional improvements, like viaducts and aqueducts to bridge waters over streams and roads, and ways to keep water in the channel. There are two broad types of canal: Historically, canals were of immense importance to commerce and the development, growth and vitality of a civilization. In 1855 the Lehigh Canal carried over 1.2 million tons of anthracite coal; by the 1930s the company which built and operated it over a century pulled the plug. The few canals still in operation in our modern age are a fraction of the numbers that once fueled and enabled economic growth, indeed were practically a prerequisite to further urbanization and industrialization. For the movement of bulk raw materials such as coal and ores are difficult and marginally affordable without water transport. Such raw materials fueled the industrial developments and new metallurgy resulting of the spiral of increasing mechanization during 17th–20th century, leading to new research disciplines, new industries and economies of scale, raising the standard of living for any industrialized society. Most ship canals today primarily service bulk cargo and large ship transportation industries, whereas the once critical smaller inland waterways conceived and engineered as boat and barge canals have largely been supplanted and filled in, abandoned and left to deteriorate, or kept in service and staffed by state employees, where dams and locks are maintained for flood control or pleasure boating. Their replacement was gradual, beginning first in the United States in the mid-1850s where canal shipping was first augmented by, then began being replaced by using much faster, less geographically constrained & limited, and generally cheaper to maintain railways. By the early 1880s, canals which had little ability to economically compete with rail transport, were off the map. In the next couple of decades, coal was increasingly diminished as the heating fuel of choice by oil, and growth of coal shipments leveled off. Later, after World War I when motor-trucks came into their own, the last small U.S. barge canals saw a steady decline in cargo ton-miles alongside many railways, the flexibility and steep slope climbing capability of lorries taking over cargo hauling increasingly as road networks were improved, and which also had the freedom to make deliveries well away from rail lined road beds or ditches in the dirt which could not operate in the winter. The longest extant canal today, the Grand Canal in northern China, still remains in heavy use, especially the portion south of the Yellow River. It stretches from Beijing to Hangzhou at 1,794 kilometres (1,115 miles). Canals are built in one of three ways, or a combination of the three, depending on available water and available path: Smaller transportation canals can carry barges or narrowboats, while ship canals allow seagoing ships to travel to an inland port (e.g., Manchester Ship Canal), or from one sea or ocean to another (e.g., Caledonian Canal, Panama Canal). At their simplest, canals consist of a trench filled with water. Depending on the stratum the canal passes through, it may be necessary to line the cut with some form of watertight material such as clay or concrete. When this is done with clay, it is known as puddling. Canals need to be level, and while small irregularities in the lie of the land can be dealt with through cuttings and embankments, for larger deviations other approaches have been adopted. The most common is the pound lock, which consists of a chamber within which the water level can be raised or lowered connecting either two pieces of canal at a different level or the canal with a river or the sea. When there is a hill to be climbed, flights of many locks in short succession may be used. Prior to the development of the pound lock in 984 AD in China by Chhaio Wei-Yo and later in Europe in the 15th century, either flash locks consisting of a single gate were used or ramps, sometimes equipped with rollers, were used to change the level. Flash locks were only practical where there was plenty of water available. Locks use a lot of water, so builders have adopted other approaches for situations where little water is available. These include boat lifts, such as the Falkirk Wheel, which use a caisson of water in which boats float while being moved between two levels; and inclined planes where a caisson is hauled up a steep railway. To cross a stream, road or valley (where the delay caused by a flight of locks at either side would be unacceptable) the valley can be spanned by a navigable aqueduct – a famous example in Wales is the Pontcysyllte Aqueduct (now a UNESCO World Heritage Site) across the valley of the River Dee. Another option for dealing with hills is to tunnel through them. An example of this approach is the Harecastle Tunnel on the Trent and Mersey Canal. Tunnels are only practical for smaller canals. Some canals attempted to keep changes in level down to a minimum. These canals known as contour canals would take longer, winding routes, along which the land was a uniform altitude. Other, generally later, canals took more direct routes requiring the use of various methods to deal with the change in level. Canals have various features to tackle the problem of water supply. In cases, like the Suez Canal, the canal is open to the sea. Where the canal is not at sea level, a number of approaches have been adopted. Taking water from existing rivers or springs was an option in some cases, sometimes supplemented by other methods to deal with seasonal variations in flow. Where such sources were unavailable, reservoirs – either separate from the canal or built into its course – and back pumping were used to provide the required water. In other cases, water pumped from mines was used to feed the canal. In certain cases, extensive "feeder canals" were built to bring water from sources located far from the canal. Where large amounts of goods are loaded or unloaded such as at the end of a canal, a canal basin may be built. This would normally be a section of water wider than the general canal. In some cases, the canal basins contain wharfs and cranes to assist with movement of goods. When a section of the canal needs to be sealed off so it can be drained for maintenance stop planks are frequently used. These consist of planks of wood placed across the canal to form a dam. They are generally placed in pre-existing grooves in the canal bank. On more modern canals, "guard locks" or gates were sometimes placed to allow a section of the canal to be quickly closed off, either for maintenance, or to prevent a major loss of water due to a canal breach. A canal fall, or canal drop, is a vertical drop in the canal bed. These are built when the natural ground slope is steeper than the desired canal gradient. They are constructed so the falling water's kinetic energy is dissipated in order to prevent it from scouring the bed and sides of the canal. A canal fall is constructed by cut and fill. It may be combined with a regulator, bridge, or other structure to save costs. There are various types of canal falls, based on their shape. One type is the ogee fall, where the drop follows an s-shaped curve to create a smooth transition and reduce turbulence. However, this smooth transition does not dissipate the water's kinetic energy, which leads to heavy scouring. As a result, the canal needs to be reinforced with concrete or masonry to protect it from eroding. Another type of canal fall is the vertical fall, which is "simple and economical". These feature a "cistern", or depressed area just downstream from the fall, to "cushion" the water by providing a deep pool for its kinetic energy to be diffused in. Vertical falls work for drops of up to 1.5 m in height, and for discharge of up to 15 cubic meters per second. The transport capacity of pack animals and carts is limited. A mule can carry an eighth-ton [250 pounds (113 kg)] maximum load over a journey measured in days and weeks, though much more for shorter distances and periods with appropriate rest. Besides, carts need roads. Transport over water is much more efficient and cost-effective for large cargoes. The oldest known canals were irrigation canals, built in Mesopotamia circa 4000 BC, in what is now Iraq. The Indus Valley civilization of ancient India (circa 3000 BC) had sophisticated irrigation and storage systems developed, including the reservoirs built at Girnar in 3000 BC. This is the first time that such planned civil project had taken place in the ancient world. In Egypt, canals date back at least to the time of Pepi I Meryre (reigned 2332–2283 BC), who ordered a canal built to bypass the cataract on the Nile near Aswan. In ancient China, large canals for river transport were established as far back as the Spring and Autumn Period (8th–5th centuries BC), the longest one of that period being the Hong Gou (Canal of the Wild Geese), which according to the ancient historian Sima Qian connected the old states of Song, Zhang, Chen, Cai, Cao, and Wei. The Caoyun System of canals was essential for imperial taxation, which was largely assessed in kind and involved enormous shipments of rice and other grains. By far the longest canal was the Grand Canal of China, still the longest canal in the world today and the oldest extant one. It is 1,794 kilometres (1,115 mi) long and was built to carry the Emperor Yang Guang between Zhuodu (Beijing) and Yuhang (Hangzhou). The project began in 605 and was completed in 609, although much of the work combined older canals, the oldest section of the canal existing since at least 486 BC. Even in its narrowest urban sections it is rarely less than 30 metres (98 ft) wide. In the 5th century BC, Achaemenid king Xerxes I of Persia ordered the construction of the Xerxes Canal through the base of Mount Athos peninsula, Chalkidiki, northern Greece. It was constructed as part of his preparations for the Second Persian invasion of Greece, a part of the Greco-Persian Wars. It is one of the few monuments left by the Persian Empire in Europe. Greek engineers were also among the first to use canal locks, by which they regulated the water flow in the Ancient Suez Canal as early as the 3rd century BC. There was little experience moving bulk loads by carts, while a pack-horse would [i.e. 'could'] carry only an eighth of a ton. On a soft road a horse might be able to draw 5/8ths of a ton. But if the load were carried by a barge on a waterway, then up to 30 tons could be drawn by the same horse.— technology historian Ronald W. Clark referring to transport realities before the industrial revolution and the Canal age. Hohokam was a society in the North American Southwest in what is now part of Arizona, United States, and Sonora, Mexico. Their irrigation systems supported the largest population in the Southwest by 1300 CE. Archaeologists working at a major archaeological dig in the 1990s in the Tucson Basin, along the Santa Cruz River, identified a culture and people that may have been the ancestors of the Hohokam. This prehistoric group occupied southern Arizona as early as 2000 BCE, and in the Early Agricultural Period grew corn, lived year-round in sedentary villages, and developed sophisticated irrigation canals. The large-scale Hohokam irrigation network in the Phoenix metropolitan area was the most complex in ancient North America. A portion of the ancient canals has been renovated for the Salt River Project and now helps to supply the city's water. In the Middle Ages, water transport was several times cheaper and faster than transport overland. Overland transport by animal drawn conveyances was used around settled areas, but unimproved roads required pack animal trains, usually of mules to carry any degree of mass, and while a mule could carry an eighth ton, it also needed teamsters to tend it and one man could only tend perhaps five mules, meaning overland bulk transport was also expensive, as men expect compensation in the form of wages, room and board. This was because long-haul roads were unpaved, more often than not too narrow for carts, much less wagons, and in poor condition, wending their way through forests, marshy or muddy quagmires as often as unimproved but dry footing. In that era, as today, greater cargoes, especially bulk goods and raw materials, could be transported by ship far more economically than by land; in the pre-railroad days of the industrial revolution, water transport was the gold standard of fast transportation. The first artificial canal in Western Europe was the Fossa Carolina built at the end of the 8th century under personal supervision of Charlemagne. In Britain, the Glastonbury Canal is believed to be the first post-Roman canal and was built in the middle of the 10th century to link the River Brue at Northover with Glastonbury Abbey, a distance of about 1.75 kilometres (1,900 yd). Its initial purpose is believed to be the transport of building stone for the abbey, but later it was used for delivering produce, including grain, wine and fish, from the abbey's outlying properties. It remained in use until at least the 14th century, but possibly as late as the mid-16th century.More lasting and of more economic impact were canals like the Naviglio Grande built between 1127 and 1257 to connect Milan with the river Ticino. The Naviglio Grande is the most important of the lombard "navigli" and the oldest functioning canal in Europe.Later, canals were built in the Netherlands and Flanders to drain the polders and assist transportation of goods and people. Canal building was revived in this age because of commercial expansion from the 12th century. River navigations were improved progressively by the use of single, or flash locks. Taking boats through these used large amounts of water leading to conflicts with watermill owners and to correct this, the pound or chamber lock first appeared, in the 10th century in China and in Europe in 1373 in Vreeswijk, Netherlands. Another important development was the mitre gate, which was, it is presumed, introduced in Italy by Bertola da Novate in the 16th century. This allowed wider gates and also removed the height restriction of guillotine locks. To break out of the limitations caused by river valleys, the first summit level canals were developed with the Grand Canal of China in 581–617 AD whilst in Europe the first, also using single locks, was the Stecknitz Canal in Germany in 1398. In the Songhai Empire of West Africa, several canals were constructed under Sunni Ali and Askia Muhammad I between Kabara and Timbuktu in the 15th century. These were used primarily for irrigation and transport. Sunni Ali also attempted to construct a canal from the Niger River to Walata to facilitate conquest of the city but his progress was halted when he went to war with the Mossi Kingdoms. Around 1500–1800 the first summit level canal to use pound locks in Europe was the Briare Canal connecting the Loire and Seine (1642), followed by the more ambitious Canal du Midi (1683) connecting the Atlantic to the Mediterranean. This included a staircase of 8 locks at Béziers, a 157 metres (515 ft) tunnel, and three major aqueducts. Canal building progressed steadily in Germany in the 17th and 18th centuries with three great rivers, the Elbe, Oder and Weser being linked by canals. In post-Roman Britain, the first early modern period canal built appears to have been the Exeter Canal, which was surveyed in 1563, and open in 1566. The oldest canal in the European settlements of North America, technically a mill race built for industrial purposes, is Mother Brook between the Boston, Massachusetts neighbourhoods of Dedham and Hyde Park connecting the higher waters of the Charles River and the mouth of the Neponset River and the sea. It was constructed in 1639 to provide water power for mills. In Russia, the Volga–Baltic Waterway, a nationwide canal system connecting the Baltic Sea and Caspian Sea via the Neva and Volga rivers, was opened in 1718. The modern canal system was mainly a product of the 18th century and early 19th century. It came into being because the Industrial Revolution (which began in Britain during the mid-18th century) demanded an economic and reliable way to transport goods and commodities in large quantities. By the early 18th century, river navigations such as the Aire and Calder Navigation were becoming quite sophisticated, with pound locks and longer and longer "cuts" (some with intermediate locks) to avoid circuitous or difficult stretches of river. Eventually, the experience of building long multi-level cuts with their own locks gave rise to the idea of building a "pure" canal, a waterway designed on the basis of where goods needed to go, not where a river happened to be. The claim for the first pure canal in Great Britain is debated between "Sankey" and "Bridgewater" supporters. The first true canal in what is now the United Kingdom was the Newry Canal in Northern Ireland constructed by Thomas Steers in 1741. The Sankey Brook Navigation, which connected St Helens with the River Mersey, is often claimed as the first modern "purely artificial" canal because although originally a scheme to make the Sankey Brook navigable, it included an entirely new artificial channel that was effectively a canal along the Sankey Brook valley. However, "Bridgewater" supporters point out that the last quarter-mile of the navigation is indeed a canalized stretch of the Brook, and that it was the Bridgewater Canal (less obviously associated with an existing river) that captured the popular imagination and inspired further canals. In the mid-eighteenth century the 3rd Duke of Bridgewater, who owned a number of coal mines in northern England, wanted a reliable way to transport his coal to the rapidly industrializing city of Manchester. He commissioned the engineer James Brindley to build a canal for that purpose. Brindley's design included an aqueduct carrying the canal over the River Irwell. This was an engineering wonder which immediately attracted tourists. The construction of this canal was funded entirely by the Duke and was called the Bridgewater Canal. It opened in 1761 and was the first major British canal. The new canals proved highly successful. The boats on the canal were horse-drawn with a towpath alongside the canal for the horse to walk along. This horse-drawn system proved to be highly economical and became standard across the British canal network. Commercial horse-drawn canal boats could be seen on the UK's canals until as late as the 1950s, although by then diesel-powered boats, often towing a second unpowered boat, had become standard. The canal boats could carry thirty tons at a time with only one horse pulling – more than ten times the amount of cargo per horse that was possible with a cart. Because of this huge increase in supply, the Bridgewater canal reduced the price of coal in Manchester by nearly two-thirds within just a year of its opening. The Bridgewater was also a huge financial success, with it earning what had been spent on its construction within just a few years. This success proved the viability of canal transport, and soon industrialists in many other parts of the country wanted canals. After the Bridgewater canal, early canals were built by groups of private individuals with an interest in improving communications. In Staffordshire the famous potter Josiah Wedgwood saw an opportunity to bring bulky cargoes of clay to his factory doors and to transport his fragile finished goods to market in Manchester, Birmingham or further away, by water, minimizing breakages. Within just a few years of the Bridgewater's opening, an embryonic national canal network came into being, with the construction of canals such as the Oxford Canal and the Trent & Mersey Canal. The new canal system was both cause and effect of the rapid industrialization of The Midlands and the north. The period between the 1770s and the 1830s is often referred to as the "Golden Age" of British canals. For each canal, an Act of Parliament was necessary to authorize construction, and as people saw the high incomes achieved from canal tolls, canal proposals came to be put forward by investors interested in profiting from dividends, at least as much as by people whose businesses would profit from cheaper transport of raw materials and finished goods. In a further development, there was often out-and-out speculation, where people would try to buy shares in a newly floated company to sell them on for an immediate profit, regardless of whether the canal was ever profitable, or even built. During this period of "canal mania", huge sums were invested in canal building, and although many schemes came to nothing, the canal system rapidly expanded to nearly 4,000 miles (over 6,400 kilometres) in length. Many rival canal companies were formed and competition was rampant. Perhaps the best example was Worcester Bar in Birmingham, a point where the Worcester and Birmingham Canal and the Birmingham Canal Navigations Main Line were only seven feet apart. For many years, a dispute about tolls meant that goods travelling through Birmingham had to be portaged from boats in one canal to boats in the other. Canal companies were initially chartered by individual states in the United States. These early canals were constructed, owned, and operated by private joint-stock companies. Four were completed when the War of 1812 broke out; these were the South Hadley Canal (opened 1795) in Massachusetts, Santee Canal (opened 1800) in South Carolina, the Middlesex Canal (opened 1802) also in Massachusetts, and the Dismal Swamp Canal (opened 1805) in Virginia. The Erie Canal (opened 1825) was chartered and owned by the state of New York and financed by bonds bought by private investors. The Erie canal runs about 363 miles (584 km) from Albany, New York, on the Hudson River to Buffalo, New York, at Lake Erie. The Hudson River connects Albany to the Atlantic port of New York City and the Erie Canal completed a navigable water route from the Atlantic Ocean to the Great Lakes. The canal contains 36 locks and encompasses a total elevation differential of around 565 ft. (169 m). The Erie Canal with its easy connections to most of the U.S. mid-west and New York City soon quickly paid back all its invested capital (US$7 million) and started turning a profit. By cutting transportation costs in half or more it became a large profit center for Albany and New York City as it allowed the cheap transportation of many of the agricultural products grown in the mid west of the United States to the rest of the world. From New York City these agricultural products could easily be shipped to other U.S. states or overseas. Assured of a market for their farm products the settlement of the U.S. mid-west was greatly accelerated by the Erie Canal. The profits generated by the Erie Canal project started a canal building boom in the United States that lasted until about 1850 when railroads started becoming seriously competitive in price and convenience. The Blackstone Canal (finished in 1828) in Massachusetts and Rhode Island fulfilled a similar role in the early industrial revolution between 1828 and 1848. The Blackstone Valley was a major contributor of the American Industrial Revolution where Samuel Slater built his first textile mill. A power canal refers to a canal used for hydraulic power generation, rather than for transport. Nowadays power canals are built almost exclusively as parts of hydroelectric power stations. Parts of the United States, particularly in the Northeast, had enough fast-flowing rivers that water power was the primary means of powering factories (usually textile mills) until after the American Civil War. For example, Lowell, Massachusetts, considered to be "The Cradle of the American Industrial Revolution," has 6 miles (9.7 km) of canals, built from around 1790 to 1850, that provided water power and a means of transportation for the city. The output of the system is estimated at 10,000 horsepower. Other cities with extensive power canal systems include Lawrence, Massachusetts, Holyoke, Massachusetts, Manchester, New Hampshire, and Augusta, Georgia. The most notable power canal was built in 1862 for the Niagara Falls Hydraulic Power and Manufacturing Company. Competition, from railways from the 1830s and roads in the 20th century, made the smaller canals obsolete for most commercial transport, and many of the British canals fell into decay. Only the Manchester Ship Canal and the Aire and Calder Canal bucked this trend. Yet in other countries canals grew in size as construction techniques improved. During the 19th century in the US, the length of canals grew from 100 miles (161 km) to over 4,000, with a complex network making the Great Lakes navigable, in conjunction with Canada, although some canals were later drained and used as railroad rights-of-way. In the United States, navigable canals reached into isolated areas and brought them in touch with the world beyond. By 1825 the Erie Canal, 363 miles (584 km) long with 36 locks, opened up a connection from the populated Northeast to the Great Lakes. Settlers flooded into regions serviced by such canals, since access to markets was available. The Erie Canal (as well as other canals) was instrumental in lowering the differences in commodity prices between these various markets across America. The canals caused price convergence between different regions because of their reduction in transportation costs, which allowed Americans to ship and buy goods from farther distances much cheaper. Ohio built many miles of canal, Indiana had working canals for a few decades, and the Illinois and Michigan Canal connected the Great Lakes to the Mississippi River system until replaced by a channelized river waterway. Three major canals with very different purposes were built in what is now Canada. The first Welland Canal, which opened in 1829 between Lake Ontario and Lake Erie, bypassing Niagara Falls and the Lachine Canal (1825), which allowed ships to skirt the nearly impassable rapids on the St. Lawrence River at Montreal, were built for commerce. The Rideau Canal, completed in 1832, connects Ottawa on the Ottawa River to Kingston, Ontario on Lake Ontario. The Rideau Canal was built as a result of the War of 1812 to provide military transportation between the British colonies of Upper Canada and Lower Canada as an alternative to part of the St. Lawrence River, which was susceptible to blockade by the United States. In France, a steady linking of all the river systems – Rhine, Rhône, Saône and Seine – and the North Sea was boosted in 1879 by the establishment of the Freycinet gauge, which specified the minimum size of locks. Canal traffic doubled in the first decades of the 20th century. Many notable sea canals were completed in this period, starting with the Suez Canal (1869) – which carries tonnage many times that of most other canals – and the Kiel Canal (1897), though the Panama Canal was not opened until 1914. In the 19th century, a number of canals were built in Japan including the Biwako canal and the Tone canal. These canals were partially built with the help of engineers from the Netherlands and other countries. A major question was how to connect the Atlantic and the Pacific with a canal through narrow Central America. (The Panama Railroad opened in 1855.) The original proposal was for a sea-level canal through what is today Nicaragua, taking advantage of the relatively large Lake Nicaragua. This canal has never been built in part because of political instability, which scared off potential investors. It remains an active project (the geography has not changed), and in the 2010s Chinese involvement was developing. The second choice for a Central American canal was a Panama canal. The De Lessups company, which ran the Suez Canal, first attempted to build a Panama Canal in the 1880s. The difficulty of the terrain and weather (rain) encountered caused the company to go bankrupt. High worker mortality from disease also discouraged further investment in the project. DeLessup's abandoned excavating equipment sits, isolated decaying machines, today tourist attractions. Twenty years later, an expansionist United States, that just acquired colonies after defeating Spain in the 1898 Spanish–American War, and whose Navy became more important, decided to reactivate the project. The United States and Colombia did not reach agreement on the terms of a canal treaty (see Hay–Herrán Treaty). Panama, which did not have (and still does not have) a land connection with the rest of Colombia, was already thinking of independence. In 1903 the United States, with support from Panamanians who expected the canal to provide substantial wages, revenues, and markets for local goods and services, took Panama province away from Colombia, and set up a puppet republic (Panama). Its currency, the Balboa – a name that suggests the country began as a way to get from one hemisphere to the other – was a replica of the US dollar. The US dollar was and remains legal tender (used as currency). A U.S. military zone, the Canal Zone, 10 miles (16 km) wide, with U.S. military stationed there (bases, 2 TV stations, channels 8 and 10, Pxs, a U.S.-style high school), split Panama in half. The Canal – a major engineering project – was built. The U.S. did not feel that conditions were stable enough to withdraw until 1979. The withdrawal from Panama contributed to President Jimmy Carter's defeat in 1980. Large-scale ship canals such as the Panama Canal and Suez Canal continue to operate for cargo transportation, as do European barge canals. Due to globalization, they are becoming increasingly important, resulting in expansion projects such as the Panama Canal expansion project. The expanded canal began commercial operation on 26 June 2016. The new set of locks allow transit of larger, Post-Panamax and New Panamax ships. The narrow early industrial canals, however, have ceased to carry significant amounts of trade and many have been abandoned to navigation, but may still be used as a system for transportation of untreated water. In some cases railways have been built along the canal route, an example being the Croydon Canal. A movement that began in Britain and France to use the early industrial canals for pleasure boats, such as hotel barges, has spurred rehabilitation of stretches of historic canals. In some cases, abandoned canals such as the Kennet and Avon Canal have been restored and are now used by pleasure boaters. In Britain, canalside housing has also proven popular in recent years. The Seine–Nord Europe Canal is being developed into a major transportation waterway, linking France with Belgium, Germany, and the Netherlands. Canals have found another use in the 21st century, as easements for the installation of fibre optic telecommunications network cabling, avoiding having them buried in roadways while facilitating access and reducing the hazard of being damaged from digging equipment. Canals are still used to provide water for agriculture. An extensive canal system exists within the Imperial Valley in the Southern California desert to provide irrigation to agriculture within the area. Canals are so deeply identified with Venice that many canal cities have been nicknamed "the Venice of…". The city is built on marshy islands, with wooden piles supporting the buildings, so that the land is man-made rather than the waterways. The islands have a long history of settlement; by the 12th century, Venice was a powerful city state. Amsterdam was built in a similar way, with buildings on wooden piles. It became a city around 1300. Many Amsterdam canals were built as part of fortifications. They became grachten when the city was enlarged and houses were built alongside the water. Its nickname as the "Venice of the North" is shared with Hamburg of Germany, St. Petersburg of Russia and Bruges of Belgium. Suzhou was dubbed the "Venice of the East" by Marco Polo during his travels there in the 13th century, with its modern canalside Pingjiang Road and Shantang Street becoming major tourist attractions. Other nearby cities including Nanjing, Shanghai, Wuxi, Jiaxing, Huzhou, Nantong, Taizhou, Yangzhou, and Changzhou are located along the lower mouth of the Yangtze River and Lake Tai, yet another source of small rivers and creeks, which have been canalized and developed for centuries. Other cities with extensive canal networks include: Alkmaar, Amersfoort, Bolsward, Brielle, Delft, Den Bosch, Dokkum, Dordrecht, Enkhuizen, Franeker, Gouda, Haarlem, Harlingen, Leeuwarden, Leiden, Sneek and Utrecht in the Netherlands; Brugge and Gent in Flanders, Belgium; Birmingham in England; Saint Petersburg in Russia; Bydgoszcz, Gdańsk, Szczecin and Wrocław in Poland; Aveiro in Portugal; Hamburg and Berlin in Germany; Fort Lauderdale and Cape Coral in Florida, United States, Wenzhou in China, Cần Thơ in Vietnam, Bangkok in Thailand, and Lahore in Pakistan. Liverpool Maritime Mercantile City was a UNESCO World Heritage Site near the centre of Liverpool, England, where a system of intertwining waterways and docks is now being developed for mainly residential and leisure use. Canal estates (sometimes known as bayous in the United States) are a form of subdivision popular in cities like Miami, Florida, Texas City, Texas and the Gold Coast, Queensland; the Gold Coast has over 890 km of residential canals. Wetlands are difficult areas upon which to build housing estates, so dredging part of the wetland down to a navigable channel provides fill to build up another part of the wetland above the flood level for houses. Land is built up in a finger pattern that provides a suburban street layout of waterfront housing blocks. Inland canals have often had boats specifically built for them. An example of this is the British narrowboat, which is up to 72 feet (21.95 m) long and 7 feet (2.13 m) wide and was primarily built for British Midland canals. In this case the limiting factor was the size of the locks. This is also the limiting factor on the Panama canal where Panamax ships were limited to a length of 289.56 m (950 ft) and a beam of 32.31 m (106 ft) until 26 June 2016 when the opening of larger locks allowed for the passage of larger New Panamax ships. For the lockless Suez Canal the limiting factor for Suezmax ships is generally draft, which is limited to 16 m (52.5 ft). At the other end of the scale, tub-boat canals such as the Bude Canal were limited to boats of under 10 tons for much of their length due to the capacity of their inclined planes or boat lifts. Most canals have a limit on height imposed either by bridges or by tunnels.
[[File:Cessna Logo.svg|100px|thumb|Logo]] '''Cessna''' is a [[company]] that makes [[airplanes]]. The company is named for their founder, Clyde Cessna. They are in [[Wichita, Kansas]] in the [[United States]]. They make small aircraft that can hold 2 to 4 people and in the last decades, they have built several models of business jets. == Cessna aircraft == * [[Cessna 172]] * [[Cessna 152]] * [[Cessna 182]] == Other websites == * [http://www.cessna.com/ Cessna website] {{Transport-stub}} [[Category:American aircraft companies]] [[Category:Wichita, Kansas]]
Cessna (/ˈsɛsnə/) is an American brand of general aviation aircraft owned by Textron Aviation since 2014, headquartered in Wichita, Kansas. Originally, it was a brand of the Cessna Aircraft Company, an American general aviation aircraft manufacturing corporation also headquartered in Wichita. The company produced small, piston-powered aircraft, as well as business jets. For much of the mid-to-late 20th century, Cessna was one of the highest-volume and most diverse producers of general aviation aircraft in the world. It was founded in 1927 by Clyde Cessna and Victor Roos and was purchased by General Dynamics in 1985, then by Textron, Inc. in 1992. In March 2014, when Textron purchased the Beechcraft and Hawker Aircraft corporations, Cessna ceased operations as a subsidiary company, and joined the others as one of the three distinct brands produced by Textron Aviation. Throughout its history, and especially in the years following World War II, Cessna became best-known for producing high-wing, small piston aircraft. Its most popular and iconic aircraft is the Cessna 172, delivered since 1956 (with a break from 1986–1996), with more sold than any other aircraft in history. Since the first model was delivered in 1972, the brand has also been well known for its Citation family of low-wing business jets which vary in size. Clyde Cessna, a farmer in Rago, Kansas, built his own aircraft and flew it in June 1911. He was the first person to do so between the Mississippi River and the Rocky Mountains. Cessna started his wood-and-fabric aircraft ventures in Enid, Oklahoma, testing many of his early planes on the salt flats. When bankers in Enid refused to lend him more money to build his planes, he moved to Wichita. Cessna Aircraft was formed when Clyde Cessna and Victor Roos became partners in the Cessna-Roos Aircraft Company in 1927. Roos resigned just one month into the partnership, selling back his interest to Cessna. Shortly afterward, Roos's name was dropped from the company name. The Cessna DC-6 earned certification on the same day as the stock market crash of 1929, October 29, 1929. In 1932, the Cessna Aircraft Company closed due to the Great Depression. However, the Cessna CR-3 custom racer made its first flight in 1933. The plane won the 1933 American Air Race in Chicago and later set a new world speed record for engines smaller than 500 cubic inches by averaging 237 mph (381 km/h). Cessna's nephews, brothers Dwane and Dwight Wallace, bought the company from Cessna in 1934. They reopened it and began the process of building it into what would become a global success. The Cessna C-37 was introduced in 1937 as Cessna's first seaplane when equipped with Edo floats. In 1940, Cessna received their largest order to date, when they signed a contract with the U.S. Army for 33 specially equipped Cessna T-50s, their first twin engine plane. Later in 1940, the Royal Canadian Air Force placed an order for 180 T-50s. Cessna returned to commercial production in 1946, after the revocation of wartime production restrictions (L-48), with the release of the Model 120 and Model 140. The approach was to introduce a new line of all-metal aircraft that used production tools, dies and jigs, rather than the hand-built tube-and-fabric construction process used before the war. The Model 140 was named by the US Flight Instructors Association as the "Outstanding Plane of the Year" in 1948. Cessna's first helicopter, the Cessna CH-1, received FAA type certification in 1955. Cessna introduced the Cessna 172 in 1956. It became the most produced airplane in history. During the post-World War II era, Cessna was known as one of the "Big Three" in general aviation aircraft manufacturing, along with Piper and Beechcraft. In 1959, Cessna acquired Aircraft Radio Corporation (ARC), of Boonton, New Jersey, a leading manufacturer of aircraft radios. During these years, Cessna expanded the ARC product line, and rebranded ARC radios as "Cessna" radios, making them the "factory option" for avionics in new Cessnas. However, during this time, ARC radios suffered a severe decline in quality and popularity. Cessna kept ARC as a subsidiary until 1983, selling it to avionics-maker Sperry. In 1960, Cessna acquired McCauley Industrial Corporation, of Ohio, a leading manufacturer of propellers for light aircraft. McCauley became the world's leading producer of general aviation aircraft propellers, largely through their installation on Cessna airplanes. In 1960, Cessna affiliated itself with Reims Aviation of Reims, France. In 1963, Cessna produced its 50,000th airplane, a Cessna 172. Cessna's first business jet, the Cessna Citation I, performed its maiden flight on September 15, 1969. Cessna produced its 100,000th single-engine airplane in 1975. In 1985, Cessna ceased to be an independent company. It was purchased by General Dynamics Corporation and became a wholly owned subsidiary. Production of the Cessna Caravan began. General Dynamics in turn sold Cessna to Textron in 1992. Late in 2007, Cessna purchased the bankrupt Columbia Aircraft company for US$26.4M and would continue production of the Columbia 350 and 400 as the Cessna 350 and Cessna 400 at the Columbia factory in Bend, Oregon. However, production of both aircraft had ended by 2018. On November 27, 2007, Cessna announced the then-new Cessna 162 would be built in the People's Republic of China by Shenyang Aircraft Corporation, which is a subsidiary of the China Aviation Industry Corporation I (AVIC I), a Chinese government-owned consortium of aircraft manufacturers. Cessna reported that the decision was made to save money and also that the company had no more plant capacity in the United States at the time. Cessna received much negative feedback for this decision, with complaints centering on the recent quality problems with Chinese production of other consumer products, China's human rights record, exporting of jobs and China's less than friendly political relationship with the United States. The customer backlash surprised Cessna and resulted in a company public relations campaign. In early 2009, the company attracted further criticism for continuing plans to build the 162 in China while laying off large numbers of workers in the United States. In the end, the Cessna 162 was not a commercial success and only a small number were delivered before production was cancelled. The company's business suffered notably during the late-2000s recession, laying off more than half its workforce between January 2009 and September 2010. On November 4, 2008, Cessna's parent company, Textron, indicated that Citation production would be reduced from the original 2009 target of 535 "due to continued softening in the global economic environment" and that this would result in an undetermined number of lay-offs at Cessna. On November 8, 2008, at the Aircraft Owners and Pilots Association (AOPA) Expo, CEO Jack Pelton indicated that sales of Cessna aircraft to individual buyers had fallen, but piston and turboprop sales to businesses had not. "While the economic slowdown has created a difficult business environment, we are encouraged by brisk activity from new and existing propeller fleet operators placing almost 200 orders for 2009 production aircraft," Pelton stated. Beginning in January 2009, a total of 665 jobs were cut at Cessna's Wichita and Bend, Oregon plants. The Cessna factory at Independence, Kansas, which builds the Cessna piston-engined aircraft and the Cessna Mustang, did not see any layoffs, but one third of the workforce at the former Columbia Aircraft facility in Bend was laid off. This included 165 of the 460 employees who built the Cessna 350 and 400. The remaining 500 jobs were eliminated at the main Cessna Wichita plant. In January 2009, the company laid off an additional 2,000 employees, bringing the total to 4,600. The job cuts included 120 at the Bend, Oregon, facility reducing the plant that built the Cessna 350 and 400 to fewer than half the number of workers that it had when Cessna bought it. Other cuts included 200 at the Independence, Kansas, plant that builds the single-engined Cessnas and the Mustang, reducing that facility to 1,300 workers. On April 29, 2009, the company suspended the Citation Columbus program and closed the Bend, Oregon, facility. The Columbus program was finally cancelled in early July 2009. The company reported, "Upon additional analysis of the business jet market related to this product offering, we decided to formally cancel further development of the Citation Columbus". With the 350 and 400 production moving to Kansas, the company indicated that it would lay off 1,600 more workers, including the remaining 150 employees at the Bend plant and up to 700 workers from the Columbus program. In early June 2009, Cessna laid off an additional 700 salaried employees, bringing the total number of lay-offs to 7,600, which was more than half the company's workers at the time. The company closed its three Columbus, Georgia, manufacturing facilities between June 2010 and December 2011. The closures included the new 100,000-square-foot (9,300 m) facility that was opened in August 2008 at a cost of US$25M, plus the McCauley Propeller Systems plant. These closures resulted in total job losses of 600 in Georgia. Some of the work was relocated to Cessna's Independence, Kansas, or Mexican facilities. Cessna's parent company, Textron, posted a loss of US$8M in the first quarter of 2010, largely driven by continuing low sales at Cessna, which were down 44%. Half of Cessna's workforce remained laid-off and CEO Jack Pelton stated that he expected the recovery to be long and slow. In September 2010, a further 700 employees were laid off, bringing the total to 8,000 jobs lost. CEO Jack Pelton indicated this round of layoffs was due to a "stalled [and] lackluster economy" and noted that while the number of orders cancelled for jets had been decreasing, new orders had not met expectations. Pelton added, "our strategy is to defend and protect our current markets while investing in products and services to secure our future, but we can do this only if we succeed in restructuring our processes and reducing our costs." On May 2, 2011, CEO Jack J. Pelton retired. The new CEO, Scott A. Ernest, started on May 31, 2011. Ernest joined Textron after 29 years at General Electric, where he had most recently served as vice president and general manager, global supply chain for GE Aviation. Ernest previously worked for Textron CEO Scott Donnelly when both worked at General Electric. In September 2011, the Federal Aviation Administration (FAA) proposed a US$2.4 million fine against the company for its failure to follow quality assurance requirements while producing fiberglass components at its plant in Chihuahua, Mexico. Excess humidity meant that the parts did not cure correctly and quality assurance did not detect the problems. The failure to follow procedures resulted in the delamination in flight of a 7 ft (2.1 m) section of one Cessna 400's wing skin from the spar while the aircraft was being flown by an FAA test pilot. The aircraft was landed safely. The FAA also discovered 82 other aircraft parts that had been incorrectly made and not detected by the company's quality assurance. The investigation resulted in an emergency Airworthiness Directive that affected 13 Cessna 400s. Since March 2012, Cessna has been pursuing building business jets in China as part of a joint venture with Aviation Industry Corporation of China (AVIC). The company stated that it intends to eventually build all aircraft models in China, saying "The agreements together pave the way for a range of business jets, utility single-engine turboprops and single-engine piston aircraft to be manufactured and certified in China." In late April 2012, the company added 150 workers in Wichita as a result of anticipated increased demand for aircraft production. Overall, they have cut more than 6000 jobs in the Wichita plant since 2009. In March 2014, Cessna ceased operations as a company and instead became a brand of Textron Aviation. During the 1950s and 1960s, Cessna's marketing department followed the lead of Detroit automakers and came up with many unique marketing terms in an effort to differentiate its product line from their competitors. Other manufacturers and the aviation press widely ridiculed and spoofed many of the marketing terms, but Cessna built and sold more aircraft than any other manufacturer during the boom years of the 1960s and 1970s. Generally, the names of Cessna models do not follow a theme, but there is usually logic to the numbering: the 100 series are the light singles, the 200s are the heftier, the 300s are light to medium twins, the 400s have "wide oval" cabin-class accommodation and the 500s are jets. Many Cessna models have names starting with C for the sake of alliteration (e.g. Citation, Crusader, Chancellor). Cessna marketing terminology includes: In October 2020, Textron Aviation was producing the following Cessna-branded models:
[[File:Paris Janvier 1860. Toilette de Mme de Dardel. Fritz von Dardel, 1860 - Nordiska Museet - NMA.0038389.jpg|thumb|right|230px|[[Paris]] 1860]] [[File:Oba Lagos060602-N-8637R-006.jpg |thumb|right|230px|The Oba ([[Nigeria]])]] [[File:Kimono backshot by sth.jpg|thumb|200px|The woman wearing traditional ''Kimono'' (Furisode)]] '''Clothing''' are items used to cover the [[human body]]. [[Human]]s are the only [[animal]]s that wear clothing. During the many thousands of years between losing body hair and learning to make clothes, humans were [[Nudity|naked]].<ref name="naked"/> Some native people in hot places continue to be naked in everyday life. Clothing is worn where the human body needs protection; from the sun and dust in hot, dry countries lacking shade and from the cold and wet in [[temperate climate]]s. Clothing such as thick wool [[coat]]s and boots keeps the human body warm in very cold [[temperature]]s (such as in the [[arctic]]). Clothing is also worn for decoration, as a [[fashion]]. People from different [[culture]]s wear different clothing, and have different beliefs and customs about what type of clothing should be worn. For many people, clothing is a status symbol. It helps people project an image. Often, clothing is a form of self-expression. Adults in different social or work situations present different views of themselves by the clothes they wear. Young people have an entirely different form of dress to express their personalities. Often people will simply follow popular fashion styles so that they will fit in. Clothing is far more than just a means to protect our bodies. [[File:Visita a las obras de Canalejas 01.jpg|left|thumb|Bright clothing so others will see]] Clothing is usually made of [[fabric]] sewn together, but may also be animal skins. Each body part has a typical item of clothing. The [[torso]] can be covered by [[shirt]]s, [[arm]]s by sleeves, [[leg]]s by [[pants]] or [[skirt]]s, hands by [[glove]]s, feet by [[footwear]], and [[head]] by [[Hat|headgear]] or [[mask]]s. In cold climates, people also wear heavy, thick [[coat]]s such as [[overcoat]]<nowiki/>s. ==Origin of clothing== There is no easy way to be sure when clothing was first developed, because it was [[Pre-history|prehistoric]] and clothing is perishable. One of the earliest that has been found is a [[cloak]] made of the [[fur]] of [[squirrel]]s, from a cave in Italy, dated to 23,000 years ago.<ref>{{cite book| first=Johannes | last=Hoops| publisher=Walter de Gruyter| year=2001| title=Reallexikon der Germanischen Altertumskunde| volume=19| page=239| isbn=9783110171631| url=https://books.google.com/books?id=d1URMz8B36oC}}</ref> Some estimates come from studying the biology of [[Louse#Lice_and_humans|lice]]. The body louse lives in clothing, and the last [[ancestor]] that it and head lice both had lived about 107,000 years ago. This suggests that clothing existed at that time.<ref>{{cite journal |author=Ralf Kittler, Manfred Kayser & Mark Stoneking |journal=Current Biology |volume=13 |pages=1414–1417 |year=2003 |title=Molecular evolution of ''Pediculus humanus'' and the origin of clothing |url=http://www.eva.mpg.de/genetics/pdf/Kittler.CurBiol.2003.pdf |doi=10.1016/S0960-9822(03)00507-4 |pmid=12932325 |issue=16 |s2cid=15277254 |access-date=2013-02-02 |archive-date=2008-09-10 |archive-url=https://web.archive.org/web/20080910120019/http://www.eva.mpg.de/genetics/pdf/Kittler.CurBiol.2003.pdf |url-status=dead |issn=0960-9822}}</ref><ref name="Mark Stoneking">{{cite journal|url= https://utsouthwestern.pure.elsevier.com/en/publications/erratum-molecular-evolution-of-pediculus-humanus-and-the-origin-o|last= Stoneking|first= Mark|title= Erratum: Molecular evolution of Pediculus humanus and the origin of clothing (Current Biology 13:16 (pp. 1414-1417))|journal= Current Biology|date= 29 December 2004|volume= 14|issue= 24|page= 2309|doi= 10.1016/j.cub.2004.12.024|accessdate= March 24, 2008}}</ref><ref>[http://mbe.oxfordjournals.org/content/28/1/29.full "...Lice Indicates Early Clothing Use ..."], Mol Biol Evol (2011) 28 (1): 29-32.</ref> Other [[louse]]-based estimates put the introduction of clothing at around 44,000–74,000 years ago.<ref name="naked">{{cite web|url=http://www.redorbit.com/news/science/13258/naked_ape_humans_lost_body_hair_long_before_finding_clothes/|title=Naked ape: Humans lost body hair long before finding clothes|accessdate=2012-03-13|date=2003-08-20}}</ref> In September 2021, scientists reported evidence of [[clothes]] being made 120,000 years ago based on findings in deposits in [[Morocco]], a country in the northwestern part of Africa.<ref name="iSC-20210916">{{cite journal |author=Hallett, Emily Y. |display-authors=et al. |title=A worked bone assemblage from 120,000–90,000 year old deposits at Contrebandiers Cave, Atlantic Coast,Morocco |date=16 September 2021 |journal=[[W:iScience|iScience]] |volume=24 |issue=9 |page=102988 |doi=10.1016/j.isci.2021.102988 |pmid=34622180 |pmc=8478944 |bibcode=2021iSci...24j2988H }}</ref><ref name="TG-20210916">{{cite news |last=Davis |first=Nicola |title=Scientists find evidence of humans making clothes 120,000 years ago - Tools and bones in Moroccan cave could be some of earliest evidence of the hallmark human behaviour |url=https://www.theguardian.com/science/2021/sep/16/scientists-find-evidence-of-humans-making-clothes-120000-years-ago |date=16 September 2021 |work=[[The Guardian]] |accessdate=16 September 2021 }}</ref> Some [[archeology|archeologists]] think [[Neanderthal]]s became extinct because they never invented the sewing needle, and were then unable to make fitted clothing or footwear to survive the extreme cold of the last ice age. == Biblical Story of Clothing == In the [[Book of Genesis]], [[Adam and Eve]] are created [[Nudity|naked]] but don’t know it until they eat a [[fruit]] that they’re not supposed to eat so they cover themselves with leaves. Then [[Names of God in Judaism|God]], sends them out of the [[Garden of Eden]] and He gives them clothing to wear. == Things that are not clothing == People often decorate their bodies with [[makeup]] or [[perfume]], and they also cut or change the hair on their heads and faces. They might also go in for [[body modification]]: [[tattoo]]s, [[scarification]]s, and [[body piercing]]s. But makeup and tattoos are not kinds of clothing. Things that are carried and not worn, like [[wallet]]s, [[purse]]s, [[cane]]s, and [[umbrella]]s, are called [[accessories]], but they are not kinds of clothing, either. [[Jewelry]] and [[eyeglasses]] are also accessories that are put on the body. [[Nail polish]] is also put on the fingertips and can be interpreted as makeup. == What clothing is made of == Clothing is often made of: * [[Cloth]] **Natural fibres, such as [[cotton]], [[flax]], [[wool]], [[hemp]], [[ramie]], [[silk]] **[[Synthetic fabric]], such as [[nylon]], [[polyester]], [[acrylic]] * [[Fur]] * [[Leather]] == Needed to make clothing == * [[Cloth]] * [[Scissors]] or [[knife|knives]]. * [[Needle]]s and [[thread]]. * [[Sewing machine]] or weaving wheel == Related pages == * [[Fashion]] * [[Fashion design]] * [[Template:Human timeline|Human timeline]] == References == {{Commons category|Clothing}} {{Reflist}} {{Clothing}} [[Category:Clothing| ]]
Clothing (also known as clothes, garments, dress, apparel, or attire) is any item worn on the body. Typically, clothing is made of fabrics or textiles, but over time it has included garments made from animal skin and other thin sheets of materials and natural products found in the environment, put together. The wearing of clothing is mostly restricted to human beings and is a feature of all human societies. The amount and type of clothing worn depends on gender, body type, social factors, and geographic considerations. Garments cover the body, footwear covers the feet, gloves cover the hands, while hats and headgear cover the head, and underwear covers the private parts. Clothing serves many purposes: it can serve as protection from the elements, rough surfaces, sharp stones, rash-causing plants, and insect bites, by providing a barrier between the skin and the environment. Clothing can insulate against cold or hot conditions, and it can provide a hygienic barrier, keeping infectious and toxic materials away from the body. It can protect feet from injury and discomfort or facilitate navigation in varied environments. Clothing also provides protection from ultraviolet radiation. It may be used to prevent glare or increase visual acuity in harsh environments, such as brimmed hats. Clothing is used for protection against injury in specific tasks and occupations, sports, and warfare. Fashioned with pockets, belts, or loops, clothing may provide a means to carry things while freeing the hands. Clothing has significant social factors as well. Wearing clothes is a variable social norm. It may connote modesty. Being deprived of clothing in front of others may be embarrassing. In many parts of the world, not wearing clothes in public so that genitals, breast, or buttocks are visible could be considered indecent exposure. Pubic area or genital coverage is the most frequently encountered minimum found cross-culturally and regardless of climate, implying social convention as the basis of customs. Clothing also may be used to communicate social status, wealth, group identity, and individualism. Some forms of personal protective equipment amount to clothing, such as coveralls, chaps or a doctor's white coat, with similar requirements for maintenance and cleaning as other textiles (boxing gloves function both as protective equipment and as a sparring weapon, so the equipment aspect rises above the glove aspect). More specialized forms of protective equipment, such as face shields are classified as protective accessories. At the far extreme, self-enclosing diving suits or space suits are form-fitting body covers, and amount to a form of dress, without being clothing per se, while containing enough high technology to amount to more of a tool than a garment. This line will continue to blur as wearable technology embeds assistive devices directly into the fabric itself; the enabling innovations are ultra low power consumption and flexible electronic substrates. Clothing also hybridizes into a personal transportation system (ice skates, roller skates, cargo pants, other outdoor survival gear, one-man band) or concealment system (stage magicians, hidden linings or pockets in tradecraft, integrated holsters for concealed carry, merchandise-laden trench coats on the black market — where the purpose of the clothing often carries over into disguise). A mode of dress fit to purpose, whether stylistic or functional, is known as an outfit or ensemble. Scientists have never agreed on when humans began wearing clothes and estimates suggested by various experts have ranged greatly, from 40,000 to as many as 3 million years ago. Recent studies by Ralf Kittler, Manfred Kayser and Mark Stoneking—anthropologists at the Max Planck Institute for Evolutionary Anthropology—have attempted to constrain the most recent date of the introduction of clothing with an indirect method relying on lice. The rationale for this method of dating stems from the fact that the human body louse cannot live outside of clothing, dying after only a few hours without shelter. This strongly implies that the date of the body louse's speciation from its parent, Pediculus humanus, can have taken place no earlier than the earliest human adoption of clothing. This date, at which the body louse (P. humanus corporis) diverged from both its parent species and its sibling subspecies, the head louse (P. humanus capitis), can be determined by the number of mutations each has developed during the intervening time. Such mutations occur at a known rate and the date of last-common-ancestor for two species can therefore be estimated from their frequency. These studies have produced dates from 40,000 to 170,000 years ago, with a greatest likelihood of speciation lying at about 107,000 years ago. Kittler, Kayser and Stoneking suggest that the invention of clothing may have coincided with the northward migration of modern Homo sapiens away from the warm climate of Africa, which is thought to have begun between 100,000 and 50,000 years ago. A second group of researchers, also relying on the genetic clock, estimate that clothing originated between 30,000 and 114,000 years ago. Dating with direct archeological evidence produces dates consistent with those hinted at by lice. In September 2021, scientists reported evidence of clothes being made 120,000 years ago based on findings in deposits in Morocco. However, despite these indications, there is no single estimate that is widely accepted. According to Anthropologists and Archaeologists, the earliest clothing likely consisted of fur, leather, leaves, or grass that was draped, wrapped, or tied around the body. Knowledge of such clothing remains inferential, as clothing materials deteriorate quickly compared with stone, bone, shell, and metal artifacts. Archeologists have identified very early sewing needles of bone and ivory from about 30,000 BC, found near Kostenki, Russia in 1988, and in 2016 a needle at least 50,000 years old from Denisova Cave in Siberia made by Denisovans. Dyed flax fibers that date back to 34,000 BC and could have been used in clothing have been found in a prehistoric cave in Georgia. Several distinct human cultures, including those residing in the Arctic Circle, have historically crafted their garments exclusively from treated and adorned animal furs and skins. In contrast, numerous other societies have complemented or substituted leather and skins with textiles woven, knitted, or twined from a diverse array of animal and plant fibers, such as wool, linen, cotton, silk, hemp, and ramie. Although modern consumers may take the production of clothing for granted, making fabric by hand is a tedious and labor-intensive process involving fiber making, spinning, and weaving. The textile industry was the first to be mechanized – with the powered loom – during the Industrial Revolution. Different cultures have evolved various ways of creating clothes out of cloth. One approach involves draping the cloth. Many people wore, and still wear, garments consisting of rectangles of cloth wrapped to fit – for example, the dhoti for men and the sari for women in the Indian subcontinent, the Scottish kilt, and the Javanese sarong. The clothes may be tied up (dhoti and sari) or implement pins or belts to hold the garments in place (kilt and sarong). The cloth remains uncut, and people of various sizes can wear the garment. Another approach involves measuring, cutting, and sewing the cloth by hand or with a sewing machine. Clothing can be cut from a sewing pattern and adjusted by a tailor to the wearer's measurements. An adjustable sewing mannequin or dress form is used to create form-fitting clothing. If the fabric is expensive, the tailor tries to use every bit of the cloth rectangle in constructing the clothing; perhaps cutting triangular pieces from one corner of the cloth, and adding them elsewhere as gussets. Traditional European patterns for shirts and chemises take this approach. These remnants can also be reused to make patchwork pockets, hats, vests, and skirts. Modern European fashion treats cloth much less conservatively, typically cutting in such a way as to leave various odd-shaped cloth remnants. Industrial sewing operations sell these as waste; domestic sewers may turn them into quilts. In the thousands of years that humans have been making clothing, they have created an astonishing array of styles, many of which have been reconstructed from surviving garments, photographs, paintings, mosaics, etc., as well as from written descriptions. Costume history can inspire current fashion designers, as well as costumiers for plays, films, television, and historical reenactment. Comfort is related to various perceptions, physiological, social, and psychological needs, and after food, it is clothing that satisfies these comfort needs. Clothing provides aesthetic, tactile, thermal, moisture, and pressure comfort. The most obvious function of clothing is to protect the wearer from the elements. It serves to prevent wind damage and provides protection from sunburn. In the cold, it offers thermal insulation. Shelter can reduce the functional need for clothing. For example, coats, hats, gloves, and other outer layers are normally removed when entering a warm place. Similarly, clothing has seasonal and regional aspects so that thinner materials and fewer layers of clothing generally are worn in warmer regions and seasons than in colder ones. Boots, hats, jackets, ponchos, and coats designed to protect from rain and snow are specialized clothing items. Clothing has been made from a wide variety of materials, ranging from leather and furs to woven fabrics, to elaborate and exotic natural and synthetic fabrics. Not all body coverings are regarded as clothing. Articles carried rather than worn normally are considered accessories rather than clothing (such as Handbags), items worn on a single part of the body and easily removed (scarves), worn purely for adornment (jewelry), or items that do not serve a protective function. For instance, corrective eyeglasses, Arctic goggles, and sunglasses would not be considered an accessory because of their protective functions. Clothing protects against many things that might injure or irritate the naked human body, including rain, snow, wind, and other weather, as well as from the sun. Garments that are too sheer, thin, small, or tight offer less protection. Appropriate clothes can also reduce risk during activities such as work or sport. Some clothing protects from specific hazards, such as insects, toxic chemicals, weather, weapons, and contact with abrasive substances. Humans have devised clothing solutions to environmental or other hazards: such as space suits, armor, diving suits, swimsuits, bee-keeper gear, motorcycle leathers, high-visibility clothing, and other pieces of protective clothing. The distinction between clothing and protective equipment is not always clear-cut since clothes designed to be fashionable often have protective value, and clothes designed for function often have corporate fashion in their design. The choice of clothes also has social implications. They cover parts of the body that social norms required to be covered, act as a form of adornment, and serve other social purposes. Someone who lacks the means to procure appropriate clothing due to poverty or affordability, or lack of inclination, sometimes is said to be worn, ragged, or shabby. Clothing performs a range of social and cultural functions, such as individual, occupational, gender differentiation, and social status. In many societies, norms about clothing reflect standards of modesty, religion, gender, and social status. Clothing may also function as adornment and an expression of personal taste or style. Serious books on clothing and its functions appear from the nineteenth century as European colonial powers interacted with new environments such as tropical ones in Asia. Some scientific research into the multiple functions of clothing in the first half of the twentieth century, with publications such as J.C. Flügel's Psychology of Clothes in 1930, and Newburgh's seminal Physiology of Heat Regulation and The Science of Clothing in 1949. By 1968, the field of Environmental Physiology had advanced and expanded significantly, but the science of clothing in relation to environmental physiology had changed little. There has since been considerable research, and the knowledge base has grown significantly, but the main concepts remain unchanged, and indeed, Newburgh's book continues to be cited by contemporary authors, including those attempting to develop thermoregulatory models of clothing development. Clothing reveals much about human history. According to Professor Kiki Smith of Smith College, garments preserved in collections are resources for study similar to books and paintings. Scholars around the world have studied a wide range of clothing topics, including the history of specific items of clothing, clothing styles in different cultural groups, and the business of clothing and fashion. The textile curator Linda Baumgarten writes that "clothing provides a remarkable picture of the daily lives, beliefs, expectations, and hopes of those who lived in the past. Clothing presents a number of challenges to historians. Clothing made of textiles or skins is subject to decay, and the erosion of physical integrity may be seen as a loss of cultural information. Costume collections often focus on important pieces of clothing considered unique or otherwise significant, limiting the opportunities scholars have to study everyday clothing. In most cultures, gender differentiation of clothing is considered appropriate. The differences are in styles, colors, fabrics, and types. In contemporary Western societies, skirts, dresses, and high-heeled shoes are usually seen as women's clothing, while neckties usually are seen as men's clothing. Trousers were once seen as exclusively men's clothing, but nowadays are worn by both genders. Men's clothes are often more practical (that is, they can function well under a wide variety of situations), but a wider range of clothing styles is available for women. Typically, men are allowed to bare their chests in a greater variety of public places. It is generally common for a woman to wear clothing perceived as masculine, while the opposite is seen as unusual. Contemporary men may sometimes choose to wear men's skirts such as togas or kilts in particular cultures, especially on ceremonial occasions. In previous times, such garments often were worn as normal daily clothing by men. In some cultures, sumptuary laws regulate what men and women are required to wear. Islam requires women to wear certain forms of attire, usually hijab. What items required varies in different Muslim societies; however, women are usually required to cover more of their bodies than men. Articles of clothing Muslim women wear under these laws or traditions range from the head-scarf to the burqa. Some contemporary clothing styles designed to be worn by either gender, such as T-shirts, have started out as menswear, but some articles, such as the fedora, originally were a style for women. During the early modern period, individuals utilized their attire as a significant method of conveying and asserting their social status. Individuals employed the utilization of high-quality fabrics and trendy designs as a means of communicating their wealth and social standing, as well as an indication of their knowledge and understanding of current fashion trends to the general public. As a result, clothing played a significant role in making the social hierarchy perceptible to all members of society. In some societies, clothing may be used to indicate rank or status. In ancient Rome, for example, only senators could wear garments dyed with Tyrian purple. In traditional Hawaiian society, only high-ranking chiefs could wear feather cloaks and palaoa, or carved whale teeth. In China, before establishment of the republic, only the emperor could wear yellow. History provides many examples of elaborate sumptuary laws that regulated what people could wear. In societies without such laws, which includes most modern societies, social status is signaled by the purchase of rare or luxury items that are limited by cost to those with wealth or status. In addition, peer pressure influences clothing choice. Some religious clothing might be considered a special case of occupational clothing. Sometimes it is worn only during the performance of religious ceremonies. However, it may be worn every day as a marker for special religious status. Sikhs wear a turban as it is a part of their religion. In some religions such as Hinduism, Sikhism, Buddhism, and Jainism the cleanliness of religious dresses is of paramount importance and considered to indicate purity. Jewish ritual requires rending (tearing) of one's upper garment as a sign of mourning. The Quran says about husbands and wives, regarding clothing: "...They are clothing/covering (Libaas) for you; and you for them" (chapter 2:187).Christian clergy members wear religious vestments during liturgical services and may wear specific non-liturgical clothing at other times. Clothing appears in numerous contexts in the Bible. The most prominent passages are: the story of Adam and Eve who made coverings for themselves out of fig leaves, Joseph's coat of many colors, and the clothing of Judah and Tamar, Mordecai and Esther. Furthermore, the priests officiating in the Temple in Jerusalem had very specific garments, the lack of which made one liable to death. The Western dress code has changed over the past 500+ years. The mechanization of the textile industry made many varieties of cloth widely available at affordable prices. Styles have changed, and the availability of synthetic fabrics has changed the definition of what is "stylish". In the latter half of the twentieth century, blue jeans became very popular, and are now worn to events that normally demand formal attire. Activewear has also become a large and growing market. Jn the Western dress code, jeans are worn by both men and women. There are several unique styles of jeans found that include: high rise jeans, mid rise jeans, low rise jeans, bootcut jeans, straight jeans, cropped jeans, skinny jeans, cuffed jeans, boyfriend jeans, and capri jeans. The licensing of designer names was pioneered by designers such as Pierre Cardin, Yves Saint Laurent, and Guy Laroche in the 1960s and has been a common practice within the fashion industry from about the 1970s. Among the more popular include Marc Jacobs and Gucci, named for Marc Jacobs Guccio Gucci respectively. By the early years of the twenty-first century, western clothing styles had, to some extent, become international styles. This process began hundreds of years earlier, during the periods of European colonialism. The process of cultural dissemination has been perpetuated over the centuries, spreading Western culture and styles, most recently as Western media corporations have penetrated markets throughout the world. Fast fashion clothing has also become a global phenomenon. These garments are less expensive, mass-produced Western clothing. Also, donated used clothing from Western countries is delivered to people in poor countries by charity organizations. People may wear ethnic or national dress on special occasions or in certain roles or occupations. For example, most Korean men and women have adopted Western-style dress for daily wear, but still wear traditional hanboks on special occasions, such as weddings and cultural holidays. Also, items of Western dress may be worn or accessorized in distinctive, non-Western ways. A Tongan man may combine a used T-shirt with a Tongan wrapped skirt, or tupenu. For practical, comfort or safety reasons, most sports and physical activities are practised wearing special clothing. Common sportswear garments include shorts, T-shirts, tennis shirts, leotards, tracksuits, and trainers. Specialized garments include wet suits (for swimming, diving, or surfing), salopettes (for skiing), and leotards (for gymnastics). Also, spandex materials often are used as base layers to soak up sweat. Spandex is preferable for active sports that require form fitting garments, such as volleyball, wrestling, track and field, dance, gymnastics, and swimming. Paris set the 1900–1940 fashion trends for Europe and North America. In the 1920s the goal was all about getting loose. Women wore dresses all day, every day. Day dresses had a drop waist, which was a sash or belt around the low waist or hip and a skirt that hung anywhere from the ankle on up to the knee, never above. Day wear had sleeves (long to mid-bicep) and a skirt that was straight, pleated, hank hemmed, or tiered. Jewelry was not conspicuous. Hair was often bobbed, giving a boyish look. In the early twenty-first century a diverse range of styles exists in fashion, varying by geography, exposure to modern media, economic conditions, and ranging from expensive haute couture, to traditional garb, to thrift store grunge. Fashion shows are events for designers to show off new and often extravagant designs. Although mechanization transformed most aspects of human clothing industry, by the mid-twentieth century, garment workers have continued to labor under challenging conditions that demand repetitive manual labor. Often, mass-produced clothing is made in what are considered by some to be sweatshops, typified by long work hours, lack of benefits, and lack of worker representation. While most examples of such conditions are found in developing countries, clothes made in industrialized nations may also be manufactured under similar conditions. Coalitions of NGOs, designers (including Katharine Hamnett, American Apparel, Veja, Quiksilver, eVocal, and Edun), and campaign groups such as the Clean Clothes Campaign (CCC) and the Institute for Global Labour and Human Rights as well as textile and clothing trade unions have sought to improve these conditions by sponsoring awareness-raising events, which draw the attention of both the media and the general public to the plight of the workers. Outsourcing production to low wage countries such as Bangladesh, China, India, Indonesia, Pakistan, and Sri Lanka became possible when the Multi Fibre Agreement (MFA) was abolished. The MFA, which placed quotas on textiles imports, was deemed a protectionist measure. Although many countries recognize treaties such as the International Labour Organization, which attempt to set standards for worker safety and rights, many countries have made exceptions to certain parts of the treaties or failed to thoroughly enforce them. India for example has not ratified sections 87 and 92 of the treaty. The production of textiles has functioned as a consistent industry for developing nations, providing work and wages, whether construed as exploitative or not, to millions of people. The use of animal fur in clothing dates to prehistoric times. Currently, although fur is still used by indigenous people in arctic zones and higher elevations for its warmth and protection, in developed countries it is associated with expensive, designer clothing. Once uncontroversial, recently it has been the focus of campaigns on the grounds that campaigners consider it cruel and unnecessary. PETA, along with other animal and animal liberation groups have called attention to fur farming and other practices they consider cruel. Real fur in fashion is contentious, with Copenhagen (2022) and London (2018) fashion weeks banning real fur in its runway shows following protests and government attention to the issue. Fashion houses such as Gucci and Chanel have banned the use of fur in its garments. Versace and Furla also stopped using fur in their collections in early 2018. In 2020, the outdoor brand Canada Goose announced it would discontinue the use of new coyote fur on parka trims following protests. Governing bodies have issued legislation banning the sale of new real fur garments. In 2021, Israel was the first government to ban the sale of real fur garments, with the exception of those worn as part of a religious faith. In 2019, the state of California banned fur trapping, with a total ban on the sale of all new fur garments except those made of sheep, cow, and rabbit fur going into effect on January 1, 2023. Clothing suffers assault both from within and without. The human body sheds skin cells and body oils, and it exudes sweat, urine, and feces that may soil clothing. From the outside, sun damage, moisture, abrasion, and dirt assault garments. Fleas and lice can hide in seams. If not cleaned and refurbished, clothing becomes worn and loses its aesthetics and functionality (as when buttons fall off, seams come undone, fabrics thin or tear, and zippers fail). Often, people wear an item of clothing until it falls apart. Some materials present problems. Cleaning leather is difficult, and bark cloth (tapa) cannot be washed without dissolving it. Owners may patch tears and rips, and brush off surface dirt, but materials such as these inevitably age. Most clothing consists of cloth, however, and most cloth can be laundered and mended (patching, darning, but compare felt). Humans have developed many specialized methods for laundering clothing, ranging from early methods of pounding clothes against rocks in running streams, to the latest in electronic washing machines and dry cleaning (dissolving dirt in solvents other than water). Hot water washing (boiling), chemical cleaning, and ironing are all traditional methods of sterilizing fabrics for hygiene purposes. Many kinds of clothing are designed to be ironed before they are worn to remove wrinkles. Most modern formal and semi-formal clothing is in this category (for example, dress shirts and suits). Ironed clothes are believed to look clean, fresh, and neat. Much contemporary casual clothing is made of knit materials that do not readily wrinkle, and do not require ironing. Some clothing is permanent press, having been treated with a coating (such as polytetrafluoroethylene) that suppresses wrinkles and creates a smooth appearance without ironing. Excess lint or debris may end up on the clothing in between launderings. In such cases, a lint remover may be useful. Once clothes have been laundered and possibly ironed, usually they are hung on clothes hangers or folded, to keep them fresh until they are worn. Clothes are folded to allow them to be stored compactly, to prevent creasing, to preserve creases, or to present them in a more pleasing manner, for instance, when they are put on sale in stores. Certain types of insects and larvae feed on clothing and textiles, such as the black carpet beetle and clothing moths. To deter such pests, clothes may be stored in cedar-lined closets or chests, or placed in drawers or containers with materials having pest repellent properties, such as lavender or mothballs. Airtight containers (such as sealed, heavy-duty plastic bags) may deter insect pest damage to clothing materials as well. A resin used for making non-wrinkle shirts releases formaldehyde, which could cause contact dermatitis for some people; no disclosure requirements exist, and in 2008 the U.S. Government Accountability Office tested formaldehyde in clothing and found that generally the highest levels were in non-wrinkle shirts and pants. In 1999, a study of the effect of washing on the formaldehyde levels found that after six months of routine washing, 7 of 27 shirts still had levels in excess of 75 ppm (the safe limit for direct skin exposure). When the raw material – cloth – was worth more than labor, it made sense to expend labor in saving it. In past times, mending was an art. A meticulous tailor or seamstress could mend rips with thread raveled from hems and seam edges so skillfully that the tear was practically invisible. Today clothing is considered a consumable item. Mass-manufactured clothing is less expensive than the labor required to repair it. Many people buy a new piece of clothing rather than spend time mending. The thrifty still replace zippers and buttons and sew up ripped hems, however. Other mending techniques include darning and invisible mending or upcycling through visible mending inspired in Japanese Sashiko. It is estimated that 80 billion to 150 billion garments are produced annually. Used, unwearable clothing can be repurposed for quilts, rags, rugs, bandages, and many other household uses. Neutral colored or undyed cellulose fibers can be recycled into paper. In Western societies, used clothing is often thrown out or donated to charity (such as through a clothing bin). It is also sold to consignment shops, dress agencies, flea markets, and in online auctions. Also, used clothing often is collected on an industrial scale to be sorted and shipped for re-use in poorer countries. Globally, used clothes are worth $4 billion, with the U.S. as the leading exporter at $575 million. Synthetics, which come primarily from petrochemicals, are not renewable or biodegradable. Excess inventory of clothing is sometimes destroyed to preserve brand value. EU member states imported €166 billion of clothes in 2018; 51% came from outside the EU (€84 billion). EU member states exported €116 billion of clothes in 2018, including 77% to other EU member states.
[[Image:Kinderschoenen.jpg|thumb|A pair of children's shoes]] A '''shoe''' is a type of [[footwear]]. It is an item of [[clothing]]. You can wear them outside. Shoes come in pairs, with one shoe for each foot. People usually wear shoes in public. They are worn for [[hygiene]], [[fashion|style]], and comfort. While there are some shoes that don't need socks to wear, shoes are usually worn with [[sock]]s. Shoes are a type of foot protection. For example, many shoes can protect a person's foot from [[sand]]; others do not. == Types of shoes == There are many different types of shoes. [[Athletic shoe]]s, for example, are lightweight to make [[running]], [[walking]] or [[jogging]] easier. The sole of the shoe is also softer. [[Slipper]]s are a kind of indoor shoe. They often serve to keep feet warm. Tall shoes are [[boot|boots]]. Rubber boots are used in wet places. Shoes are also worn for fashion. Shoes with a high heel are called [[high heels]]. Shoes for formal wear are made of [[leather]]. They are worn for business and ceremonial functions. == Other websites == * {{commonscat-inline|Shoes}} [[Category:Basic English 850 words]] [[Category:Footwear]] {{stub}}
A shoe is an item of footwear intended to protect and comfort the human foot. Though the human foot can adapt to varied terrains and climate conditions, it is vulnerable, and shoes provide protection. Form was originally tied to function but over time shoes also became fashion items. Some shoes are worn as safety equipment, such as steel-toe boots, which are required footwear at industrial worksites. Additionally, fashion has often evolved into many different designs, such as high heels, which are most commonly worn by women during fancy occasions. Contemporary footwear varies vastly in style, complexity and cost. Basic sandals may consist of only a thin sole and simple strap and be sold for a low cost. High fashion shoes made by famous designers may be made of expensive materials, use complex construction and sell for large sums of money. Some shoes are designed for specific purposes, such as boots designed specifically for mountaineering or skiing, while others have more generalized usage such as sneakers which have transformed from a special purpose sport shoe into a general use shoe. Traditionally, shoes have been made from leather, wood or canvas, but are increasingly being made from rubber, plastics, and other petrochemical-derived materials. Globally, the shoe industry is a $200 billion a year industry. 90% of shoes end up in landfills, because the materials are hard to separate, recycle or otherwise reuse. The earliest known shoes are sagebrush bark sandals dating from approximately 7000 or 8000 BC, found in the Fort Rock Cave in the US state of Oregon in 1938. The world's oldest leather shoe, made from a single piece of cowhide laced with a leather cord along seams at the front and back, was found in the Areni-1 cave complex in Armenia in 2008 and is believed to date to 3500 BC. Ötzi the Iceman's shoes, dating to 3300 BC, featured brown bearskin bases, deerskin side panels, and a bark-string net, which pulled tight around the foot. The Jotunheimen shoe was discovered in August 2006: archaeologists estimate that this leather shoe was made between 1800 and 1100 BC, making it the oldest article of clothing discovered in Scandinavia. Sandals and other plant fiber based tools were found in Cueva de los Murciélagos in Albuñol in southern Spain in 2023, dating to approximately 7500 to 4200 BC, making them what are believed to be the oldest shoes found in Europe. It is thought that shoes may have been used long before this, but because the materials used were highly perishable, it is difficult to find evidence of the earliest footwear. Footprints suggestive of shoes or sandals due to having crisp edges, no signs of toes found and three small divots where leather tying laces/straps would have been attached have been at Garden Route National Park, Addo Elephant National Park and Goukamma Nature Reserve in South Africa. These date back to between 73,000 and 136,000 PB. Consistent with the existence of such shoe is the finding of bone awls dating back to this period that could have made simple footwear. Another source of evidence is the study of the bones of the smaller toes (as opposed to the big toe), it was observed that their thickness decreased approximately 40,000 to 26,000 years ago. This led archaeologists to deduce the existence of common rather than an occasional wearing of shoes as this would lead to less bone growth, resulting in shorter, thinner toes. These earliest designs were very simple, often mere "foot bags" of leather to protect the feet from rocks, debris, and cold. Many early natives in North America wore a similar type of footwear, known as the moccasin. These are tight-fitting, soft-soled shoes typically made out of leather or bison hides. Many moccasins were also decorated with various beads and other adornments. Moccasins were not designed to be waterproof, and in wet weather and warm summer months, most Native Americans went barefoot. The leaves of the sisal plant were used to make twine for sandals in South America while the natives of Mexico used the Yucca plant. As civilizations began to develop, thong sandals (precursors to the modern flip-flop) were worn. This practice dates back to pictures of them in ancient Egyptian murals from 4000 BC. "Thebet" may have been the term used to describe these sandals in Egyptian times, possibly from the city Thebes. The Middle Kingdom is when the first of these thebets were found, but it is possible that it debuted in the Early Dynastic Period. One pair found in Europe was made of papyrus leaves and dated to be approximately 1,500 years old. They were also worn in Jerusalem during the first century of the Christian era. Thong sandals were worn by many civilizations and made from a vast variety of materials. Ancient Egyptian sandals were made from papyrus and palm leaves. The Masai of Africa made them out of rawhide. In India they were made from wood. While thong sandals were commonly worn, many people in ancient times, such as the Egyptians, Hindus and Greeks, saw little need for footwear, and most of the time, preferred being barefoot. The Egyptians and Hindus made some use of ornamental footwear, such as a soleless sandal known as a "Cleopatra", which did not provide any practical protection for the foot. The ancient Greeks largely viewed footwear as self-indulgent, unaesthetic and unnecessary. Shoes were primarily worn in the theater, as a means of increasing stature, and many preferred to go barefoot. Athletes in the Ancient Olympic Games participated barefoot—and naked. Even the gods and heroes were primarily depicted barefoot, as well as the hoplite warriors. They fought battles in bare feet and Alexander the Great conquered his vast empire with barefoot armies. The runners of Ancient Greece had also been believed to have run barefoot. The Romans, who eventually conquered the Greeks and adopted many aspects of their culture, did not adopt the Greek perception of footwear and clothing. Roman clothing was seen as a sign of power, and footwear was seen as a necessity of living in a civilized world, although the slaves and paupers usually went barefoot. Roman soldiers were issued with chiral (left and right shoe different) footwear. Shoes for soldiers had riveted insoles to extend the life of the leather, increase comfort, and provide better traction. The design of these shoes also designated the rank of the officers. The more intricate the insignia and the higher up the boot went on the leg, the higher the rank of the soldier. There are references to shoes being worn in the Bible. In China and Japan, rice straws were used. Starting around 4 BC, the Greeks began wearing symbolic footwear. These were heavily decorated to clearly indicate the status of the wearer. Courtesans wore leather shoes colored with white, green, lemon or yellow dyes, and young woman betrothed or newly married wore pure white shoes. Because of the cost to lighten leather, shoes of a paler shade were a symbol of wealth in the upper class. Often, the soles would be carved with a message so it would imprint on the ground. Cobblers became a notable profession around this time, with Greek shoemakers becoming famed in the Roman empire. A common casual shoe in the Pyrenees during the Middle Ages was the espadrille. This is a sandal with braided jute soles and a fabric upper portion, and often includes fabric laces that tie around the ankle. The term is French and comes from the esparto grass. The shoe originated in the Catalonian region of Spain as early as the 13th century, and was commonly worn by peasants in the farming communities in the area. New styles began to develop during the Song dynasty in China, one of them being the debut of foot straps. It was first used by the noble Han classes, but soon developed throughout society. Women would use these shoes to develop their "lotus feet", which would entice the males. The practice allegedly started during the Shang dynasty, but it grew popular by c. AD 960. When the Mongols conquered China, they dissolved the practice in 1279, and the Manchus banned foot binding in 1644. The Han people, however, continued to use the style without much government intervention. In medieval times shoes could be up to two feet long, with their toes sometimes filled with hair, wool, moss, or grass. Many medieval shoes were made using the turnshoe method of construction, in which the upper was turned flesh side out, and was lasted onto the sole and joined to the edge by a seam. The shoe was then turned inside-out so that the grain was outside. Some shoes were developed with toggled flaps or drawstrings to tighten the leather around the foot for a better fit. Surviving medieval turnshoes often fit the foot closely, with the right and left shoe being mirror images. Around 1500, the turnshoe method was largely replaced by the welted rand method (where the uppers are sewn to a much stiffer sole and the shoe cannot be turned inside-out). The turn shoe method is still used for some dance and specialty shoes. By the 15th century, pattens became popular by both men and women in Europe. These are commonly seen as the predecessor of the modern high-heeled shoe, while the poor and lower classes in Europe, as well as slaves in the New World, were barefoot. In the 15th century, the Crakow was fashionable in Europe. This style of shoe is named because it is thought to have originated in Kraków, the capital of Poland. The style is characterized by the point of the shoe, known as the "polaine", which often was supported by a whalebone tied to the knee to prevent the point getting in the way while walking. Also during the 15th century, chopines were created in Turkey, and were usually 7–8 in (180–200 mm) high. These shoes became popular in Venice and throughout Europe, as a status symbol revealing wealth and social standing. During the 16th century, royalty, such as Catherine de Medici or Mary I of England, started wearing high-heeled shoes to make them look taller or larger than life. By 1580, even men wore them, and a person with authority or wealth was often referred to as, "well-heeled". In 17th century France, heels were exclusively worn by aristocrats. Louis XIV of France outlawed anybody from wearing red high heels except for himself and his royal court. Eventually the modern shoe, with a sewn-on sole, was devised. Since the 17th century, most leather shoes have used a sewn-on sole. This remains the standard for finer-quality dress shoes today. Until around 1800, welted rand shoes were commonly made without differentiation for the left or right foot. Such shoes are now referred to as "straights". Only gradually did the modern foot-specific shoe become standard. Shoemaking became more commercialized in the mid-18th century, as it expanded as a cottage industry. Large warehouses began to stock footwear, made by many small manufacturers from the area. Until the 19th century, shoemaking was a traditional handicraft, but by the century's end, the process had been almost completely mechanized, with production occurring in large factories. Despite the obvious economic gains of mass production, the factory system produced shoes without the individual differentiation that the traditional shoemaker was able to provide. The 19th century was when Chinese feminists called for an end to the use of foot straps, and a ban in 1902 was implemented. The ban was soon repealed until it was banned again in 1911 by the new Nationalist government. It was effective in coastal cities, but countryside cities continued without much regulation. Mao Zedong enforced the rule in 1949 and it continues throughout contemporary times. A number of people still have lotus feet today. The first steps towards mechanisation were taken during the Napoleonic Wars by the engineer, Marc Brunel. He developed machinery for the mass production of boots for the soldiers of the British Army. In 1812, he devised a scheme for making nailed-boot-making machinery that automatically fastened soles to uppers by means of metallic pins or nails. With the support of the Duke of York, the shoes were manufactured, and, due to their strength, cheapness, and durability, were introduced for the use of the army. In the same year, the use of screws and staples was patented by Richard Woodman. Brunel's system was described by Sir Richard Phillips as a visitor to his factory in Battersea as follows: In another building I was shown his manufactory of shoes, which, like the other, is full of ingenuity, and, in regard to subdivision of labour, brings this fabric on a level with the oft-admired manufactory of pins. Every step in it is affected by the most elegant and precise machinery; while, as each operation is performed by one hand, so each shoe passes through twenty-five hands, who complete from the hide, as supplied by the currier, a hundred pairs of strong and well-finished shoes per day. All the details are performed by the ingenious application of the mechanic powers; and all the parts are characterised by precision, uniformity, and accuracy. As each man performs but one step in the process, which implies no knowledge of what is done by those who go before or follow him, so the persons employed are not shoemakers, but wounded soldiers, who are able to learn their respective duties in a few hours. The contract at which these shoes are delivered to Government is 6s. 6d. per pair, being at least 2s. less than what was paid previously for an unequal and cobbled article. However, when the war ended in 1815, manual labour became much cheaper, and the demand for military equipment subsided. As a consequence, Brunel's system was no longer profitable and it soon ceased business. Similar exigencies at the time of the Crimean War stimulated a renewed interest in methods of mechanization and mass-production, which proved longer lasting. A shoemaker in Leicester, Tomas Crick, patented the design for a riveting machine in 1853. His machine used an iron plate to push iron rivets into the sole. The process greatly increased the speed and efficiency of production. He also introduced the use of steam-powered rolling-machines for hardening leather and cutting-machines, in the mid-1850s. The sewing machine was introduced in 1846, and provided an alternative method for the mechanization of shoemaking. By the late 1850s, the industry was beginning to shift towards the modern factory, mainly in the US and areas of England. A shoe-stitching machine was invented by the American Lyman Blake in 1856 and perfected by 1864. Entering into a partnership with McKay, his device became known as the McKay stitching machine and was quickly adopted by manufacturers throughout New England. As bottlenecks opened up in the production line due to these innovations, more and more of the manufacturing stages, such as pegging and finishing, became automated. By the 1890s, the process of mechanisation was largely complete. On January 24, 1899, Humphrey O'Sullivan of Lowell, Massachusetts, was awarded a patent for a rubber heel for boots and shoes. A process for manufacturing stitchless, that is, glued, shoes—AGO—was developed in 1910. Since the mid-20th century, advances in rubber, plastics, synthetic cloth, and industrial adhesives have allowed manufacturers to create shoes that stray considerably from traditional crafting techniques. Leather, which had been the primary material in earlier styles, has remained standard in expensive dress shoes, but athletic shoes often have little or no real leather. Soles, which were once laboriously hand-stitched on, are now more often machine stitched or simply glued on. Many of these newer materials, such as rubber and plastics, have made shoes less biodegradable. It is estimated that most mass-produced shoes require 1000 years to degrade in a landfill. In the late 2000s, some shoemakers picked up on the issue and began to produce shoes made entirely from degradable materials, such as the Nike Considered. In 2007, the global shoe industry had an overall market of $107.4 billion, in terms of revenue, and is expected to grow to $122.9 billion by the end of 2012. Shoe manufacturers in the People's Republic of China account for 63% of production, 40.5% of global exports and 55% of industry revenue. However, many manufacturers in Europe dominate the higher-priced, higher value-added end of the market. As an integral part of human culture and civilization, shoes have found their way into our culture, folklore, and art. A popular 18th-century nursery rhyme is There was an Old Woman Who Lived in a Shoe. This story tells about an old woman living in a shoe with a lot of children. In 1948, Mahlon Haines, a shoe salesman in Hallam, Pennsylvania, built an actual house shaped like a work boot as a form of advertisement. The Haines Shoe House was rented to newlyweds and the elderly until his death in 1962. Since then, it has served as an ice cream parlor, a bed and breakfast, and a museum. It still stands today and is a popular roadside attraction. Shoes also play an important role in the fairy tales Cinderella and The Red Shoes. In the movie adaption of the children's book The Wonderful Wizard of Oz, a pair of red ruby slippers play a key role in the plot. The 1985 comedy The Man with One Red Shoe features an eccentric man wearing one normal business shoe and one red shoe that becomes central to the plot. One poem, written by Phebus Etienne with the title "Shoes", focuses on them. It describes religious messages and is 3 stanzas long. The first stanza is one line, whereas the second is 13 lines and the third being 14 lines. Throughout the poem the main character talks about their dead mother and their routine with her grave. Haitians are said to "not put shoes on the dead." as it makes spirits easier to "step over the offerings". Athletic sneaker collection has also existed as a part of urban subculture in the United States for several decades. Recent decades have seen this trend spread to European nations such as the Czech Republic. A Sneakerhead is a person who owns multiple pairs of shoes as a form of collection and fashion. A contributor to the growth of sneaker collecting is the continued global popularity of the Air Jordan line of sneakers designed by Nike for Basketball star Michael Jordan. In the Bible's Old Testament, the shoe is used to symbolize something that is worthless or of little value. In the New Testament, the act of removing one's shoes symbolizes servitude. Ancient Semitic-speaking peoples regarded the act of removing their shoes as a mark of reverence when approaching a sacred person or place. In the Book of Exodus, Moses was instructed to remove his shoes before approaching the burning bush: Put off thy shoes from off thy feet, for the place whereon thou standest [is] holy ground. The removal of the shoe also symbolizes the act of giving up a legal right. In Hebrew custom, if a man chose not to marry his childless brother's widow, the widow removed her brother-in-law's shoe to symbolize that he had abandoned his duty. In Arab custom, the removal of one's shoe also symbolized the dissolution of marriage. In Arab culture, showing the sole of one's shoe is considered an insult, and to throw a shoe and hit someone with it is considered an even greater insult. Shoes are considered to be dirty as they frequently touch the ground, and are associated with the lowest part of the body—the foot. As such, shoes are forbidden in mosques, and it is also considered unmannerly to cross the legs and display the soles of one's shoes during conversation. This insult was demonstrated in Iraq, first when Saddam Hussein's statue was toppled in 2003, Iraqis gathered around it and struck the statue with their shoes. In 2008, United States President George W. Bush had a shoe thrown at him by a journalist as a statement against the war in Iraq. More generally, shoe-throwing or shoeing, showing the sole of one's shoe or using shoes to insult are forms of protest in many parts of the world. Incidents where shoes were thrown at political figures have taken place in Australia, India, Ireland, Taiwan, Hong Kong, Pakistan, the United Kingdom, the United States, and most notably the Arab world. Empty shoes may also symbolize death. In Greek culture, empty shoes are the equivalent of the American funeral wreath. For example, empty shoes placed outside of a Greek home would tell others that the family's son has died in battle. At an observation memorializing the 10th anniversary of the September 11 attacks, 3,000 pairs of empty shoes were used to recognize those killed. The Shoes on the Danube Bank is a memorial in Budapest, Hungary. Conceived by film director Can Togay, he created it on the east bank of the Danube River with sculptor Gyula Pauer to honor the Jews who were killed by fascist Arrow Cross militiamen in Budapest during World War II. They were ordered to take off their shoes and were shot at the edge of the water so that their bodies fell into the river and were carried away. The memorial represents their shoes left behind on the bank. The basic anatomy of a shoe is recognizable, regardless of the specific style of footwear. All shoes have a sole, which is the bottom of a shoe, in contact with the ground. Soles can be made from a variety of materials, although most modern shoes have soles made from natural rubber, polyurethane, or polyvinyl chloride (PVC) compounds. Soles can be simple—a single material in a single layer—or they can be complex, with multiple structures or layers and materials. When various layers are used, soles may consist of an insole, midsole, and an outsole. The insole is the interior bottom of a shoe, which sits directly beneath the foot under the footbed (also known as sock liner). The purpose of the insole is to attach to the lasting margin of the upper, which is wrapped around the last during the closing of the shoe during the lasting operation. Insoles are usually made of cellulosic paper board or synthetic non woven insole board. Many shoes have removable and replaceable footbeds. Extra cushioning is often added for comfort (to control the shape, moisture, or smell of the shoe) or health reasons (to help deal with differences in the natural shape of the foot or positioning of the foot during standing or walking). The outsole is the layer in direct contact with the ground. Dress shoes often have leather or resin rubber outsoles; casual or work-oriented shoes have outsoles made of natural rubber or a synthetic material like polyurethane. The outsole may comprise a single piece or may be an assembly of separate pieces, often of different materials. On some shoes, the heel of the sole has a rubber plate for durability and traction, while the front is leather for style. Specialized shoes will often have modifications on this design: athletic or so-called cleated shoes like soccer, rugby, baseball and golf shoes have spikes embedded in the outsole to improve traction. The midsole is the layer in between the outsole and the insole, typically there for shock absorption. Some types of shoes, like running shoes, have additional material for shock absorption, usually beneath the heel of the foot, where one puts the most pressure down. Some shoes may not have a midsole at all. The heel is the bottom rear part of a shoe. Its function is to support the heel of the foot. They are often made of the same material as the sole of the shoe. This part can be high for fashion or to make the person look taller, or flat for more practical and comfortable use. On some shoes the inner forward point of the heel is chiselled off, a feature known as a "gentleman's corner". This piece of design is intended to alleviate the problem of the points catching the bottom of trousers and was first observed in the 1930s. A heel is the projection at the back of a shoe which rests below the heel bone. The shoe heel is used to improve the balance of the shoe, increase the height of the wearer, alter posture or other decorative purposes. Sometimes raised, the high heel is common to a form of shoe often worn by women, but sometimes by men too. See also stiletto heel. The upper helps hold the shoe onto the foot. In the simplest cases, such as sandals or flip-flops, this may be nothing more than a few straps for holding the sole in place. Closed footwear, such as boots, trainers and most men's shoes, will have a more complex upper. This part is often decorated or is made in a certain style to look attractive. The upper is connected to the sole by a strip of leather, rubber, or plastic that is stitched between it and the sole, known as a welt. Most uppers have a mechanism, such as laces, straps with buckles, zippers, elastic, velcro straps, buttons, or snaps, for tightening the upper on the foot. Uppers with laces usually have a tongue that helps seal the laced opening and protect the foot from abrasion by the laces. Uppers with laces also have eyelets or hooks to make it easier to tighten and loosen the laces and to prevent the lace from tearing through the upper material. An aglet is the protective wrapping on the end of the lace. The vamp is the front part of the shoe, starting behind the toe, extending around the eyelets and tongue and towards back part of the shoe. The medial is the part of the shoe closest to a person's center of symmetry, and the lateral is on the opposite side, away from their center of symmetry. This can be in reference to either the outsole or the vamp. Most shoes have shoelaces on the upper, connecting the medial and lateral parts after one puts their shoes on and aiding in keeping their shoes on their feet. In 1968, Puma SE introduced the first pair of sneakers with Velcro straps in lieu of shoelaces, and these became popular by the 1980s, especially among children and the elderly. The toe box is the part that covers and protects the toes. People with toe deformities, or individuals who experience toe swelling (such as long-distance runners) usually require a larger toe box. There are a vast variety of different types of shoes. Most types of shoes are designed for specific activities. For example, boots are typically designed for work or heavy outdoor use. Athletic shoes are designed for particular sports such as running, walking, or other sports. Some shoes are designed to be worn at more formal occasions, and others are designed for casual wear. There are also a vast variety of shoes designed for different types of dancing. Orthopedic shoes are special types of footwear designed for individuals with particular foot problems or special needs. Clinicians evaluate patient's footwear as a part of their clinical examination. However, it is often based on each individual's needs, with attention to the choice of footwear worn and if the shoe is adequate for the purpose of completing their activities of daily living. Other animals, such as dogs and horses, may also wear special shoes to protect their feet as well. Depending on the activity for which they are designed, some types of footwear may fit into multiple categories. For example, Cowboy boots are considered boots, but may also be worn in more formal occasions and used as dress shoes. Hiking boots incorporate many of the protective features of boots, but also provide the extra flexibility and comfort of many athletic shoes. Flip-flops are considered casual footwear, but have also been worn in formal occasions, such as visits to the White House. Athletic shoes are specifically designed to be worn for participating in various sports. Since friction between the foot and the ground is an important force in most sports, modern athletic shoes are designed to maximize this force, and materials, such as rubber, are used. Participants in sports in which sliding is desirable, such as dancing or bowling, wear shoes with lower coefficients of friction. The earliest athletic shoes, dating to the mid-19th century, were track spikes—leather shoes with metal cleats on the soles to provide increased friction during running. They were developed by J.W. Foster & Sons, which later become known as Reebok. By the end of the 19th century, Spalding also manufactured these shoes as well. Adidas started selling shoes with track spikes in them for running and soccer in 1925. Spikes were eventually added to shoes for baseball and American football in the 20th century. Golfers also use shoes with small metal spikes on their soles to prevent slipping during their swing. The earliest rubber-soled athletic shoes date back to 1876 in the United Kingdom, when the New Liverpool Rubber Company made plimsolls, or sandshoes, designed for the sport of croquet. Similar rubber-soled shoes were made in 1892 in the United States by Humphrey O'Sullivan, based on Charles Goodyear's technology. The United States Rubber Company was founded the same year and produced rubber-soled and heeled shoes under a variety of brand names, which were later consolidated in 1916 under the name, Keds. These shoes became known as, "sneakers", because the rubber sole allowed the wearer to sneak up on another person. In 1964, the founding of Nike by Phil Knight and Bill Bowerman of the University of Oregon introduced many new improvements common in modern running shoes, such as rubber waffle soles, breathable nylon uppers, and cushioning in the mid-sole and heel. During the 1970s, the expertise of podiatrists also became important in athletic shoe design, to implement new design features based on how feet reacted to specific actions, such as running, jumping, or side-to-side movement. Athletic shoes for women were also designed for their specific physiological differences. Shoes specific to the sport of basketball were developed by Chuck Taylor, and are popularly known as Chuck Taylor All-Stars. These shoes, first sold in 1917, are double-layer canvas shoes with rubber soles and toe caps, and a high heel (known as a "high top") for added support. In 1969, Taylor was inducted into the Naismith Memorial Basketball Hall of Fame in recognition of this development, and in the 1970s, other shoe manufacturers, such as Nike, Adidas, Reebok, and others began imitating this style of athletic shoe. In April 1985, Nike introduced its own brand of basketball shoe which would become popular in its own right, the Air Jordan, named after the then-rookie Chicago Bulls basketball player, Michael Jordan. The Air Jordan line of shoes sold $100 million in their first year. As barefoot running became popular by the late 20th and early 21st century, many modern shoe manufacturers have recently designed footwear that mimic this experience, maintaining optimum flexibility and natural walking while also providing some degree of protection. Termed as Minimalist shoes, their purpose is to allow one's feet and legs to feel more subtly the impacts and forces involved in running, allowing finer adjustments in running style. Some of these shoes include the Vibram FiveFingers, Nike Free, and Saucony's Kinvara and Hattori. Mexican huaraches are also very simple running shoes, similar to the shoes worn by the Tarahumara people of northern Mexico, who are known for their distance running abilities. Wrestling shoes are also very light and flexible shoes that are designed to mimic bare feet while providing additional traction and protection. Many athletic shoes are designed with specific features for specific activities. One of these includes roller skates, which have metal or plastic wheels on the bottom specific for the sport of roller skating. Similarly, ice skates have a metal blade attached to the bottom for locomotion across ice. Skate shoes have also been designed to provide a comfortable, flexible and durable shoe for the sport of skateboarding. Climbing shoes are rubber-soled, tight-fitting shoes designed to fit in the small cracks and crevices for rock climbing. Cycling shoes are similarly designed with rubber soles and a tight fit, but also are equipped with a metal or plastic cleat to interface with clipless pedals, as well as a stiff sole to maximize power transfer and support the foot. Some shoes are made specifically to improve a person's ability to weight train. Sneakers that are a mix between an activity-centered and a more standard design have also been produced: examples include roller shoes, which feature wheels that can be used to roll on hard ground, and Soap shoes, which feature a hard plastic sole that can be used for grinding. A boot is a special type of shoe which covers the foot and the ankle and extends up the leg, sometimes as far as the knee or even the hip. Most boots have a heel that is clearly distinguishable from the rest of the sole, even if the two are made of one piece. They are typically made of leather or rubber, although they may be made from a variety of different materials. Boots are worn both for their functionality—protecting the foot and leg from water, snow, mud or hazards or providing additional ankle support for strenuous activities—as well as for reasons of style and fashion. Cowboy boots are a specific style of riding boots that combine function with fashion. They became popular among cowboys in the western United States during the 19th century. Traditional cowboy boots have a Cuban heel, rounded to pointed toe, high shaft, and, traditionally, no lacing. They are normally made from cowhide leather but may be made from more exotic skins such as ostrich, anaconda, or elephant skins. Hiking boots are designed to provide extra ankle and arch support, as well as extra padding for comfort during hiking. They are constructed to provide comfort for miles of walking over rough terrains, and protect the hiker's feet against water, mud, rocks, and other wilderness obstacles. These boots support the ankle to avoid twisting but do not restrict the ankle's movement too much. They are fairly stiff to support the foot. A properly fitted boot and/or friction-reducing patches applied to troublesome areas ensures protection against blisters and other discomforts associated with long hikes on rugged terrain. During wet or snowy weather, snow boots are worn to keep the foot warm and dry. They are typically made of rubber or other water-resistant material, have multiple layers of insulation, and a high heel to keep snow out. Boots may also be attached to snowshoes to increase the distribution of weight over a larger surface area for walking in snow. Ski boots are a specialized snow boot which are used in alpine or cross-country skiing and designed to provide a way to attach the skier to his/her skis using ski bindings. The ski/boot/binding combination is used to effectively transmit control inputs from the skier's legs to the snow. Ice skates are another specialized boot with a metal blade attached to the bottom which is used to propel the wearer across a sheet of ice. Inline skates are similar to ice skates but with a set of three to four wheels in lieu of the blade, which are designed to mimic ice skating on solid surfaces such as wood or concrete. Boots are designed to withstand heavy wear to protect the wearer and provide good traction. They are generally made from sturdy leather uppers and non-leather outsoles. They may be used for uniforms of the police or military, as well as for protection in industrial settings such as mining and construction. Protective features may include steel-tipped toes and soles or ankle guards. Dress shoes are characterized by smooth and supple leather uppers, leather soles, and narrow sleek figure. Casual shoes are characterized by sturdy leather uppers, non-leather outsoles, and wide profile. Some designs of dress shoes can be worn by either gender. The majority of dress shoes have an upper covering, commonly made of leather, enclosing most of the lower foot, but not covering the ankles. This upper part of the shoe is often made without apertures or openings, but may also be made with openings or even itself consist of a series of straps, e.g. an open toe featured in women's shoes. Shoes with uppers made high to cover the ankles are also available; a shoe with the upper rising above the ankle is usually considered a boot but certain styles may be referred to as high-topped shoes or high-tops. Usually, a high-topped shoe is secured by laces or zippers, although some styles have elastic inserts to ease slipping the shoe on. Men's shoes can be categorized by how they are closed: Men's shoes can also be decorated in various ways: Formal high-end men's shoes are manufactured by several companies around the world, amongst others in Great Britain, France, Hungary, Romania, Croatia, Italy, and to a lesser extent in the United States. Notable British brands include: Church's English Shoes (est. 1873), John Lobb Bootmaker (est. 1849), Edward Green Shoes (est. 1890), and Crockett & Jones (est. 1879). Both John Lobb and Edward Green offer bespoke products. In between the world wars, men's footwear received significant innovation and design, led by cobblers and cordwainers in London's West End. A well-known French maker is J.M. Weston. Armani of Italy was a major influence on men's shoe design in the 1960s–1980s until they returned to the larger proportions of its forebears, the welt-constructed Anglo-American dress shoe originally created in Edwardian England. Another well-known Italian company is Salvatore Ferragamo Italia S.p.A. Higher end companies in the United States are Allen Edmonds and Alden Shoe Company. Alden, located in New England, specializes in genuine shell cordovan leather from the only remaining horse tannery in the US, in Chicago and is completely manufactured domestically, whereas Allen Edmonds of Wisconsin is a larger company that outsources some of its production. There is a large variety of shoes available for women, in addition to most of the men's styles being more accepted as unisex. Some broad categories are: A vast variety of footwear is used by dancers. The choice of dance shoe type depends on the style of dance that is to be performed and, in many cases, the characteristics of the surface that will be danced on. Orthopedic shoes are specially-designed footwear to relieve discomfort associated with many foot and ankle disorders, such as blisters, bunions, calluses and corns, hammer toes, plantar fasciitis, or heel spurs. Certain types of therapeutic footwear are prescribed for children with mobility issues. They may also be worn by individuals with diabetes or people with unequal leg length. These shoes typically have a low heel, tend to be wide with a particularly wide toe box, and have a firm heel to provide extra support. Some may also have a removable insole, or orthotic, to provide extra arch support. The measure of a foot for a shoe is from the heel to the longest toe. Shoe size is an alphanumerical indication of the fitting size of a shoe for a person. Often it just consists of a number indicating the length because many shoemakers only provide a standard width for economic reasons. Globally, several different shoe-size systems are used, differing in their units of measurement and in the position of sizes 0 and 1. Only a few systems also take the width of the feet into account. Some regions use different shoe-size systems for different types of shoes (e.g., men's, women's, children's, sport, or safety shoes). Units for shoe sizes vary vastly around the world. European sizes are measured in Paris Points, each measuring two-thirds of a centimeter. The UK and American units result in whole-number sizes spaced at one barleycorn (1⁄3 inch), with UK adult sizes starting at size 1 = 8+2⁄3 in (22.0 cm). In the US, this is size 2. Men's and women's shoe sizes often have different scales. Shoe size is often measured using a Brannock Device, which can determine both the width and length size values of the foot. A metric standard for shoe sizing, the Mondopoint system, was introduced in the 1970s by International Standard ISO 2816:1973 "Fundamental characteristics of a system of shoe sizing to be known as Mondopoint" and ISO 3355:1975 "Shoe sizes – System of length grading (for use in the Mondopoint system)". the current version of the standard is ISO 9407:2019, "Shoe sizes—Mondopoint system of sizing and marking". The Mondopoint system includes measurements of both length and width of the foot. In many places in the world, shoes are removed when moving from exteriors to interiors, particularly in homes and religious buildings. In many Asian countries, outdoor shoes are exchanged for indoor shoes or slippers. Some fitness centres require that shoes be exchanged for indoor shoes to prevent dirt and grime from being transferred to the equipment.
{{No sources}}[[File:Reebok Royal Glide Ripple Clip shoe.jpg|thumb|[[Sport shoes]]]] '''Footwear''' is an item of [[clothing]] made by humans that covers and protects the [[foot]], including the [[sole]]s of the feet. Footwear allows people to walk on rough surfaces such as [[gravel]] [[road]]s without hurting their feet. Some types of footwear such as [[boot]]s help to keep people's feet dry, or help to keep people's feet warm in [[arctic|cold]] weather. ==Types of Footwear== *[[Shoe]]s are a type of footwear that protects the foot and the [[sole]]s of the feet. Shoes are made of many different types of materials, such as [[leather]], [[fabric]], and [[plastic]]. There are several types of shoes: **'''Athletic shoes''' (also called "running shoes") are made of fabric, rubber, and plastic. They are made for people to wear while they are doing [[sports]] or [[recreation]]al activities. Athletic shoes like spikes, give grip to sprinting on the athletic track. **'''Dress shoes''' are made for people to wear when they are wearing [[formal]] clothes such as a [[suit and tie]] or a [[tuxedo]]. Dress shoes are normally made from dark-colored leather that is shined using shoe polish. People working in professions such as [[business]], [[lawyer|law]], and [[politician|politics]] wear dress shoes. People working in [[restaurant]]s or [[retail store]]s may also wear dress shoes. **'''Work shoes''' are designed for people who work in construction or factory jobs. Work shoes often have a steel toe cover to protect the person's toes. Work shoes are usually made of strong leather, to protect the person's foot from sharp objects or dangerous chemicals. [[File:Chalcolithic leather shoe from Areni-1 cave.jpg|thumb|The [[Areni-1 shoe]] is a 5,500-year-old leather shoe that was found in 2008 in excellent condition in the Areni-1 cave located in the Vayots Dzor province of [[Armenia]].]] *[[Sandal]]s are a type of footwear that consists of a covering for the sole of the feet and straps or ties which hold the sandal onto the foot. Sandals are worn in warm countries, such as the [[Philippines]] and [[Cuba]]. Sandals are also worn indoors in some places, such as in [[sauna]]s. * '''[[Boot]]s''' are a type of footwear that protects the foot and [[ankle]]. Boots are higher and larger than [[shoe]]s and [[sandal]]s. Some boots are high enough to protect the [[leg|calves]] (lower part of the leg) as well. **'''Rain boots''' (or rubber boots) are made from [[rubber]] or [[plastic]]. Rain boots protect a person's feet from [[water]] and [[rain]]. People who work on [[fishing]] boats and [[farmer]]s wear rubber boots to keep their feet dry. People who work in [[chemical industry|chemical]] factories wear rubber boots to protect their feet from dangerous chemicals. **'''Winter boots''' are boots that keep a person's feet warm in cold weather. People in cold countries such as [[Canada]] and [[Sweden]] wear winter boots during the cold season. Winter boots can be made from many different materials, such as [[leather]], [[fabric]], or [[plastic]]. Winter boots are [[Thermal insulation|insulated]] with wool or fur to keep the feet warm. Most winter boots also keep people's feet dry. **'''Work boots''' (or "[[construction]] boots") are designed for people who work in construction or [[factory]] jobs. Work boots often have a steel toe cover to protect the person's toes. Work boots are usually made of strong leather, to protect the person's foot from sharp objects or dangerous chemicals. Some work boots have a flat piece of [[steel]] in the sole to protect the foot from sharp [[nail]]s. Many countries require construction workers to wear work boots when they are on a construction site. *[[Sock]]s are an article of [[clothing]] for the foot, [[ankle]], and calf (the lower part of the [[leg]], below the [[knee]]). Socks absorb [[perspiration]] ([[sweat]]) from the feet. Socks also protect the foot from chafing (rubbing) against the shoe or boot, which can cause [[blister]]s. Both men and women wear socks. Socks are usually made from [[cotton]] or synthetic fabrics. There are several types of socks: **'''Athletic socks''' are made of cotton or [[synthetic fabric]]. They are made for people to wear while they are doing [[sports]] or [[recreation]]al activities. **'''Dress socks''' are made for people to wear when they are wearing dress shoes and [[formal]] clothes such as a [[suit and tie]] or a [[tuxedo]]. Dress socks are normally made from dark-colored fabric (often black or dark blue-colored). People working in professions such as [[business]], [[lawyer|law]], and [[politician|politics]] wear dress socks. People working in [[restaurant]]s or [[retail store]]s may also wear dress socks. **'''Winter socks''' are designed to keep a person's feet warm in cold weather. People in cold countries such as [[Canada]] and [[Sweden]] wear winter socks during the cold season. Winter socks can be made from many different materials, but most winter socks are made with [[wool]]. Wool is a warm fabric which helps to [[Thermal insulation|insulate]] the feet and keep them warm. **There are also foot and leg coverings for [[women]] which are related to socks. These articles of clothing, which are made from [[polyester]] or other synthetic materials, are called [[pantyhose]], [[nylons]], or [[tights]]. ==How footwear is made == ===By hand=== People in many countries make their own footwear by hand, using simple [[tool]]s. A simple pair of sandals can be made by hand [[knife|cutting]] a foot-shaped sole out of a thick, flexible material such as rubber. Next, straps of fabric, rope or leather can be added with a [[needle]] and [[thread]]. A simple pair of boots can be made by hand by using animal hide with fur, and sewing it with strong thread. ===In a factory=== Many people wear footwear that is made in a [[factory]]. The machines in shoe factories and boot factories can make footwear much more quickly than people who are making footwear by hand with a needle and thread. {{commonscat|Footwear}} [[Category:Footwear| ]]
Footwear refers to garments worn on the feet, which typically serve the purpose of protection against adversities of the environment such as wear from rough ground; stability on slippery ground; and temperature. Cultures have different customs regarding footwear. These include not using any in some situations, usually bearing a symbolic meaning. This can however also be imposed on specific individuals to place them at a practical disadvantage against shod people, if they are excluded from having footwear available or are prohibited from using any. This usually takes place in situations of captivity, such as imprisonment or slavery, where the groups are among other things distinctly divided by whether or not footwear is being worn. In some cultures, people remove their shoes before entering a home. Bare feet are also seen as a sign of humility and respect, and adherents of many religions worship or mourn while barefoot. Some religious communities explicitly require people to remove shoes before they enter holy buildings, such as temples. In several cultures people remove their shoes as a sign of respect towards someone of higher standing. Similarly, deliberately forcing other people to go barefoot while being shod oneself has been used to clearly showcase and convey one's superiority within a setting of power disparity. Practitioners of the craft of shoemaking are called shoemakers, cobblers, or cordwainers. Footwear has been used by humans since prehistoric times, with paleoclimatology suggesting that they would have been needed in some areas of human settlement by at least 50,000 years ago during the Last Glacial Period. Osteologists have found evidence of the effect of footwear on human remains by around 40,000 years ago. The oldest shoes so far recovered were found by a team under Luther Cressman in Fort Rock Cave, Oregon, US, in 1938. They had been preserved under the Mazama Ash deposited c. 5025 BC during the volcanic eruption that formed Crater Lake. In 1999, they were dated to around 10,500–9,300 BP. Egyptian butchers sometimes wore platform sandals with thicker soles than usual to raise their feet out of the gore. Wealthier Egyptians also sometimes wore platforms. The Greeks distinguished a great variety of footwear, particularly different styles of sandals. The heeled cothurnus was part of the standard costume for tragedians, and the effeminate soccus for comedians. Going barefoot, however, was frequently lauded: Spartan boys undergoing military training, Socrates, and Olympic athletes all went without shoes most of the time. Similarly, ancient China considered footwear an important aspect of civilization—particularly embroidered slippers—but often depicted Taoist immortals and gods like Xuanwu barefoot. The Book of Exodus records Moses reverentially removing his shoes at Mount Sinai and the priests likewise went barefoot at the Temple of Solomon before Babylonian customs prevailed and entering houses of worship in footwear became common in Judaism and Christianity. The Etruscans experienced several footwear trends, including the prominently pointed shoe or boot now known as the calceus repandus. The Romans saw clothing and footwear as unmistakable signs of power and status in society. Patricians typically wore dyed and ornamented shoes of tanned leather with their togas or armor, while plebeians wore rawhide or hobnail boots and slaves were usually required to be barefoot. These class distinctions in footwear seem to have lessened during the imperial period, however, as the emperors appropriated more and more symbols of high status for themselves. The Romans were the earliest people currently known to have shaped their right and left shoes distinctly during creation, rather than pulling them tight and allowing them to wear into shape. The Catholic patron saints of shoemaking—Crispin and Crispinian—were martyred during the Diocletianic Persecution. In medieval Europe, leather shoes and boots became more common. At first most were simply pieces of leather sewn together and then held tight around the foot with a toggle or drawstring. This developed into the turnshoe, where the sole and upper were sewn together and then turned inside-out to hide and protect the seam and improve water resistance. From the reign of Charlemagne, Byzantine fashions began to influence the west and the pontificalia of the popes and other bishops began to feature greater luxury, including embroidered silk and velvet slippers. By the High Middle Ages, fashion trends periodically prompted sumptuary taxes or regulations and church condemnation for vanity. The 12th-century pigache and 14th- and 15th-century poulaine had elongated toes, often stuffed to maintain their shape. Around the same time, several mendicant orders began practicing discalceation as an aspect of their vows of humility and poverty, going entirely barefoot at all times or only wearing sandals in any weather. From the 1480s, the poulaine was replaced by the duckbill, which had a flat front but soon became impractically wide. The stiff hose of the era usually required fairly soft footwear, which in turn was easier to damage in the dirt and muck of the street and outdoors. This led many people to use wooden-soled calopedes, pattens, or galoshes, overshoes that served as a platform while walking. Particularly in Venice, these platforms were combined with the shoe to make chopines, sometimes so awkwardly high that the wearer required servants to help support them. (Turkish sources, meanwhile, credit the chopines directly to the nalins worn in Ottoman baths and whose height was considered to be a marker of status.) By the early modern period, the development of better socks and less stiff hose allowed European footwear to become firmer and more durable. Welting was developed, using a narrow band of leather between the uppers and sole to improve appearance and comfort, increase water resistance, and simplify repair, particularly resoling worn shoes. Beginning with the 1533 marriage of the 14-year-old Florentine Catherine de Medici to Prince Henry of France, both male and female royalty and nobles began wearing high heels, giving rise to the expression "well heeled". This was done sometimes for display or appearance and sometimes as an aid to riding in stirrups. For the most part, male footwear was more ornate and expensive because women's feet were usually covered by the large dresses of the era. Shoe fetishism was first publicized in the work of Nicolas-Edme Rétif in prerevolutionary France. 17th-century Cavalier boots developed into upper-class fashion and into sailing boots prized by fishermen and pirates before being replaced as military gear by the 18th-century Hessian and 19th-century Wellington boot. In Ming and Qing China, foot binding led to the development of lotus shoes for Han women and then flowerpot shoes for the Manchu women who wanted to emulate the characteristic walk of women with bound feet without undergoing the process themselves. In Africa, North America, and Spanish and Portuguese South America, slave codes often mandated slaves should be barefoot at all times without exception. Following its independence, the American South was an exception. Its demand for masses of low-quality shoes for its slaves was met by workshops around Boston, Philadelphia, and New York, a dependence that later hobbled the Confederate Army during the Civil War and was responsible in legend for the decisive Battle of Gettysburg. Amid the Industrial Revolution, John Adam Dagyr's introduction of assembly line production and tight quality control to the "ten-footer" workshops in Lynn, Massachusetts, US, around 1760 is sometimes credited as the first shoe factory. However, although mechanized textile mills greatly reduced the price of proper socks, each step of the shoemaking process still needed to be done by hand in a slowly optimized putting-out system. The first mechanized systems—developed by Marc Isambard Brunel in 1810 to supply boots to the British Army amid the Napoleonic Wars—failed commercially as soon as the wars were over because the demobilized soldiers reduced the price of manual labor. John Brooks Nichols's 1850 adaptation of Howe and Singer's sewing machines to handle binding uppers to soles and the Surinamese immigrant Jan Ernst Matzeliger's 1880 invention of an automatic lasting machine finally allowed true industrialization, taking the productivity of individual workers from 20 or 50 pairs a day to as many as 700, halving prices, and briefly making Lynn the center of world shoe production. As late as 1865, most men in the industry identified in the census and city directory as general purpose "cordwainers" or "shoemakers"; by 1890, they were almost universally described as "shoe workers" or—more often—by the specific name of their work within the industry: "edgesetter", "heel trimmer", "McKay machine operator". Many were replaced by cheaper immigrants; the Czech Tomáš Baťa joined these workers at Lynn in 1904 and then returned to his own factory in Zlín, Moravia, mechanizing and rationalizing its production while guiding the factory town that developed into a garden city. By the early 20th century, vulcanization had led to the development of plimsolls, deck shoes, rubber boots, galoshes, and waders. The prevalence of trench foot in World War I focused attention on the importance of providing of adequate footwear in following conflicts, although this was not always possible. Millions of Chinese soldiers in both the NRA and PLA were obliged to use straw and rope shoes to allow easy replacement on long marches during both World War II and the following civil war, contributing to disease and desertion, particularly among the Nationalists. Following the world wars, the increasing importance of professional sports greatly popularized a variety of athletic shoes, particularly sneakers. Major brands such as Converse, Adidas, and Nike used celebrity endorsements from Chuck Taylor, Michael Jordan, Lionel Messi, and others to promote their products. Fashion houses periodically prompted new trends in women's and high-end fashion. In particular, while working for Christian Dior, Roger Vivier popularized the stiletto heel in 1954. (Men's dress shoes have tended to retain 19th-century British looks such as the Oxford shoe and loafers.) Various subcultures have employed distinctive footwear as part of their identity, including winklepickers, Doc Martens, and skate shoes. The international trade in footwear was at first chiefly restricted to American exports to Europe and Europe's exports to its various colonial empires. Assisted by the Marshall Plan after World War II, Italy became the major shoe exporting country in the 1950s. It was joined in the 1960s by Japan, which offshored its production to Taiwan, South Korea, and Hong Kong as its own labor became too expensive. In their turn, the Hong Kong manufacturers began moving production to Guangdong in mainland China almost immediately after the establishment of Deng Xiaoping's Opening Up Policy in the early 1980s. Competitors were soon forced to follow suit, including removal of Taiwanese and Korean production to Fujian and to Wenzhou in southern Zhejiang. Similarly, amid Perestroika and the Fall of Communism, Italy dismantled its domestic industry, outsourcing its work to Eastern Europe, which proved less dependable than the Chinese and further eroded their market share. Beginning around the year 2000, China has constantly produced more than half of the world's shoes. As of 2021, footwear is the 30th most traded category internationally; but, while China produces well over 60% of exported footwear, it currently earns less than 36% of the value of the total trade owing to the continuing importance of American, German, and other brands in the North American and European markets. Modern footwear is usually made of leather or plastic, and rubber. In fact, leather was one of the original materials used for the first versions of a shoe. The soles can be made of rubber or plastic, sometimes with the addition of a sheet of metal on the inside. Roman sandals had sheets of metal on their soles so that they would not bend out of shape. In more recent times, footwear suppliers such as Nike have begun to source environmentally friendly materials. In Europe, recent decades have seen a decline in the footwear industry. While about 27,000 firms were in business in 2005, only 21,700 remained in 2009. Not only have these firms decreased in number, but direct employment has also reduced within the sector. In the U.S., the annual footwear industry revenue was $48 billion in 2012. In 2015, there were about 29,000 shoe stores in the U.S. and the shoe industry employed about 189,000 people. Due to rising imports, these numbers are also declining. The only way of staying afloat in the shoe market is to establish a presence in niche markets. To ensure high quality and safety of footwear, manufacturers have to make sure all products comply to existing and relevant standards. By producing footwear in accordance with national and international regulations, potential risks can be minimized and the interest of both textile manufacturers and consumers can be protected. The following standards/regulations apply to footwear products: Footwear can create two types of impressions: two-dimensional and three-dimensional impressions. When footwear places material onto a solid surface, it creates a two-dimensional impression. These types of impressions can be made with a variety of substances, like dirt and sand. When footwear removes material from a soft surface, it creates a three-dimensional impression. These types of impressions can be made in a variety of soft substances, like snow and dirt. Two-dimensional impressions also differ from three-dimensional impressions because the latter demonstrate length, width, and depth whereas two-dimensional impressions only demonstrate the first two aspects. Media related to Footwear at Wikimedia Commons
[[File:CountBasieEthelWatersStageDoorCanteen.jpg|thumb|250px|[[Count Basie]] with his [[orchestra]] and [[vocalist]] [[Ethel Waters]], in the movie ''Stage Door Canteen'', 1943]] {{Infobox music genre | Name = Jazz | color = black | bgcolor = #FFC0CB | image = |caption = | stylistic_origins = {{hlist|[[Blues]]|[[Folk music|folk]]|[[American march music|marches]]|[[ragtime]]|[[classical music]]}} | cultural_origins = Late 19th-century United States | instruments = {{hlist|[[Double bass]]|[[Drum kit|drums]]|[[guitar]] (typically [[electric guitar]])|[[piano]]|[[saxophone]]|[[trumpet]]|[[clarinet]]|[[trombone]]|[[Singing|vocals]]|[[vibraphone]]|[[Hammond organ]]|[[harmonica]]}}. In [[jazz fusion]] of the 1970s, [[electric bass]], [[electric piano]] and [[synthesizer]] were common. | derivatives = {{hlist|[[Funk]]|[[jump blues]]|[[reggae]]|[[rhythm and blues]]|[[ska]]|[[Hip hop music|hip hop]]}} | subgenrelist = | subgenres = {{hlist|[[Bebop]]|[[Big band]]|[[Free jazz]]|[[Gypsy jazz]]|[[Hard bop]]|[[Latin jazz]]|[[Mainstream jazz]]|[[Soul jazz]]|[[Stride (music)|Stride piano]]|[[Swing music|Swing]]|[[Traditional jazz]]}}}} '''Jazz''' is a type of [[music]] which was invented in the [[United States]]. Jazz music combines [[African-American people|African-American]] music with [[Europe]]an music. Jazz first became popular in the 1910s. Some common jazz instruments include the [[saxophone]], [[trumpet]], [[guitar]], [[Jazz piano|piano]], [[double bass]], and [[drum]]s. ==Definition== It is difficult to give an exact definition for "jazz".<ref>{{cite book|url=https://books.google.com/books?id=kfFgNABSuuUC|title=The History and Tradition of Jazz|author=Larson, Thomas|date=2002|publisher=Kendall Hunt|isbn=9780787275747}}</ref> A singer Nina Simone said, "Jazz is not just music, it is a way of life, it is a way of being, a way of thinking".<ref name=":0" /> But when we talk about jazz as music, one important part of jazz is [[improvisation]] (improv), which means the person playing is making music up as they go along. If a jazz band is playing a song, the song may have several solos where one player will improvise while the rest of the band, except for the rhythm section (such as the piano, bass, or drums), does not play. Most jazz is very rhythmic, which is called "swing," and uses "blue" notes.<ref>{{Cite web|date=2015-03-09|title=What is Jazz?|url=https://americanhistory.si.edu/smithsonian-jazz/education/what-jazz|access-date=2020-08-11|website=National Museum of American History|language=en}}</ref> Blue notes are notes that are played at a slightly different pitch from normal. == History == [[File:BennyGoodmanStageDoorCanteen.jpg|thumb|left|[[Benny Goodman]], one of the first swing [[big band]] leaders to become widely popular.]] Jazz began in the United States in the early 20th century. Jazz music was first based on the music of African slaves who were forced to work in the plantations of the [[southern United States]]. This included [[call and response (music)|call and response]] songs, spirituals, [[chant]]s and blue notes. These characteristics are what developed [[blues]], a sad song that slaves sung during their labor. These influences were indirect, through earlier musical forms such as [[ragtime]]. Jazz also has musical styles from European music, as well as the brass and stringed instruments and (sometimes) the use of [[musical notation]]. There have been different types of jazz through time. [[New Orleans]] jazz began in the early 1910s. Dixieland jazz was also popular. In the 1930s, there was [[swing jazz]], which was also called big band jazz. In the 1940s, [[bebop]] became a major type of jazz, with fast songs and [[wikt:hard|complex]] [[harmony]]. Large jazz bands, which are called [[big band]]s, were also popular in the 1940s. Big bands usually have 5 saxophone players, 4 or 5 trumpet players, 4 [[trombone]] players, a [[piano]] player or [[guitar]] player, an acoustic bass player, a drummer, and sometimes a singer. In the 1950s, there was [[hard bop]] jazz. In the 1960s, there was [[modern jazz]] and [[free jazz]]. In the 1970s, [[jazz fusion]] began to blend jazz music with [[Rock and roll|rock]] music. Some jazz is still played with the same improv methods as it did at its beginning, except with modern [[Synthesizer|electronic instruments]]. == International Jazz Day == April 30 is an "International Jazz Day".<ref>{{Cite web|title=VOA Observes Jazz Appreciation Month {{!}} Voice of America - English|url=https://www.voanews.com/episode/voa-observes-jazz-appreciation-month-3750766|access-date=2020-08-11|website=www.voanews.com|language=en}}</ref> Each year jazz concerts and educational events take place around the world. This is the day to increase understanding to jazz. The United Nations Education and Science Organization (UNESCO) also take part in this event.<ref name=":0">{{Cite web|date=2020-06-08|title=International Jazz Day|url=https://en.unesco.org/commemorations/jazzday|access-date=2020-08-11|website=UNESCO|language=en}}</ref> People celebrate the day, because they believe the history of jazz has something to do with peace, freedom and equality.<ref name=":0" /><ref>{{Cite web|last=Nations|first=United|title=International Jazz Day|url=https://www.un.org/en/observances/jazz-day|access-date=2020-08-11|website=United Nations|language=en}}</ref> ==References== {{reflist}} == Other websites == * [https://www.pbs.org/kenburns/jazz/home/ ''Jazz''] — A documentary movie by [[Ken Burns]] on [[PBS]] * [https://archive.org/details/Free_20s_Jazz_Collection Free 1920s Jazz Collection] available for downloading at [[Archive.org]] * [http://www.jazzfoundation.org/ Jazz Foundation of America] * [http://www.smithsonianjazz.org/ Jazz at the Smithsonian Museum] * [http://www.jazzhall.com/ Alabama Jazz Hall of Fame website] * [https://americanhistory.si.edu/smithsonian-jazz/collections-and-archives/smithsonian-jazz-oral-history-program Smithsonian Jazz Oral History Program | National Museum of American History] {{authority control}} [[Category:Jazz music| ]] [[Category:African-American history]]
Jazz is a music genre that originated in the African-American communities of New Orleans, Louisiana, in the late 19th and early 20th centuries, with its roots in blues and ragtime. Since the 1920s Jazz Age, it has been recognized as a major form of musical expression in traditional and popular music. Jazz is characterized by swing and blue notes, complex chords, call and response vocals, polyrhythms and improvisation. Jazz has roots in European harmony and African rhythmic rituals. As jazz spread around the world, it drew on national, regional, and local musical cultures, which gave rise to different styles. New Orleans jazz began in the early 1910s, combining earlier brass band marches, French quadrilles, biguine, ragtime and blues with collective polyphonic improvisation. But jazz did not begin as a single musical tradition in New Orleans or elsewhere. In the 1930s, arranged dance-oriented swing big bands, Kansas City jazz (a hard-swinging, bluesy, improvisational style), and gypsy jazz (a style that emphasized musette waltzes) were the prominent styles. Bebop emerged in the 1940s, shifting jazz from danceable popular music toward a more challenging "musician's music" which was played at faster tempos and used more chord-based improvisation. Cool jazz developed near the end of the 1940s, introducing calmer, smoother sounds and long, linear melodic lines. The mid-1950s saw the emergence of hard bop, which introduced influences from rhythm and blues, gospel, and blues to small groups and particularly to saxophone and piano. Modal jazz developed in the late 1950s, using the mode, or musical scale, as the basis of musical structure and improvisation, as did free jazz, which explored playing without regular meter, beat and formal structures. Jazz-rock fusion appeared in the late 1960s and early 1970s, combining jazz improvisation with rock music's rhythms, electric instruments, and highly amplified stage sound. In the early 1980s, a commercial form of jazz fusion called smooth jazz became successful, garnering significant radio airplay. Other styles and genres abound in the 21st century, such as Latin and Afro-Cuban jazz. The origin of the word jazz has resulted in considerable research, and its history is well documented. It is believed to be related to jasm, a slang term dating back to 1860 meaning "pep, energy". The earliest written record of the word is in a 1912 article in the Los Angeles Times in which a minor league baseball pitcher described a pitch which he called a 'jazz ball' "because it wobbles and you simply can't do anything with it". The use of the word in a musical context was documented as early as 1915 in the Chicago Daily Tribune. Its first documented use in a musical context in New Orleans was in a November 14, 1916, Times-Picayune article about "jas bands". In an interview with National Public Radio, musician Eubie Blake offered his recollections of the slang connotations of the term, saying: "When Broadway picked it up, they called it 'J-A-Z-Z'. It wasn't called that. It was spelled 'J-A-S-S'. That was dirty, and if you knew what it was, you wouldn't say it in front of ladies." The American Dialect Society named it the Word of the 20th Century. Jazz is difficult to define because it encompasses a wide range of music spanning a period of over 100 years, from ragtime to rock-infused fusion. Attempts have been made to define jazz from the perspective of other musical traditions, such as European music history or African music. But critic Joachim-Ernst Berendt argues that its terms of reference and its definition should be broader, defining jazz as a "form of art music which originated in the United States through the confrontation of the Negro with European music" and arguing that it differs from European music in that jazz has a "special relationship to time defined as 'swing'". Jazz involves "a spontaneity and vitality of musical production in which improvisation plays a role" and contains a "sonority and manner of phrasing which mirror the individuality of the performing jazz musician". A broader definition that encompasses different eras of jazz has been proposed by Travis Jackson: "it is music that includes qualities such as swing, improvising, group interaction, developing an 'individual voice', and being open to different musical possibilities". Krin Gibbard argued that "jazz is a construct" which designates "a number of musics with enough in common to be understood as part of a coherent tradition". Duke Ellington, one of jazz's most famous figures, said, "It's all music." Although jazz is considered difficult to define, in part because it contains many subgenres, improvisation is one of its defining elements. The centrality of improvisation is attributed to the influence of earlier forms of music such as blues, a form of folk music which arose in part from the work songs and field hollers of African-American slaves on plantations. These work songs were commonly structured around a repetitive call-and-response pattern, but early blues was also improvisational. Classical music performance is evaluated more by its fidelity to the musical score, with less attention given to interpretation, ornamentation, and accompaniment. The classical performer's goal is to play the composition as it was written. In contrast, jazz is often characterized by the product of interaction and collaboration, placing less value on the contribution of the composer, if there is one, and more on the performer. The jazz performer interprets a tune in individual ways, never playing the same composition twice. Depending on the performer's mood, experience, and interaction with band members or audience members, the performer may change melodies, harmonies, and time signatures. In early Dixieland, a.k.a. New Orleans jazz, performers took turns playing melodies and improvising countermelodies. In the swing era of the 1920s–'40s, big bands relied more on arrangements which were written or learned by ear and memorized. Soloists improvised within these arrangements. In the bebop era of the 1940s, big bands gave way to small groups and minimal arrangements in which the melody was stated briefly at the beginning and most of the piece was improvised. Modal jazz abandoned chord progressions to allow musicians to improvise even more. In many forms of jazz, a soloist is supported by a rhythm section of one or more chordal instruments (piano, guitar), double bass, and drums. The rhythm section plays chords and rhythms that outline the composition structure and complement the soloist. In avant-garde and free jazz, the separation of soloist and band is reduced, and there is license, or even a requirement, for the abandoning of chords, scales, and meters. Since the emergence of bebop, forms of jazz that are commercially oriented or influenced by popular music have been criticized. According to Bruce Johnson, there has always been a "tension between jazz as a commercial music and an art form". Regarding the Dixieland jazz revival of the 1940s, Black musicians rejected it as being shallow nostalgia entertainment for white audiences. On the other hand, traditional jazz enthusiasts have dismissed bebop, free jazz, and jazz fusion as forms of debasement and betrayal. An alternative view is that jazz can absorb and transform diverse musical styles. By avoiding the creation of norms, jazz allows avant-garde styles to emerge. For some African Americans, jazz has drawn attention to African-American contributions to culture and history. For others, jazz is a reminder of "an oppressive and racist society and restrictions on their artistic visions". Amiri Baraka argues that there is a "white jazz" genre that expresses whiteness. White jazz musicians appeared in the Midwest and in other areas throughout the U.S. Papa Jack Laine, who ran the Reliance band in New Orleans in the 1910s, was called "the father of white jazz". The Original Dixieland Jazz Band, whose members were white, were the first jazz group to record, and Bix Beiderbecke was one of the most prominent jazz soloists of the 1920s. The Chicago Style was developed by white musicians such as Eddie Condon, Bud Freeman, Jimmy McPartland, and Dave Tough. Others from Chicago such as Benny Goodman and Gene Krupa became leading members of swing during the 1930s. Many bands included both Black and white musicians. These musicians helped change attitudes toward race in the U.S. Female jazz performers and composers have contributed to jazz throughout its history. Although Betty Carter, Ella Fitzgerald, Adelaide Hall, Billie Holiday, Peggy Lee, Abbey Lincoln, Anita O'Day, Dinah Washington, and Ethel Waters were recognized for their vocal talent, less familiar were bandleaders, composers, and instrumentalists such as pianist Lil Hardin Armstrong, trumpeter Valaida Snow, and songwriters Irene Higginbotham and Dorothy Fields. Women began playing instruments in jazz in the early 1920s, drawing particular recognition on piano. When male jazz musicians were drafted during World War II, many all-female bands replaced them. The International Sweethearts of Rhythm, which was founded in 1937, was a popular band that became the first all-female integrated band in the U.S. and the first to travel with the USO, touring Europe in 1945. Women were members of the big bands of Woody Herman and Gerald Wilson. Beginning in the 1950s, many women jazz instrumentalists were prominent, some sustaining long careers. Some of the most distinctive improvisers, composers, and bandleaders in jazz have been women. Trombonist Melba Liston is acknowledged as the first female horn player to work in major bands and to make a real impact on jazz, not only as a musician but also as a respected composer and arranger, particularly through her collaborations with Randy Weston from the late 1950s into the 1990s. Jewish Americans played a significant role in jazz. As jazz spread, it developed to encompass many different cultures, and the work of Jewish composers in Tin Pan Alley helped shape the many different sounds that jazz came to incorporate. Jewish Americans were able to thrive in Jazz because of the probationary whiteness that they were allotted at the time. George Bornstein wrote that African Americans were sympathetic to the plight of the Jewish American and vice versa. As disenfranchised minorities themselves, Jewish composers of popular music saw themselves as natural allies with African Americans. The Jazz Singer with Al Jolson is one example of how Jewish Americans were able to bring jazz, music that African Americans developed, into popular culture. Benny Goodman was a vital Jewish American to the progression of Jazz. Goodman was the leader of a racially integrated band named King of Swing. His jazz concert in the Carnegie Hall in 1938 was the first ever to be played there. The concert was described by Bruce Eder as "the single most important jazz or popular music concert in history". Shep Fields also helped to popularize "Sweet" Jazz music through his appearances and Big band remote broadcasts from such landmark venues as Chicago's Palmer House, Broadway's Paramount Theater and the Starlight Roof at the famed Waldorf-Astoria Hotel. He entertained audiences with a light elegant musical style which remained popular with audiences for nearly three decades from the 1930s until the late 1950s. Jazz originated in the late-19th to early-20th century. It developed out of many forms of music, including blues, spirituals, hymns, marches, vaudeville song, ragtime, and dance music. It also incorporated interpretations of American and European classical music, entwined with African and slave folk songs and the influences of West African culture. Its composition and style have changed many times throughout the years with each performer's personal interpretation and improvisation, which is also one of the greatest appeals of the genre. By the 18th century, slaves in the New Orleans area gathered socially at a special market, in an area which later became known as Congo Square, famous for its African dances. By 1866, the Atlantic slave trade had brought nearly 400,000 Africans to North America. The slaves came largely from West Africa and the greater Congo River basin and brought strong musical traditions with them. The African traditions primarily use a single-line melody and call-and-response pattern, and the rhythms have a counter-metric structure and reflect African speech patterns. An 1885 account says that they were making strange music (Creole) on an equally strange variety of 'instruments'—washboards, washtubs, jugs, boxes beaten with sticks or bones and a drum made by stretching skin over a flour-barrel. Lavish festivals with African-based dances to drums were organized on Sundays at Place Congo, or Congo Square, in New Orleans until 1843. There are historical accounts of other music and dance gatherings elsewhere in the southern United States. Robert Palmer said of percussive slave music: Usually such music was associated with annual festivals, when the year's crop was harvested and several days were set aside for celebration. As late as 1861, a traveler in North Carolina saw dancers dressed in costumes that included horned headdresses and cow tails and heard music provided by a sheepskin-covered "gumbo box", apparently a frame drum; triangles and jawbones furnished the auxiliary percussion. There are quite a few [accounts] from the southeastern states and Louisiana dating from the period 1820–1850. Some of the earliest [Mississippi] Delta settlers came from the vicinity of New Orleans, where drumming was never actively discouraged for very long and homemade drums were used to accompany public dancing until the outbreak of the Civil War. Another influence came from the harmonic style of hymns of the church, which black slaves had learned and incorporated into their own music as spirituals. The origins of the blues are undocumented, though they can be seen as the secular counterpart of the spirituals. However, as Gerhard Kubik points out, whereas the spirituals are homophonic, rural blues and early jazz "was largely based on concepts of heterophony". During the early 19th century an increasing number of black musicians learned to play European instruments, particularly the violin, which they used to parody European dance music in their own cakewalk dances. In turn, European American minstrel show performers in blackface popularized the music internationally, combining syncopation with European harmonic accompaniment. In the mid-1800s the white New Orleans composer Louis Moreau Gottschalk adapted slave rhythms and melodies from Cuba and other Caribbean islands into piano salon music. New Orleans was the main nexus between the Afro-Caribbean and African American cultures. The Black Codes outlawed drumming by slaves, which meant that African drumming traditions were not preserved in North America, unlike in Cuba, Haiti, and elsewhere in the Caribbean. African-based rhythmic patterns were retained in the United States in large part through "body rhythms" such as stomping, clapping, and patting juba dancing. In the opinion of jazz historian Ernest Borneman, what preceded New Orleans jazz before 1890 was "Afro-Latin music", similar to what was played in the Caribbean at the time. A three-stroke pattern known in Cuban music as tresillo is a fundamental rhythmic figure heard in many different slave musics of the Caribbean, as well as the Afro-Caribbean folk dances performed in New Orleans Congo Square and Gottschalk's compositions (for example "Souvenirs From Havana" (1859)). Tresillo (shown below) is the most basic and most prevalent duple-pulse rhythmic cell in sub-Saharan African music traditions and the music of the African Diaspora. Tresillo is heard prominently in New Orleans second line music and in other forms of popular music from that city from the turn of the 20th century to present. "By and large the simpler African rhythmic patterns survived in jazz ... because they could be adapted more readily to European rhythmic conceptions," jazz historian Gunther Schuller observed. "Some survived, others were discarded as the Europeanization progressed." In the post-Civil War period (after 1865), African Americans were able to obtain surplus military bass drums, snare drums and fifes, and an original African-American drum and fife music emerged, featuring tresillo and related syncopated rhythmic figures. This was a drumming tradition that was distinct from its Caribbean counterparts, expressing a uniquely African-American sensibility. "The snare and bass drummers played syncopated cross-rhythms," observed the writer Robert Palmer, speculating that "this tradition must have dated back to the latter half of the nineteenth century, and it could have not have developed in the first place if there hadn't been a reservoir of polyrhythmic sophistication in the culture it nurtured." African-American music began incorporating Afro-Cuban rhythmic motifs in the 19th century when the habanera (Cuban contradanza) gained international popularity. Musicians from Havana and New Orleans would take the twice-daily ferry between both cities to perform, and the habanera quickly took root in the musically fertile Crescent City. John Storm Roberts states that the musical genre habanera "reached the U.S. twenty years before the first rag was published." For the more than quarter-century in which the cakewalk, ragtime, and proto-jazz were forming and developing, the habanera was a consistent part of African-American popular music. Habaneras were widely available as sheet music and were the first written music which was rhythmically based on an African motif (1803). From the perspective of African-American music, the "habanera rhythm" (also known as "congo"), "tango-congo", or tango. can be thought of as a combination of tresillo and the backbeat. The habanera was the first of many Cuban music genres which enjoyed periods of popularity in the United States and reinforced and inspired the use of tresillo-based rhythms in African-American music. New Orleans native Louis Moreau Gottschalk's piano piece "Ojos Criollos (Danse Cubaine)" (1860) was influenced by the composer's studies in Cuba: the habanera rhythm is clearly heard in the left hand. In Gottschalk's symphonic work "A Night in the Tropics" (1859), the tresillo variant cinquillo appears extensively. The figure was later used by Scott Joplin and other ragtime composers. Comparing the music of New Orleans with the music of Cuba, Wynton Marsalis observes that tresillo is the New Orleans "clavé", a Spanish word meaning "code" or "key", as in the key to a puzzle, or mystery. Although the pattern is only half a clave, Marsalis makes the point that the single-celled figure is the guide-pattern of New Orleans music. Jelly Roll Morton called the rhythmic figure the Spanish tinge and considered it an essential ingredient of jazz. The abolition of slavery in 1865 led to new opportunities for the education of freed African Americans. Although strict segregation limited employment opportunities for most blacks, many were able to find work in entertainment. Black musicians were able to provide entertainment in dances, minstrel shows, and in vaudeville, during which time many marching bands were formed. Black pianists played in bars, clubs, and brothels, as ragtime developed. Ragtime appeared as sheet music, popularized by African-American musicians such as the entertainer Ernest Hogan, whose hit songs appeared in 1895. Two years later, Vess Ossman recorded a medley of these songs as a banjo solo known as "Rag Time Medley". Also in 1897, the white composer William Krell published his "Mississippi Rag" as the first written piano instrumental ragtime piece, and Tom Turpin published his "Harlem Rag", the first rag published by an African-American. Classically trained pianist Scott Joplin produced his "Original Rags" in 1898 and, in 1899, had an international hit with "Maple Leaf Rag", a multi-strain ragtime march with four parts that feature recurring themes and a bass line with copious seventh chords. Its structure was the basis for many other rags, and the syncopations in the right hand, especially in the transition between the first and second strain, were novel at the time. The last four measures of Scott Joplin's "Maple Leaf Rag" (1899) are shown below. African-based rhythmic patterns such as tresillo and its variants, the habanera rhythm and cinquillo, are heard in the ragtime compositions of Joplin and Turpin. Joplin's "Solace" (1909) is generally considered to be in the habanera genre: both of the pianist's hands play in a syncopated fashion, completely abandoning any sense of a march rhythm. Ned Sublette postulates that the tresillo/habanera rhythm "found its way into ragtime and the cakewalk," whilst Roberts suggests that "the habanera influence may have been part of what freed black music from ragtime's European bass". In the northeastern United States, a "hot" style of playing ragtime had developed, notably James Reese Europe's symphonic Clef Club orchestra in New York City, which played a benefit concert at Carnegie Hall in 1912. The Baltimore rag style of Eubie Blake influenced James P. Johnson's development of stride piano playing, in which the right hand plays the melody, while the left hand provides the rhythm and bassline. In Ohio and elsewhere in the mid-west the major influence was ragtime, until about 1919. Around 1912, when the four-string banjo and saxophone came in, musicians began to improvise the melody line, but the harmony and rhythm remained unchanged. A contemporary account states that blues could only be heard in jazz in the gut-bucket cabarets, which were generally looked down upon by the Black middle-class. Blues is the name given to both a musical form and a music genre, which originated in African-American communities of primarily the Deep South of the United States at the end of the 19th century from their spirituals, work songs, field hollers, shouts and chants and rhymed simple narrative ballads. The African use of pentatonic scales contributed to the development of blue notes in blues and jazz. As Kubik explains: Many of the rural blues of the Deep South are stylistically an extension and merger of basically two broad accompanied song-style traditions in the west central Sudanic belt: W. C. Handy became interested in folk blues of the Deep South while traveling through the Mississippi Delta. In this folk blues form, the singer would improvise freely within a limited melodic range, sounding like a field holler, and the guitar accompaniment was slapped rather than strummed, like a small drum which responded in syncopated accents, functioning as another "voice". Handy and his band members were formally trained African-American musicians who had not grown up with the blues, yet he was able to adapt the blues to a larger band instrument format and arrange them in a popular music form. Handy wrote about his adopting of the blues: The primitive southern Negro, as he sang, was sure to bear down on the third and seventh tone of the scale, slurring between major and minor. Whether in the cotton field of the Delta or on the Levee up St. Louis way, it was always the same. Till then, however, I had never heard this slur used by a more sophisticated Negro, or by any white man. I tried to convey this effect ... by introducing flat thirds and sevenths (now called blue notes) into my song, although its prevailing key was major ... , and I carried this device into my melody as well. The publication of his "Memphis Blues" sheet music in 1912 introduced the 12-bar blues to the world (although Gunther Schuller argues that it is not really a blues, but "more like a cakewalk"). This composition, as well as his later "St. Louis Blues" and others, included the habanera rhythm, and would become jazz standards. Handy's music career began in the pre-jazz era and contributed to the codification of jazz through the publication of some of the first jazz sheet music. The music of New Orleans, Louisiana had a profound effect on the creation of early jazz. In New Orleans, slaves could practice elements of their culture such as voodoo and playing drums. Many early jazz musicians played in the bars and brothels of the red-light district around Basin Street called Storyville. In addition to dance bands, there were marching bands which played at lavish funerals (later called jazz funerals). The instruments used by marching bands and dance bands became the instruments of jazz: brass, drums, and reeds tuned in the European 12-tone scale. Small bands contained a combination of self-taught and formally educated musicians, many from the funeral procession tradition. These bands traveled in black communities in the deep south. Beginning in 1914, Louisiana Creole and African-American musicians played in vaudeville shows which carried jazz to cities in the northern and western parts of the U.S. Jazz became international in 1914, when the Creole Band with cornettist Freddie Keppard performed the first ever jazz concert outside the United States, at the Pantages Playhouse Theatre in Winnipeg, Canada. In New Orleans, a white bandleader named Papa Jack Laine integrated blacks and whites in his marching band. He was known as "the father of white jazz" because of the many top players he employed, such as George Brunies, Sharkey Bonano, and future members of the Original Dixieland Jass Band. During the early 1900s, jazz was mostly performed in African-American and mulatto communities due to segregation laws. Storyville brought jazz to a wider audience through tourists who visited the port city of New Orleans. Many jazz musicians from African-American communities were hired to perform in bars and brothels. These included Buddy Bolden and Jelly Roll Morton in addition to those from other communities, such as Lorenzo Tio and Alcide Nunez. Louis Armstrong started his career in Storyville and found success in Chicago. Storyville was shut down by the U.S. government in 1917. Cornetist Buddy Bolden played in New Orleans from 1895 to 1906. No recordings by him exist. His band is credited with creating the big four: the first syncopated bass drum pattern to deviate from the standard on-the-beat march. As the example below shows, the second half of the big four pattern is the habanera rhythm. Afro-Creole pianist Jelly Roll Morton began his career in Storyville. Beginning in 1904, he toured with vaudeville shows to southern cities, Chicago, and New York City. In 1905, he composed "Jelly Roll Blues", which became the first jazz arrangement in print when it was published in 1915. It introduced more musicians to the New Orleans style. Morton considered the tresillo/habanera, which he called the Spanish tinge, an essential ingredient of jazz. "Now in one of my earliest tunes, "New Orleans Blues," you can notice the Spanish tinge. In fact, if you can't manage to put tinges of Spanish in your tunes, you will never be able to get the right seasoning, I call it, for jazz." An excerpt of "New Orleans Blues" is shown below. In the excerpt, the left hand plays the tresillo rhythm, while the right hand plays variations on cinquillo. Morton was a crucial innovator in the evolution from the early jazz form known as ragtime to jazz piano, and could perform pieces in either style; in 1938, Morton made a series of recordings for the Library of Congress in which he demonstrated the difference between the two styles. Morton's solos, however, were still close to ragtime, and were not merely improvisations over chord changes as in later jazz, but his use of the blues was of equal importance. Morton loosened ragtime's rigid rhythmic feeling, decreasing its embellishments and employing a swing feeling. Swing is the most important and enduring African-based rhythmic technique used in jazz. An oft quoted definition of swing by Louis Armstrong is: "if you don't feel it, you'll never know it." The New Harvard Dictionary of Music states that swing is: "An intangible rhythmic momentum in jazz...Swing defies analysis; claims to its presence may inspire arguments." The dictionary does nonetheless provide the useful description of triple subdivisions of the beat contrasted with duple subdivisions: swing superimposes six subdivisions of the beat over a basic pulse structure or four subdivisions. This aspect of swing is far more prevalent in African-American music than in Afro-Caribbean music. One aspect of swing, which is heard in more rhythmically complex Diaspora musics, places strokes in-between the triple and duple-pulse "grids". New Orleans brass bands are a lasting influence, contributing horn players to the world of professional jazz with the distinct sound of the city whilst helping black children escape poverty. The leader of New Orleans' Camelia Brass Band, D'Jalma Ganier, taught Louis Armstrong to play trumpet; Armstrong would then popularize the New Orleans style of trumpet playing, and then expand it. Like Jelly Roll Morton, Armstrong is also credited with the abandonment of ragtime's stiffness in favor of swung notes. Armstrong, perhaps more than any other musician, codified the rhythmic technique of swing in jazz and broadened the jazz solo vocabulary. The Original Dixieland Jass Band made the music's first recordings early in 1917, and their "Livery Stable Blues" became the earliest released jazz record. That year, numerous other bands made recordings featuring "jazz" in the title or band name, but most were ragtime or novelty records rather than jazz. In February 1918 during World War I, James Reese Europe's "Hellfighters" infantry band took ragtime to Europe, then on their return recorded Dixieland standards including "Darktown Strutters' Ball". From 1920 to 1933, Prohibition in the United States banned the sale of alcoholic drinks, resulting in illicit speakeasies which became lively venues of the "Jazz Age", hosting popular music, dance songs, novelty songs, and show tunes. Jazz began to get a reputation as immoral, and many members of the older generations saw it as a threat to the old cultural values by promoting the decadent values of the Roaring 20s. Henry van Dyke of Princeton University wrote, "... it is not music at all. It's merely an irritation of the nerves of hearing, a sensual teasing of the strings of physical passion." The New York Times reported that Siberian villagers used jazz to scare away bears, but the villagers had used pots and pans; another story claimed that the fatal heart attack of a celebrated conductor was caused by jazz. In 1919, Kid Ory's Original Creole Jazz Band of musicians from New Orleans began playing in San Francisco and Los Angeles, where in 1922 they became the first black jazz band of New Orleans origin to make recordings. During the same year, Bessie Smith made her first recordings. Chicago was developing "Hot Jazz", and King Oliver joined Bill Johnson. Bix Beiderbecke formed The Wolverines in 1924. Despite its Southern black origins, there was a larger market for jazzy dance music played by white orchestras. In 1918, Paul Whiteman and his orchestra became a hit in San Francisco. He signed a contract with Victor and became the top bandleader of the 1920s, giving hot jazz a white component, hiring white musicians such as Bix Beiderbecke, Jimmy Dorsey, Tommy Dorsey, Frankie Trumbauer, and Joe Venuti. In 1924, Whiteman commissioned George Gershwin's Rhapsody in Blue, which was premiered by his orchestra. Jazz began to be recognized as a notable musical form. Olin Downes, reviewing the concert in The New York Times, wrote, "This composition shows extraordinary talent, as it shows a young composer with aims that go far beyond those of his ilk, struggling with a form of which he is far from being master. ... In spite of all this, he has expressed himself in a significant and, on the whole, highly original form. ... His first theme ... is no mere dance-tune ... it is an idea, or several ideas, correlated and combined in varying and contrasting rhythms that immediately intrigue the listener." After Whiteman's band successfully toured Europe, huge hot jazz orchestras in theater pits caught on with other whites, including Fred Waring, Jean Goldkette, and Nathaniel Shilkret. According to Mario Dunkel, Whiteman's success was based on a "rhetoric of domestication" according to which he had elevated and rendered valuable (read "white") a previously inchoate (read "black") kind of music. Whiteman's success caused black artists to follow suit, including Earl Hines (who opened in The Grand Terrace Cafe in Chicago in 1928), Duke Ellington (who opened at the Cotton Club in Harlem in 1927), Lionel Hampton, Fletcher Henderson, Claude Hopkins, and Don Redman, with Henderson and Redman developing the "talking to one another" formula for "hot" swing music. In 1924, Louis Armstrong joined the Fletcher Henderson dance band for a year, as featured soloist. The original New Orleans style was polyphonic, with theme variation and simultaneous collective improvisation. Armstrong was a master of his hometown style, but by the time he joined Henderson's band, he was already a trailblazer in a new phase of jazz, with its emphasis on arrangements and soloists. Armstrong's solos went well beyond the theme-improvisation concept and extemporized on chords, rather than melodies. According to Schuller, by comparison, the solos by Armstrong's bandmates (including a young Coleman Hawkins), sounded "stiff, stodgy", with "jerky rhythms and a grey undistinguished tone quality". The following example shows a short excerpt of the straight melody of "Mandy, Make Up Your Mind" by George W. Meyer and Arthur Johnston (top), compared with Armstrong's solo improvisations (below) (recorded 1924). Armstrong's solos were a significant factor in making jazz a true 20th-century language. After leaving Henderson's group, Armstrong formed his Hot Five band, where he popularized scat singing. The 1930s belonged to popular swing big bands, in which some virtuoso soloists became as famous as the band leaders. Key figures in developing the "big" jazz band included bandleaders and arrangers Count Basie, Cab Calloway, Jimmy and Tommy Dorsey, Duke Ellington, Benny Goodman, Fletcher Henderson, Earl Hines, Harry James, Jimmie Lunceford, Glenn Miller and Artie Shaw. Although it was a collective sound, swing also offered individual musicians a chance to "solo" and improvise melodic, thematic solos which could at times be complex "important" music. Over time, social strictures regarding racial segregation began to relax in America: white bandleaders began to recruit black musicians and black bandleaders white ones. In the mid-1930s, Benny Goodman hired pianist Teddy Wilson, vibraphonist Lionel Hampton and guitarist Charlie Christian to join small groups. In the 1930s, Kansas City Jazz as exemplified by tenor saxophonist Lester Young marked the transition from big bands to the bebop influence of the 1940s. An early 1940s style known as "jumping the blues" or jump blues used small combos, uptempo music and blues chord progressions, drawing on boogie-woogie from the 1930s. While swing was reaching the height of its popularity, Duke Ellington spent the late 1920s and 1930s developing an innovative musical idiom for his orchestra. Abandoning the conventions of swing, he experimented with orchestral sounds, harmony, and musical form with complex compositions that still translated well for popular audiences; some of his tunes became hits, and his own popularity spanned from the United States to Europe. Ellington called his music American Music, rather than jazz, and liked to describe those who impressed him as "beyond category". These included many musicians from his orchestra, some of whom are considered among the best in jazz in their own right, but it was Ellington who melded them into one of the most popular jazz orchestras in the history of jazz. He often composed for the style and skills of these individuals, such as "Jeep's Blues" for Johnny Hodges, "Concerto for Cootie" for Cootie Williams (which later became "Do Nothing Till You Hear from Me" with Bob Russell's lyrics), and "The Mooche" for Tricky Sam Nanton and Bubber Miley. He also recorded compositions written by his bandsmen, such as Juan Tizol's "Caravan" and "Perdido", which brought the "Spanish Tinge" to big-band jazz. Several members of the orchestra remained with him for several decades. The band reached a creative peak in the early 1940s, when Ellington and a small hand-picked group of his composers and arrangers wrote for an orchestra of distinctive voices who displayed tremendous creativity. As only a limited number of American jazz records were released in Europe, European jazz traces many of its roots to American artists such as James Reese Europe, Paul Whiteman, and Lonnie Johnson, who visited Europe during and after World War I. It was their live performances which inspired European audiences' interest in jazz, as well as the interest in all things American (and therefore exotic) which accompanied the economic and political woes of Europe during this time. The beginnings of a distinct European style of jazz began to emerge in this interwar period. British jazz began with a tour by the Original Dixieland Jazz Band in 1919. In 1926, Fred Elizalde and His Cambridge Undergraduates began broadcasting on the BBC. Thereafter jazz became an important element in many leading dance orchestras, and jazz instrumentalists became numerous. This style entered full swing in France with the Quintette du Hot Club de France, which began in 1934. Much of this French jazz was a combination of African-American jazz and the symphonic styles in which French musicians were well-trained; in this, it is easy to see the inspiration taken from Paul Whiteman since his style was also a fusion of the two. Belgian guitarist Django Reinhardt popularized gypsy jazz, a mix of 1930s American swing, French dance hall "musette", and Eastern European folk with a languid, seductive feel; the main instruments were steel stringed guitar, violin, and double bass. Solos pass from one player to another as guitar and bass form the rhythm section. Some researchers believe Eddie Lang and Joe Venuti pioneered the guitar-violin partnership characteristic of the genre, which was brought to France after they had been heard live or on Okeh Records in the late 1920s. The outbreak of World War II marked a turning point for jazz. The swing-era jazz of the previous decade had challenged other popular music as being representative of the nation's culture, with big bands reaching the height of the style's success by the early 1940s; swing acts and big bands traveled with U.S. military overseas to Europe, where it also became popular. Stateside, however, the war presented difficulties for the big-band format: conscription shortened the number of musicians available; the military's need for shellac (commonly used for pressing gramophone records) limited record production; a shortage of rubber (also due to the war effort) discouraged bands from touring via road travel; and a demand by the musicians' union for a commercial recording ban limited music distribution between 1942 and 1944. Many of the big bands who were deprived of experienced musicians because of the war effort began to enlist young players who were below the age for conscription, as was the case with saxophonist Stan Getz's entry in a band as a teenager. This coincided with a nationwide resurgence in the Dixieland style of pre-swing jazz; performers such as clarinetist George Lewis, cornetist Bill Davison, and trombonist Turk Murphy were hailed by conservative jazz critics as more authentic than the big bands. Elsewhere, with the limitations on recording, small groups of young musicians developed a more uptempo, improvisational style of jazz, collaborating and experimenting with new ideas for melodic development, rhythmic language, and harmonic substitution, during informal, late-night jam sessions hosted in small clubs and apartments. Key figures in this development were largely based in New York and included pianists Thelonious Monk and Bud Powell, drummers Max Roach and Kenny Clarke, saxophonist Charlie Parker, and trumpeter Dizzy Gillespie. This musical development became known as bebop. Bebop and subsequent post-war jazz developments featured a wider set of notes, played in more complex patterns and at faster tempos than previous jazz. According to Clive James, bebop was "the post-war musical development which tried to ensure that jazz would no longer be the spontaneous sound of joy ... Students of race relations in America are generally agreed that the exponents of post-war jazz were determined, with good reason, to present themselves as challenging artists rather than tame entertainers." The end of the war marked "a revival of the spirit of experimentation and musical pluralism under which it had been conceived", along with "the beginning of a decline in the popularity of jazz music in America", according to American academic Michael H. Burchett. With the rise of bebop and the end of the swing era after the war, jazz lost its cachet as pop music. Vocalists of the famous big bands moved on to being marketed and performing as solo pop singers; these included Frank Sinatra, Peggy Lee, Dick Haymes, and Doris Day. Older musicians who still performed their pre-war jazz, such as Armstrong and Ellington, were gradually viewed in the mainstream as passé. Other younger performers, such as singer Big Joe Turner and saxophonist Louis Jordan, who were discouraged by bebop's increasing complexity, pursued more lucrative endeavors in rhythm and blues, jump blues, and eventually rock and roll. Some, including Gillespie, composed intricate yet danceable pieces for bebop musicians in an effort to make them more accessible, but bebop largely remained on the fringes of American audiences' purview. "The new direction of postwar jazz drew a wealth of critical acclaim, but it steadily declined in popularity as it developed a reputation as an academic genre that was largely inaccessible to mainstream audiences", Burchett said. "The quest to make jazz more relevant to popular audiences, while retaining its artistic integrity, is a constant and prevalent theme in the history of postwar jazz." During its swing period, jazz had been an uncomplicated musical scene; according to Paul Trynka, this changed in the post-war years: Suddenly jazz was no longer straightforward. There was bebop and its variants, there was the last gasp of swing, there were strange new brews like the progressive jazz of Stan Kenton, and there was a completely new phenomenon called revivalism – the rediscovery of jazz from the past, either on old records or performed live by aging players brought out of retirement. From now on it was no good saying that you liked jazz, you had to specify what kind of jazz. And that is the way it has been ever since, only more so. Today, the word 'jazz' is virtually meaningless without further definition. In the early 1940s, bebop-style performers began to shift jazz from danceable popular music toward a more challenging "musician's music". The most influential bebop musicians included saxophonist Charlie Parker, pianists Bud Powell and Thelonious Monk, trumpeters Dizzy Gillespie and Clifford Brown, and drummer Max Roach. Divorcing itself from dance music, bebop established itself more as an art form, thus lessening its potential popular and commercial appeal. Composer Gunther Schuller wrote: "In 1943 I heard the great Earl Hines band which had Bird in it and all those other great musicians. They were playing all the flatted fifth chords and all the modern harmonies and substitutions and Dizzy Gillespie runs in the trumpet section work. Two years later I read that that was 'bop' and the beginning of modern jazz ... but the band never made recordings." Dizzy Gillespie wrote: "People talk about the Hines band being 'the incubator of bop' and the leading exponents of that music ended up in the Hines band. But people also have the erroneous impression that the music was new. It was not. The music evolved from what went before. It was the same basic music. The difference was in how you got from here to here to here...naturally each age has got its own shit." Since bebop was meant to be listened to, not danced to, it could use faster tempos. Drumming shifted to a more elusive and explosive style, in which the ride cymbal was used to keep time while the snare and bass drum were used for accents. This led to a highly syncopated music with a linear rhythmic complexity. Bebop musicians employed several harmonic devices which were not previously typical in jazz, engaging in a more abstracted form of chord-based improvisation. Bebop scales are traditional scales with an added chromatic passing note; bebop also uses "passing" chords, substitute chords, and altered chords. New forms of chromaticism and dissonance were introduced into jazz, and the dissonant tritone (or "flatted fifth") interval became the "most important interval of bebop" Chord progressions for bebop tunes were often taken directly from popular swing-era tunes and reused with a new and more complex melody or reharmonized with more complex chord progressions to form new compositions, a practice which was already well-established in earlier jazz, but came to be central to the bebop style. Bebop made use of several relatively common chord progressions, such as blues (at base, I–IV–V, but often infused with ii–V motion) and "rhythm changes" (I–VI–ii–V) – the chords to the 1930s pop standard "I Got Rhythm". Late bop also moved towards extended forms that represented a departure from pop and show tunes. The harmonic development in bebop is often traced back to a moment experienced by Charlie Parker while performing "Cherokee" at Clark Monroe's Uptown House, New York, in early 1942. "I'd been getting bored with the stereotyped changes that were being used...and I kept thinking there's bound to be something else. I could hear it sometimes. I couldn't play it...I was working over 'Cherokee,' and, as I did, I found that by using the higher intervals of a chord as a melody line and backing them with appropriately related changes, I could play the thing I'd been hearing. It came alive." Gerhard Kubik postulates that harmonic development in bebop sprang from blues and African-related tonal sensibilities rather than 20th-century Western classical music. "Auditory inclinations were the African legacy in [Parker's] life, reconfirmed by the experience of the blues tonal system, a sound world at odds with the Western diatonic chord categories. Bebop musicians eliminated Western-style functional harmony in their music while retaining the strong central tonality of the blues as a basis for drawing upon various African matrices." Samuel Floyd states that blues was both the bedrock and propelling force of bebop, bringing about a new harmonic conception using extended chord structures that led to unprecedented harmonic and melodic variety, a developed and even more highly syncopated, linear rhythmic complexity and a melodic angularity in which the blue note of the fifth degree was established as an important melodic-harmonic device; and reestablishment of the blues as the primary organizing and functional principle. Kubik wrote: While for an outside observer, the harmonic innovations in bebop would appear to be inspired by experiences in Western "serious" music, from Claude Debussy to Arnold Schoenberg, such a scheme cannot be sustained by the evidence from a cognitive approach. Claude Debussy did have some influence on jazz, for example, on Bix Beiderbecke's piano playing. And it is also true that Duke Ellington adopted and reinterpreted some harmonic devices in European contemporary music. West Coast jazz would run into such debts as would several forms of cool jazz, but bebop has hardly any such debts in the sense of direct borrowings. On the contrary, ideologically, bebop was a strong statement of rejection of any kind of eclecticism, propelled by a desire to activate something deeply buried in self. Bebop then revived tonal-harmonic ideas transmitted through the blues and reconstructed and expanded others in a basically non-Western harmonic approach. The ultimate significance of all this is that the experiments in jazz during the 1940s brought back to African-American music several structural principles and techniques rooted in African traditions. These divergences from the jazz mainstream of the time met a divided, sometimes hostile response among fans and musicians, especially swing players who bristled at the new harmonic sounds. To hostile critics, bebop seemed filled with "racing, nervous phrases". But despite the friction, by the 1950s bebop had become an accepted part of the jazz vocabulary. The general consensus among musicians and musicologists is that the first original jazz piece to be overtly based in clave was "Tanga" (1943), composed by Cuban-born Mario Bauza and recorded by Machito and his Afro-Cubans in New York City. "Tanga" began as a spontaneous descarga (Cuban jam session), with jazz solos superimposed on top. This was the birth of Afro-Cuban jazz. The use of clave brought the African timeline, or key pattern, into jazz. Music organized around key patterns convey a two-celled (binary) structure, which is a complex level of African cross-rhythm. Within the context of jazz, however, harmony is the primary referent, not rhythm. The harmonic progression can begin on either side of clave, and the harmonic "one" is always understood to be "one". If the progression begins on the "three-side" of clave, it is said to be in 3–2 clave (shown below). If the progression begins on the "two-side", it is in 2–3 clave. Mario Bauzá introduced bebop innovator Dizzy Gillespie to Cuban conga drummer and composer Chano Pozo. Gillespie and Pozo's brief collaboration produced some of the most enduring Afro-Cuban jazz standards. "Manteca" (1947) is the first jazz standard to be rhythmically based on clave. According to Gillespie, Pozo composed the layered, contrapuntal guajeos (Afro-Cuban ostinatos) of the A section and the introduction, while Gillespie wrote the bridge. Gillespie recounted: "If I'd let it go like [Chano] wanted it, it would have been strictly Afro-Cuban all the way. There wouldn't have been a bridge. I thought I was writing an eight-bar bridge, but ... I had to keep going and ended up writing a sixteen-bar bridge." The bridge gave "Manteca" a typical jazz harmonic structure, setting the piece apart from Bauza's modal "Tanga" of a few years earlier. Gillespie's collaboration with Pozo brought specific African-based rhythms into bebop. While pushing the boundaries of harmonic improvisation, cu-bop also drew from African rhythm. Jazz arrangements with a Latin A section and a swung B section, with all choruses swung during solos, became common practice with many Latin tunes of the jazz standard repertoire. This approach can be heard on pre-1980 recordings of "Manteca", "A Night in Tunisia", "Tin Tin Deo", and "On Green Dolphin Street". Another jazz composition critical to the development of Afro-Cuban jazz was Bud Powell's "Un Poco Loco," recorded with Curley Russell on bass and Max Roach on drums. Noted for its "frenetic energy" and "clanging cowbell and polyrhythmic accompaniment," the composition combined Afro-Cuban rhythm with polytonality and preceded further use of modality and avant-garde harmony in Latin jazz. Cuban percussionist Mongo Santamaria first recorded his composition "Afro Blue" in 1959. "Afro Blue" was the first jazz standard built upon a typical African three-against-two (3:2) cross-rhythm, or hemiola. The piece begins with the bass repeatedly playing 6 cross-beats per each measure of 8, or 6 cross-beats per 4 main beats—6:4 (two cells of 3:2). The following example shows the original ostinato "Afro Blue" bass line. The cross noteheads indicate the main beats (not bass notes). When John Coltrane covered "Afro Blue" in 1963, he inverted the metric hierarchy, interpreting the tune as a 4 jazz waltz with duple cross-beats superimposed (2:3). Originally a B♭ pentatonic blues, Coltrane expanded the harmonic structure of "Afro Blue". Perhaps the most respected Afro-cuban jazz combo of the late 1950s was vibraphonist Cal Tjader's band. Tjader had Mongo Santamaria, Armando Peraza, and Willie Bobo on his early recording dates. In the late 1940s, there was a revival of Dixieland, harking back to the contrapuntal New Orleans style. This was driven in large part by record company reissues of jazz classics by the Oliver, Morton, and Armstrong bands of the 1930s. There were two types of musicians involved in the revival: the first group was made up of those who had begun their careers playing in the traditional style and were returning to it (or continuing what they had been playing all along), such as Bob Crosby's Bobcats, Max Kaminsky, Eddie Condon, and Wild Bill Davison. Most of these players were originally Midwesterners, although there were a small number of New Orleans musicians involved. The second group of revivalists consisted of younger musicians, such as those in the Lu Watters band, Conrad Janis, and Ward Kimball and his Firehouse Five Plus Two Jazz Band. By the late 1940s, Louis Armstrong's Allstars band became a leading ensemble. Through the 1950s and 1960s, Dixieland was one of the most commercially popular jazz styles in the US, Europe, and Japan, although critics paid little attention to it. Hard bop is an extension of bebop (or "bop") music that incorporates influences from blues, rhythm and blues, and gospel, especially in saxophone and piano playing. Hard bop was developed in the mid-1950s, coalescing in 1953 and 1954; it developed partly in response to the vogue for cool jazz in the early 1950s and paralleled the rise of rhythm and blues. It has been described as "funky" and can be considered a relative of soul jazz. Some elements of the genre were simplified from their bebop roots. Miles Davis' 1954 performance of "Walkin'" at the first Newport Jazz Festival introduced the style to the jazz world. Further leaders of hard bop's development included the Clifford Brown/Max Roach Quintet, Art Blakey's Jazz Messengers, the Horace Silver Quintet, and trumpeters Lee Morgan and Freddie Hubbard. The late 1950s to early 1960s saw hard boppers form their own bands as a new generation of blues- and bebop-influenced musicians entered the jazz world, from pianists Wynton Kelly and Tommy Flanagan to saxophonists Joe Henderson and Hank Mobley. Coltrane, Johnny Griffin, Mobley, and Morgan all participated on the album A Blowin' Session (1957), considered by Al Campbell to have been one of the high points of the hard bop era. Hard bop was prevalent within jazz for about a decade spanning from 1955 to 1965, but has remained highly influential on mainstream or "straight-ahead" jazz. It went into decline in the late 1960s through the 1970s due to the emergence of other styles such as jazz fusion, but again became influential following the Young Lions Movement and the emergence of neo-bop. Modal jazz is a development which began in the later 1950s which takes the mode, or musical scale, as the basis of musical structure and improvisation. Previously, a solo was meant to fit into a given chord progression, but with modal jazz, the soloist creates a melody using one (or a small number of) modes. The emphasis is thus shifted from harmony to melody: "Historically, this caused a seismic shift among jazz musicians, away from thinking vertically (the chord), and towards a more horizontal approach (the scale)", explained pianist Mark Levine. The modal theory stems from a work by George Russell. Miles Davis introduced the concept to the greater jazz world with Kind of Blue (1959), an exploration of the possibilities of modal jazz which would become the best selling jazz album of all time. In contrast to Davis' earlier work with hard bop and its complex chord progression and improvisation, Kind of Blue was composed as a series of modal sketches in which the musicians were given scales that defined the parameters of their improvisation and style. "I didn't write out the music for Kind of Blue, but brought in sketches for what everybody was supposed to play because I wanted a lot of spontaneity," recalled Davis. The track "So What" has only two chords: D-7 and E♭-7. Other innovators in this style include Jackie McLean, and two of the musicians who had also played on Kind of Blue: John Coltrane and Bill Evans. Free jazz, and the related form of avant-garde jazz, broke through into an open space of "free tonality" in which meter, beat, and formal symmetry all disappeared, and a range of world music from India, Africa, and Arabia were melded into an intense, even religiously ecstatic or orgiastic style of playing. While loosely inspired by bebop, free jazz tunes gave players much more latitude; the loose harmony and tempo was deemed controversial when this approach was first developed. The bassist Charles Mingus is also frequently associated with the avant-garde in jazz, although his compositions draw from myriad styles and genres. The first major stirrings came in the 1950s with the early work of Ornette Coleman (whose 1960 album Free Jazz: A Collective Improvisation coined the term) and Cecil Taylor. In the 1960s, exponents included Albert Ayler, Gato Barbieri, Carla Bley, Don Cherry, Larry Coryell, John Coltrane, Bill Dixon, Jimmy Giuffre, Steve Lacy, Michael Mantler, Sun Ra, Roswell Rudd, Pharoah Sanders, and John Tchicai. In developing his late style, Coltrane was especially influenced by the dissonance of Ayler's trio with bassist Gary Peacock and drummer Sunny Murray, a rhythm section honed with Cecil Taylor as leader. In November 1961, Coltrane played a gig at the Village Vanguard, which resulted in the classic Chasin' the 'Trane, which DownBeat magazine panned as "anti-jazz". On his 1961 tour of France, he was booed, but persevered, signing with the new Impulse! Records in 1960 and turning it into "the house that Trane built", while championing many younger free jazz musicians, notably Archie Shepp, who often played with trumpeter Bill Dixon, who organized the 4-day "October Revolution in Jazz" in Manhattan in 1964, the first free jazz festival. A series of recordings with the Classic Quartet in the first half of 1965 show Coltrane's playing becoming increasingly abstract, with greater incorporation of devices like multiphonics, utilization of overtones, and playing in the altissimo register, as well as a mutated return to Coltrane's sheets of sound. In the studio, he all but abandoned his soprano to concentrate on the tenor saxophone. In addition, the quartet responded to the leader by playing with increasing freedom. The group's evolution can be traced through the recordings The John Coltrane Quartet Plays, Living Space and Transition (both June 1965), New Thing at Newport (July 1965), Sun Ship (August 1965), and First Meditations (September 1965). In June 1965, Coltrane and 10 other musicians recorded Ascension, a 40-minute-long piece without breaks that included adventurous solos by young avant-garde musicians as well as Coltrane, and was controversial primarily for the collective improvisation sections that separated the solos. Dave Liebman later called it "the torch that lit the free jazz thing". After recording with the quartet over the next few months, Coltrane invited Pharoah Sanders to join the band in September 1965. While Coltrane used over-blowing frequently as an emotional exclamation-point, Sanders would opt to overblow his entire solo, resulting in a constant screaming and screeching in the altissimo range of the instrument. Free jazz was played in Europe in part because musicians such as Ayler, Taylor, Steve Lacy, and Eric Dolphy spent extended periods of time there, and European musicians such as Michael Mantler and John Tchicai traveled to the U.S. to experience American music firsthand. European contemporary jazz was shaped by Peter Brötzmann, John Surman, Krzysztof Komeda, Zbigniew Namysłowski, Tomasz Stanko, Lars Gullin, Joe Harriott, Albert Mangelsdorff, Kenny Wheeler, Graham Collier, Michael Garrick and Mike Westbrook. They were eager to develop approaches to music that reflected their heritage. Since the 1960s, creative centers of jazz in Europe have developed, such as the creative jazz scene in Amsterdam. Following the work of drummer Han Bennink and pianist Misha Mengelberg, musicians started to explore by improvising collectively until a form (melody, rhythm, a famous song) is found Jazz critic Kevin Whitehead documented the free jazz scene in Amsterdam and some of its main exponents such as the ICP (Instant Composers Pool) orchestra in his book New Dutch Swing. Since the 1990s Keith Jarrett has defended free jazz from criticism. British writer Stuart Nicholson has argued European contemporary jazz has an identity different from American jazz and follows a different trajectory. Latin jazz is jazz that employs Latin American rhythms and is generally understood to have a more specific meaning than simply jazz from Latin America. A more precise term might be Afro-Latin jazz, as the jazz subgenre typically employs rhythms that either have a direct analog in Africa or exhibit an African rhythmic influence beyond what is ordinarily heard in other jazz. The two main categories of Latin jazz are Afro-Cuban jazz and Brazilian jazz. In the 1960s and 1970s, many jazz musicians had only a basic understanding of Cuban and Brazilian music, and jazz compositions which used Cuban or Brazilian elements were often referred to as "Latin tunes", with no distinction between a Cuban son montuno and a Brazilian bossa nova. Even as late as 2000, in Mark Gridley's Jazz Styles: History and Analysis, a bossa nova bass line is referred to as a "Latin bass figure". It was not uncommon during the 1960s and 1970s to hear a conga playing a Cuban tumbao while the drumset and bass played a Brazilian bossa nova pattern. Many jazz standards such as "Manteca", "On Green Dolphin Street" and "Song for My Father" have a "Latin" A section and a swung B section. Typically, the band would only play an even-eighth "Latin" feel in the A section of the head and swing throughout all of the solos. Latin jazz specialists like Cal Tjader tended to be the exception. For example, on a 1959 live Tjader recording of "A Night in Tunisia", pianist Vince Guaraldi soloed through the entire form over an authentic mambo. For most of its history, Afro-Cuban jazz had been a matter of superimposing jazz phrasing over Cuban rhythms. But by the end of the 1970s, a new generation of New York City musicians had emerged who were fluent in both salsa dance music and jazz, leading to a new level of integration of jazz and Cuban rhythms. This era of creativity and vitality is best represented by the Gonzalez brothers Jerry (congas and trumpet) and Andy (bass). During 1974–1976, they were members of one of Eddie Palmieri's most experimental salsa groups: salsa was the medium, but Palmieri was stretching the form in new ways. He incorporated parallel fourths, with McCoy Tyner-type vamps. The innovations of Palmieri, the Gonzalez brothers and others led to an Afro-Cuban jazz renaissance in New York City. This occurred in parallel with developments in Cuba The first Cuban band of this new wave was Irakere. Their "Chékere-son" (1976) introduced a style of "Cubanized" bebop-flavored horn lines that departed from the more angular guajeo-based lines which were typical of Cuban popular music and Latin jazz up until that time. It was based on Charlie Parker's composition "Billie's Bounce", jumbled together in a way that fused clave and bebop horn lines. In spite of the ambivalence of some band members towards Irakere's Afro-Cuban folkloric / jazz fusion, their experiments forever changed Cuban jazz: their innovations are still heard in the high level of harmonic and rhythmic complexity in Cuban jazz and in the jazzy and complex contemporary form of popular dance music known as timba. Brazilian jazz, such as bossa nova, is derived from samba, with influences from jazz and other 20th-century classical and popular music styles. Bossa is generally moderately paced, with melodies sung in Portuguese or English, whilst the related jazz-samba is an adaptation of street samba into jazz. The bossa nova style was pioneered by Brazilians João Gilberto and Antônio Carlos Jobim and was made popular by Elizete Cardoso's recording of "Chega de Saudade" on the Canção do Amor Demais LP. Gilberto's initial releases, and the 1959 film Black Orpheus, achieved significant popularity in Latin America; this spread to North America via visiting American jazz musicians. The resulting recordings by Charlie Byrd and Stan Getz cemented bossa nova's popularity and led to a worldwide boom, with 1963's Getz/Gilberto, numerous recordings by famous jazz performers such as Ella Fitzgerald and Frank Sinatra, and the eventual entrenchment of the bossa nova style as a lasting influence in world music. Brazilian percussionists such as Airto Moreira and Naná Vasconcelos also influenced jazz internationally by introducing Afro-Brazilian folkloric instruments and rhythms into a wide variety of jazz styles, thus attracting a greater audience to them. While bossa nova has been labeled as jazz by music critics, namely those from outside of Brazil, it has been rejected by many prominent bossa nova musicians such as Jobim, who once said "Bossa nova is not Brazilian jazz." The first jazz standard composed by a non-Latino to use an overt African 8 cross-rhythm was Wayne Shorter's "Footprints" (1967). On the version recorded on Miles Smiles by Miles Davis, the bass switches to a 4 tresillo figure at 2:20. "Footprints" is not, however, a Latin jazz tune: African rhythmic structures are accessed directly by Ron Carter (bass) and Tony Williams (drums) via the rhythmic sensibilities of swing. Throughout the piece, the four beats, whether sounded or not, are maintained as the temporal referent. The following example shows the 8 and 4 forms of the bass line. The slashed noteheads indicate the main beats (not bass notes), where one ordinarily taps their foot to "keep time". The use of pentatonic scales was another trend associated with Africa. The use of pentatonic scales in Africa probably goes back thousands of years. McCoy Tyner perfected the use of the pentatonic scale in his solos, and also used parallel fifths and fourths, which are common harmonies in West Africa. The minor pentatonic scale is often used in blues improvisation, and like a blues scale, a minor pentatonic scale can be played over all of the chords in a blues. The following pentatonic lick was played over blues changes by Joe Henderson on Horace Silver's "African Queen" (1965). Jazz pianist, theorist, and educator Mark Levine refers to the scale generated by beginning on the fifth step of a pentatonic scale as the V pentatonic scale. Levine points out that the V pentatonic scale works for all three chords of the standard II–V–I jazz progression. This is a very common progression, used in pieces such as Miles Davis' "Tune Up". The following example shows the V pentatonic scale over a II–V–I progression. Accordingly, John Coltrane's "Giant Steps" (1960), with its 26 chords per 16 bars, can be played using only three pentatonic scales. Coltrane studied Nicolas Slonimsky's Thesaurus of Scales and Melodic Patterns, which contains material that is virtually identical to portions of "Giant Steps". The harmonic complexity of "Giant Steps" is on the level of the most advanced 20th-century art music. Superimposing the pentatonic scale over "Giant Steps" is not merely a matter of harmonic simplification, but also a sort of "Africanizing" of the piece, which provides an alternate approach for soloing. Mark Levine observes that when mixed in with more conventional "playing the changes", pentatonic scales provide "structure and a feeling of increased space". As noted above, jazz has incorporated from its inception aspects of African-American sacred music including spirituals and hymns. Secular jazz musicians often performed renditions of spirituals and hymns as part of their repertoire or isolated compositions such as "Come Sunday", part of "Black and Beige Suite" by Duke Ellington. Later many other jazz artists borrowed from black gospel music. However, it was only after World War II that a few jazz musicians began to compose and perform extended works intended for religious settings or as religious expression. Since the 1950s, sacred and liturgical music has been performed and recorded by many prominent jazz composers and musicians. The "Abyssinian Mass" by Wynton Marsalis (Blueengine Records, 2016) is a recent example. Relatively little has been written about sacred and liturgical jazz. In a 2013 doctoral dissertation, Angelo Versace examined the development of sacred jazz in the 1950s using disciplines of musicology and history. He noted that the traditions of black gospel music and jazz were combined in the 1950s to produce a new genre, "sacred jazz". Versace maintained that the religious intent separates sacred from secular jazz. Most prominent in initiating the sacred jazz movement were pianist and composer Mary Lou Williams, known for her jazz masses in the 1950s and Duke Ellington. Prior to his death in 1974 in response to contacts from Grace Cathedral in San Francisco, Duke Ellington wrote three Sacred Concerts: 1965 – A Concert of Sacred Music; 1968 – Second Sacred Concert; 1973 – Third Sacred Concert. The most prominent form of sacred and liturgical jazz is the jazz mass. Although most often performed in a concert setting rather than church worship setting, this form has many examples. An eminent example of composers of the jazz mass was Mary Lou Williams. Williams converted to Catholicism in 1957, and proceeded to compose three masses in the jazz idiom. One was composed in 1968 to honor the recently assassinated Martin Luther King Jr. and the third was commissioned by a pontifical commission. It was performed once in 1975 in St Patrick's Cathedral in New York City. However the Catholic Church has not embraced jazz as appropriate for worship. In 1966 Joe Masters recorded "Jazz Mass" for Columbia Records. A jazz ensemble was joined by soloists and choir using the English text of the Roman Catholic Mass. Other examples include "Jazz Mass in Concert" by Lalo Schiffrin (Aleph Records, 1998, UPC 0651702632725) and "Jazz Mass" by Vince Guaraldi (Fantasy Records, 1965). In England, classical composer Will Todd recorded his "Jazz Missa Brevis" with a jazz ensemble, soloists and the St Martin's Voices on a 2018 Signum Records release, "Passion Music/Jazz Missa Brevis" also released as "Mass in Blue", and jazz organist James Taylor composed "The Rochester Mass" (Cherry Red Records, 2015). In 2013, Versace put forth bassist Ike Sturm and New York composer Deanna Witkowski as contemporary exemplars of sacred and liturgical jazz. In the late 1960s and early 1970s, the hybrid form of jazz-rock fusion was developed by combining jazz improvisation with rock rhythms, electric instruments and the highly amplified stage sound of rock musicians such as Jimi Hendrix and Frank Zappa. Jazz fusion often uses mixed meters, odd time signatures, syncopation, complex chords, and harmonies. According to AllMusic: ... until around 1967, the worlds of jazz and rock were nearly completely separate. [However, ...] as rock became more creative and its musicianship improved, and as some in the jazz world became bored with hard bop and did not want to play strictly avant-garde music, the two different idioms began to trade ideas and occasionally combine forces. In 1969, Davis fully embraced the electric instrument approach to jazz with In a Silent Way, which can be considered his first fusion album. Composed of two side-long suites edited heavily by producer Teo Macero, this quiet, static album would be equally influential to the development of ambient music. As Davis recalls: The music I was really listening to in 1968 was James Brown, the great guitar player Jimi Hendrix, and a new group who had just come out with a hit record, "Dance to the Music", Sly and the Family Stone ... I wanted to make it more like rock. When we recorded In a Silent Way I just threw out all the chord sheets and told everyone to play off of that. Two contributors to In a Silent Way also joined organist Larry Young to create one of the early acclaimed fusion albums: Emergency! (1969) by The Tony Williams Lifetime. Weather Report's self-titled electronic and psychedelic Weather Report debut album caused a sensation in the jazz world on its arrival in 1971, thanks to the pedigree of the group's members (including percussionist Airto Moreira), and their unorthodox approach to music. The album featured a softer sound than would be the case in later years (predominantly using acoustic bass with Shorter exclusively playing soprano saxophone, and with no synthesizers involved), but is still considered a classic of early fusion. It built on the avant-garde experiments which Joe Zawinul and Shorter had pioneered with Miles Davis on Bitches Brew, including an avoidance of head-and-chorus composition in favor of continuous rhythm and movement – but took the music further. To emphasize the group's rejection of standard methodology, the album opened with the inscrutable avant-garde atmospheric piece "Milky Way", which featured by Shorter's extremely muted saxophone inducing vibrations in Zawinul's piano strings while the latter pedaled the instrument. DownBeat described the album as "music beyond category", and awarded it Album of the Year in the magazine's polls that year. Weather Report's subsequent releases were creative funk-jazz works. Although some jazz purists protested against the blend of jazz and rock, many jazz innovators crossed over from the contemporary hard bop scene into fusion. As well as the electric instruments of rock (such as electric guitar, electric bass, electric piano and synthesizer keyboards), fusion also used the powerful amplification, "fuzz" pedals, wah-wah pedals and other effects that were used by 1970s-era rock bands. Notable performers of jazz fusion included Miles Davis, Eddie Harris, keyboardists Joe Zawinul, Chick Corea, and Herbie Hancock, vibraphonist Gary Burton, drummer Tony Williams, violinist Jean-Luc Ponty, guitarists Larry Coryell, Al Di Meola, John McLaughlin, Ryo Kawasaki, and Frank Zappa, saxophonist Wayne Shorter and bassists Jaco Pastorius and Stanley Clarke. Jazz fusion was also popular in Japan, where the band Casiopea released more than thirty fusion albums. According to jazz writer Stuart Nicholson, "just as free jazz appeared on the verge of creating a whole new musical language in the 1960s ... jazz-rock briefly suggested the promise of doing the same" with albums such as Williams' Emergency! (1970) and Davis' Agharta (1975), which Nicholson said "suggested the potential of evolving into something that might eventually define itself as a wholly independent genre quite apart from the sound and conventions of anything that had gone before." This development was stifled by commercialism, Nicholson said, as the genre "mutated into a peculiar species of jazz-inflected pop music that eventually took up residence on FM radio" at the end of the 1970s. Although jazz-rock fusion reached the height of its popularity in the 1970s, the use of electronic instruments and rock-derived musical elements in jazz continued in the 1990s and 2000s. Musicians using this approach include Pat Metheny, John Abercrombie, John Scofield and the Swedish group e.s.t. Since the beginning of the 1990s, electronic music had significant technical improvements that popularized and created new possibilities for the genre. Jazz elements such as improvisation, rhythmic complexities and harmonic textures were introduced to the genre and consequently had a big impact in new listeners and in some ways kept the versatility of jazz relatable to a newer generation that did not necessarily relate to what the traditionalists call real jazz (bebop, cool and modal jazz). Artists such as Squarepusher, Aphex Twin, Flying Lotus and sub genres like IDM, drum 'n' bass, jungle and techno ended up incorporating a lot of these elements. Squarepusher being cited as one big influence for jazz performers drummer Mark Guiliana and pianist Brad Mehldau, showing the correlations between jazz and electronic music are a two-way street. By the mid-1970s, the sound known as jazz-funk had developed, characterized by a strong back beat (groove), electrified sounds and, often, the presence of electronic analog synthesizers. Jazz-funk also draws influences from traditional African music, Afro-Cuban rhythms and Jamaican reggae, notably Kingston bandleader Sonny Bradshaw. Another feature is the shift of emphasis from improvisation to composition: arrangements, melody and overall writing became important. The integration of funk, soul, and R&B music into jazz resulted in the creation of a genre whose spectrum is wide and ranges from strong jazz improvisation to soul, funk or disco with jazz arrangements, jazz riffs and jazz solos, and sometimes soul vocals. Early examples are Herbie Hancock's Headhunters band and Miles Davis' On the Corner album, which, in 1972, began Davis' foray into jazz-funk and was, he claimed, an attempt at reconnecting with the young black audience which had largely forsaken jazz for rock and funk. While there is a discernible rock and funk influence in the timbres of the instruments employed, other tonal and rhythmic textures, such as the Indian tambora and tablas and Cuban congas and bongos, create a multi-layered soundscape. The album was a culmination of sorts of the musique concrète approach that Davis and producer Teo Macero had begun to explore in the late 1960s. The 1980s saw something of a reaction against the fusion and free jazz that had dominated the 1970s. Trumpeter Wynton Marsalis emerged early in the decade, and strove to create music within what he believed was the tradition, rejecting both fusion and free jazz and creating extensions of the small and large forms initially pioneered by artists such as Louis Armstrong and Duke Ellington, as well as the hard bop of the 1950s. It is debatable whether Marsalis' critical and commercial success was a cause or a symptom of the reaction against Fusion and Free Jazz and the resurgence of interest in the kind of jazz pioneered in the 1960s (particularly modal jazz and post-bop); nonetheless there were many other manifestations of a resurgence of traditionalism, even if fusion and free jazz were by no means abandoned and continued to develop and evolve. For example, several musicians who had been prominent in the fusion genre during the 1970s began to record acoustic jazz once more, including Chick Corea and Herbie Hancock. Other musicians who had experimented with electronic instruments in the previous decade had abandoned them by the 1980s; for example, Bill Evans, Joe Henderson, and Stan Getz. Even the 1980s music of Miles Davis, although certainly still fusion, adopted a far more accessible and recognizably jazz-oriented approach than his abstract work of the mid-1970s, such as a return to a theme-and-solos approach. A similar reaction took place against free jazz. According to Ted Gioia: the very leaders of the avant garde started to signal a retreat from the core principles of free jazz. Anthony Braxton began recording standards over familiar chord changes. Cecil Taylor played duets in concert with Mary Lou Williams, and let her set out structured harmonies and familiar jazz vocabulary under his blistering keyboard attack. And the next generation of progressive players would be even more accommodating, moving inside and outside the changes without thinking twice. Musicians such as David Murray or Don Pullen may have felt the call of free-form jazz, but they never forgot all the other ways one could play African-American music for fun and profit. Pianist Keith Jarrett—whose bands of the 1970s had played only original compositions with prominent free jazz elements—established his so-called 'Standards Trio' in 1983, which, although also occasionally exploring collective improvisation, has primarily performed and recorded jazz standards. Chick Corea similarly began exploring jazz standards in the 1980s, having neglected them for the 1970s. In 1987, the United States House of Representatives and Senate passed a bill proposed by Democratic Representative John Conyers Jr. to define jazz as a unique form of American music, stating "jazz is hereby designated as a rare and valuable national American treasure to which we should devote our attention, support and resources to make certain it is preserved, understood and promulgated." It passed in the House on September 23, 1987, and in the Senate on November 4, 1987. In 2001, Ken Burns's documentary Jazz premiered on PBS, featuring Wynton Marsalis and other experts reviewing the entire history of American jazz to that time. It received some criticism, however, for its failure to reflect the many distinctive non-American traditions and styles in jazz that had developed, and its limited representation of US developments in the last quarter of the 20th century. The emergence of young jazz talent beginning to perform in older, established musicians' groups further impacted the resurgence of traditionalism in the jazz community. In the 1970s, the groups of Betty Carter and Art Blakey and the Jazz Messengers retained their conservative jazz approaches in the midst of fusion and jazz-rock, and in addition to difficulty booking their acts, struggled to find younger generations of personnel to authentically play traditional styles such as hard bop and bebop. In the late 1970s, however, a resurgence of younger jazz players in Blakey's band began to occur. This movement included musicians such as Valery Ponomarev and Bobby Watson, Dennis Irwin and James Williams. In the 1980s, in addition to Wynton and Branford Marsalis, the emergence of pianists in the Jazz Messengers such as Donald Brown, Mulgrew Miller, and later, Benny Green, bassists such as Charles Fambrough, Lonnie Plaxico (and later, Peter Washington and Essiet Essiet) horn players such as Bill Pierce, Donald Harrison and later Javon Jackson and Terence Blanchard emerged as talented jazz musicians, all of whom made significant contributions in the 1990s and 2000s. The young Jazz Messengers' contemporaries, including Roy Hargrove, Marcus Roberts, Wallace Roney and Mark Whitfield were also influenced by Wynton Marsalis's emphasis toward jazz tradition. These younger rising stars rejected avant-garde approaches and instead championed the acoustic jazz sound of Charlie Parker, Thelonious Monk and early recordings of the first Miles Davis quintet. This group of "Young Lions" sought to reaffirm jazz as a high art tradition comparable to the discipline of classical music. In addition, Betty Carter's rotation of young musicians in her group foreshadowed many of New York's preeminent traditional jazz players later in their careers. Among these musicians were Jazz Messenger alumni Benny Green, Branford Marsalis and Ralph Peterson Jr., as well as Kenny Washington, Lewis Nash, Curtis Lundy, Cyrus Chestnut, Mark Shim, Craig Handy, Greg Hutchinson and Marc Cary, Taurus Mateen and Geri Allen. O.T.B. ensemble included a rotation of young jazz musicians such as Kenny Garrett, Steve Wilson, Kenny Davis, Renee Rosnes, Ralph Peterson Jr., Billy Drummond, and Robert Hurst. Starting in the 1990s, a number of players from largely straight-ahead or post-bop backgrounds emerged as a result of the rise of neo-traditionalist jazz, including pianists Jason Moran and Vijay Iyer, guitarist Kurt Rosenwinkel, vibraphonist Stefon Harris, trumpeters Roy Hargrove and Terence Blanchard, saxophonists Chris Potter and Joshua Redman, clarinetist Ken Peplowski and bassist Christian McBride. In the early 1980s, a commercial form of jazz fusion called "pop fusion" or "smooth jazz" became successful, garnering significant radio airplay in "quiet storm" time slots at radio stations in urban markets across the U.S. This helped to establish or bolster the careers of vocalists including Al Jarreau, Anita Baker, Chaka Khan, and Sade, as well as saxophonists including Grover Washington Jr., Kenny G, Kirk Whalum, Boney James, and David Sanborn. In general, smooth jazz is downtempo (the most widely played tracks are of 90–105 beats per minute), and has a lead melody-playing instrument (saxophone, especially soprano and tenor, and legato electric guitar are popular). In his Newsweek article "The Problem With Jazz Criticism", Stanley Crouch considers Miles Davis' playing of fusion to be a turning point that led to smooth jazz. Critic Aaron J. West has countered the often negative perceptions of smooth jazz, stating: I challenge the prevalent marginalization and malignment of smooth jazz in the standard jazz narrative. Furthermore, I question the assumption that smooth jazz is an unfortunate and unwelcomed evolutionary outcome of the jazz-fusion era. Instead, I argue that smooth jazz is a long-lived musical style that merits multi-disciplinary analyses of its origins, critical dialogues, performance practice, and reception. Acid jazz developed in the UK in the 1980s and 1990s, influenced by jazz-funk and electronic dance music. Acid jazz often contains various types of electronic composition (sometimes including sampling or live DJ cutting and scratching), but it is just as likely to be played live by musicians, who often showcase jazz interpretation as part of their performance. Richard S. Ginell of AllMusic considers Roy Ayers "one of the prophets of acid jazz". Nu jazz is influenced by jazz harmony and melodies, and there are usually no improvisational aspects. It can be very experimental in nature and can vary widely in sound and concept. It ranges from the combination of live instrumentation with the beats of jazz house (as exemplified by St Germain, Jazzanova, and Fila Brazillia) to more band-based improvised jazz with electronic elements (for example, The Cinematic Orchestra, Kobol and the Norwegian "future jazz" style pioneered by Bugge Wesseltoft, Jaga Jazzist, and Nils Petter Molvær). Jazz rap developed in the late 1980s and early 1990s and incorporates jazz influences into hip hop. In 1988, Gang Starr released the debut single "Words I Manifest", which sampled Dizzy Gillespie's 1962 "Night in Tunisia", and Stetsasonic released "Talkin' All That Jazz", which sampled Lonnie Liston Smith. Gang Starr's debut LP No More Mr. Nice Guy (1989) and their 1990 track "Jazz Thing" sampled Charlie Parker and Ramsey Lewis. The groups which made up the Native Tongues Posse tended toward jazzy releases: these include the Jungle Brothers' debut Straight Out the Jungle (1988), and A Tribe Called Quest's People's Instinctive Travels and the Paths of Rhythm (1990) and The Low End Theory (1991). Rap duo Pete Rock & CL Smooth incorporated jazz influences on their 1992 debut Mecca and the Soul Brother. Rapper Guru's Jazzmatazz series began in 1993 using jazz musicians during the studio recordings. Although jazz rap had achieved little mainstream success, Miles Davis' final album Doo-Bop (released posthumously in 1992) was based on hip hop beats and collaborations with producer Easy Mo Bee. Davis' ex-bandmate Herbie Hancock also absorbed hip-hop influences in the mid-1990s, releasing the album Dis Is Da Drum in 1994. The mid-2010s saw an increased influence of R&B, hip-hop, and pop music on jazz. In 2015, Kendrick Lamar released his third studio album, To Pimp a Butterfly. The album heavily featured prominent contemporary jazz artists such as Thundercat and redefined jazz rap with a larger focus on improvisation and live soloing rather than simply sampling. In that same year, saxophonist Kamasi Washington released his nearly three-hour long debut, The Epic. Its hip-hop inspired beats and R&B vocal interludes was not only acclaimed by critics for being innovative in keeping jazz relevant, but also sparked a small resurgence in jazz on the internet. The relaxation of orthodoxy which was concurrent with post-punk in London and New York City led to a new appreciation of jazz. In London, the Pop Group began to mix free jazz and dub reggae into their brand of punk rock. In New York, No Wave took direct inspiration from both free jazz and punk. Examples of this style include Lydia Lunch's Queen of Siam, Gray, the work of James Chance and the Contortions (who mixed Soul with free jazz and punk) and the Lounge Lizards (the first group to call themselves "punk jazz"). John Zorn took note of the emphasis on speed and dissonance that was becoming prevalent in punk rock, and incorporated this into free jazz with the release of the Spy vs. Spy album in 1986, a collection of Ornette Coleman tunes done in the contemporary thrashcore style. In the same year, Sonny Sharrock, Peter Brötzmann, Bill Laswell, and Ronald Shannon Jackson recorded the first album under the name Last Exit, a similarly aggressive blend of thrash and free jazz. These developments are the origins of jazzcore, the fusion of free jazz with hardcore punk. The M-Base movement started in the 1980s, when a loose collective of young African-American musicians in New York which included Steve Coleman, Greg Osby, and Gary Thomas developed a complex but grooving sound. In the 1990s, most M-Base participants turned to more conventional music, but Coleman, the most active participant, continued developing his music in accordance with the M-Base concept. Coleman's audience decreased, but his music and concepts influenced many musicians, according to pianist Vijay Iver and critic Ben Ratlifff of The New York Times. M-Base changed from a movement of a loose collective of young musicians to a kind of informal Coleman "school", with a much advanced but already originally implied concept. Steve Coleman's music and M-Base concept gained recognition as "next logical step" after Charlie Parker, John Coltrane, and Ornette Coleman. Since the 1990s, jazz has been characterized by a pluralism in which no one style dominates, but rather a wide range of styles and genres are popular. Individual performers often play in a variety of styles, sometimes in the same performance. Pianist Brad Mehldau and The Bad Plus have explored contemporary rock music within the context of the traditional jazz acoustic piano trio, recording instrumental jazz versions of songs by rock musicians. The Bad Plus have also incorporated elements of free jazz into their music. A firm avant-garde or free jazz stance has been maintained by some players, such as saxophonists Greg Osby and Charles Gayle, while others, such as James Carter, have incorporated free jazz elements into a more traditional framework. Harry Connick Jr. began his career playing stride piano and the Dixieland jazz of his home, New Orleans, beginning with his first recording when he was 10 years old. Some of his earliest lessons were at the home of pianist Ellis Marsalis. Connick had success on the pop charts after recording the soundtrack to the movie When Harry Met Sally, which sold over two million copies. Crossover success has also been achieved by Diana Krall, Norah Jones, Cassandra Wilson, Kurt Elling, and Jamie Cullum. Additionally, the era saw the release of recordings and videos from the previous century, such as a Just Jazz tape broadcast by a band led by Gene Ammons and studio archives such as Just Coolin' by Art Blakey and the Jazz Messengers. An internet-aided trend of 2010's jazz was that of extreme reharmonization, inspired by both virtuosic players known for their speed and rhythm such as Art Tatum, as well as players known for their ambitious voicings and chords such as Bill Evans. Supergroup Snarky Puppy adopted this trend, allowing players like Cory Henry to shape the grooves and harmonies of modern jazz soloing. YouTube phenomenon Jacob Collier also gained recognition for his ability to play an incredibly large number of instruments and his ability to use microtones, advanced polyrhythms, and blend a spectrum of genres in his largely homemade production process. Other jazz musicians gained popularity through social media during the 2010s and 2020s. These included Joan Chamorro, a bassist and bandleader based in Barcelona whose big band and jazz combo videos have received tens of millions of views on YouTube, and Emmet Cohen, who broadcast a series of performances live from New York starting in March 2020.
{{hatnote|For the [[slang]] term, see [[:wikt:beef|Beef<sup> (n,2)</sup>]] in Wiktionary.}} [[File:Standing-rib-roast.jpg|thumb|An uncooked rib roast]] '''Beef''' is a type of [[meat]] that comes from [[cattle]]. Beef could also include meats from other [[bovine]]s. There are different types of beef. Beef is popular in [[Canada]]. {{wiktionary}} == Types of beef == * [[Ground beef]] * [[Steak]] * [[Corned beef]] * [[Steak rump]] Beef can also be in different types of grading such as: * [[Prime Beef|Prime]] * [[Choice Beef|Choice]] * [[Select Beef|Select]] {{Meat}} [[Category:Beef| ]] {{Food-stub}}
Beef is the culinary name for meat from cattle (Bos taurus). Beef can be prepared in various ways; cuts are often used for steak, which can be cooked to varying degrees of doneness, while trimmings are often ground or minced, as found in most hamburgers. Beef contains protein, iron, and vitamin B12. Along with other kinds of red meat, high consumption is associated with an increased risk of colorectal cancer and coronary heart disease, especially when processed. Beef has a high environmental impact, being a primary driver of deforestation with the highest greenhouse gas emissions of any agricultural product. In prehistoric times, humankind hunted aurochs and later domesticated them. Since that time, numerous breeds of cattle have been bred specifically for the quality or quantity of their meat. Today, beef is the third most widely consumed meat in the world, after pork and poultry. As of 2018, the United States, Brazil, and China were the largest producers of beef. Some religions and cultures prohibit beef consumption, especially Indic religions like Hinduism. Buddhists and Sikhs are also against animal slaughtering, but they do not have a wrongful eating doctrine. The word beef is from the Latin word bōs, in contrast to cow which is from Middle English cou (both words have the same Indo-European root *gʷou-). After the Norman Conquest, the French-speaking nobles who ruled England naturally used French words to refer to the meats they were served. Thus, various Anglo-Saxon words were used for the animal (such as nēat, or cu for adult females) by the peasants, but the meat was called boef (ox) (Modern French bœuf) by the French nobles — who did not often deal with the live animal — when it was served to them. This is one example of the common English dichotomy between the words for animals (with largely Germanic origins) and their meat (with Romanic origins) that is also found in such English word-pairs as pig/pork, deer/venison, sheep/mutton and chicken/poultry (also the less common goat/chevon). Beef is cognate with bovine through the Late Latin bovīnus. The rarely used plural form of beef is beeves. People have eaten the flesh of bovines since prehistoric times; some of the earliest known cave paintings, such as those of Lascaux, show aurochs in hunting scenes. People domesticated cattle to provide ready access to beef, milk, and leather. Cattle have been domesticated at least twice over the course of evolutionary history. The first domestication event occurred around 10,500 years ago with the evolution of Bos taurus. The second was more recent, around 7,000 years ago, with the evolution of Bos indicus in the Indian subcontinent. There is a possible third domestication event 8,500 years ago, with a potential third species Bos africanus arising in Africa. In the United States, the growth of the beef business was largely due to expansion in the Southwest. Upon the acquisition of grasslands through the Mexican–American War of 1848, and later the expulsion of the Plains Indians from this region and the Midwest, the American livestock industry began, starting primarily with the taming of wild longhorn cattle. Chicago and New York City were the first to benefit from these developments in their stockyards and in their meat markets. Beef cattle are raised and fed using a variety of methods, including feedlots, free range, ranching, backgrounding and intensive animal farming. Concentrated Animal Feeding Operations (CAFOs), commonly referred to as factory farms, are commonly used to meet the demand of beef production. CAFOs supply 70.4% of cows in the US market and 99% of all meat in the United States supply. Cattle CAFOs can also be a source of E. coli contamination in the food supply due to the prevalence of manure in CAFOs. These E. coli contaminations include one strain, E. coli O157:H7, which can be toxic to humans, because cattle typically hold this strain in their digestive system. Another consequence of unsanitary conditions created by high-density confinement systems is increased use of antibiotics in order to prevent illness. An analysis of FDA sales data by the Natural Resources Defense Council found 42% of medically important antibiotic use in the U.S. was on cattle, posing concerns about the development of antibiotic resistant bacteria. In 2023 production was forecast to peak by 2035. The consumption of beef poses numerous threats to the natural environment. Of all agricultural products, beef requires some of the most land and water, and its production results in the greatest amount of greenhouse gas emissions (GHG), air pollution, and water pollution. A 2021 study added up GHG emissions from the entire lifecycle, including production, transportation, and consumption, and estimated that beef contributed about 4 billion tonnes (9%) of anthropogenic greenhouse gases in 2010. Cattle populations graze around 26% of all land on Earth, not including the large agricultural fields that are used to grow cattle feed. According to FAO, "Ranching-induced deforestation is one of the main causes of loss of some unique plant and animal species in the tropical rainforests of Central and South America as well as carbon release in the atmosphere." Beef is also the primary driver of deforestation in the Amazon, with around 80% of all converted land being used to rear cattle. 91% of Amazon land deforested since 1970 has been converted to cattle ranching. 41% of global deforestation from 2005 to 2013 has been attributed to the expansion of beef production. This is due to the higher ratio of net energy of gain to net energy of maintenance where metabolizable energy intake is higher. The ratio of feed required to produce an equivalent amount of beef (live weight) has been estimated at 7:1 to 43:1, compared with about 2:1 for chicken. However, assumptions about feed quality are implicit in such generalizations. For example, production of a kilogram of beef cattle live weight may require between 4 and 5 kilograms of feed high in protein and metabolizable energy content, or more than 20 kilograms of feed of much lower quality. A simple exchange of beef to soy beans (a common feed source for cattle) in Americans' diets would, according to one estimate, result in meeting between 46 and 74 percent of the reductions needed to meet the 2020 greenhouse gas emission goals of the United States as pledged in 2009. A 2021 CSIRO trial concluded that feeding cattle a 3% diet of the seaweed Asparagopsis taxiformis could reduce the methane component of their emissions by 80%. While such feed options are still experimental, even when looking at the most widely used feeds around the globe, there is high variability in efficiency. One study found that shifting compositions of current feeds, production areas, and informed land restoration could enable greenhouse gas emissions reductions of 34–85% annually (612–1,506 MtCO2e yr−1) without increasing costs to global beef production. Some scientists claim that the demand for beef is contributing to significant biodiversity loss as it is a significant driver of deforestation and habitat destruction; species-rich habitats, such as significant portions of the Amazon region, are being converted to agriculture for meat production. The 2019 IPBES Global Assessment Report on Biodiversity and Ecosystem Services also concurs that the beef industry plays a significant role in biodiversity loss. Around 25% to nearly 40% of global land surface is being used for livestock farming, which is mostly cattle. Some kinds of beef may receive special certifications or designations based on criteria including their breed (Certified Angus Beef, Certified Hereford Beef), origin (Kobe beef, Carne de Ávila, Belgian Blue), or the way the cattle are treated, fed or slaughtered (organic, grass-fed, Kosher, or Halal beef). Some countries regulate the marketing and sale of beef by observing criteria post-slaughter and classifying the observed quality of the meat. In 2018, the United States, Brazil, and China produced the most beef with 12.22 million tons, 9.9 million tons, and 6.46 million tons respectively. The top 3 beef exporting countries in 2019 were Australia (14.8% of total exports), the United States (13.4% of total exports), and Brazil (12.6% of total exports). Beef production is also important to the economies of Japan, Argentina, Uruguay, Canada, Paraguay, Mexico, Belarus and Nicaragua. As per 2020, Brazil was the largest beef exporter in the world followed by Australia, United States, India (Includes Carabeef only) and Argentina. Brazil, Australia, the United States and India accounted for roughly 61% of the world's beef exports. The world produced 60.57 million metric tons of beef in 2020, down 950K metric tons from the prior year. Major decline for production of beef was from India up to 510k and Australia down to 309K metric tons from the prior year. Most beef can be used as is by merely cutting into certain parts, such as roasts, short ribs or steak (filet mignon, sirloin steak, rump steak, rib steak, rib eye steak, hanger steak, etc.), while other cuts are processed (corned beef or beef jerky). Trimmings, on the other hand, which are usually mixed with meat from older, leaner (therefore tougher) cattle, are ground, minced or used in sausages. The blood is used in some varieties called blood sausage. Other parts that are eaten include other muscles and offal, such as the oxtail, liver, tongue, tripe from the reticulum or rumen, glands (particularly the pancreas and thymus, referred to as sweetbread), the heart, the brain (although forbidden where there is a danger of bovine spongiform encephalopathy, BSE, commonly referred to as mad cow disease), the kidneys, and the tender testicles of the bull (known in the United States as calf fries, prairie oysters, or Rocky Mountain oysters). Some intestines are cooked and eaten as is, but are more often cleaned and used as natural sausage casings. The bones are used for making beef stock. Meat from younger cows (calves) is called veal. Beef from steers and heifers is similar. Beef is first divided into primal cuts, large pieces of the animal initially separated by butchering. These are basic sections from which steaks and other subdivisions are cut. The term "primal cut" is quite different from "prime cut", used to characterize cuts considered to be of higher quality. Since the animal's legs and neck muscles do the most work, they are the toughest; the meat becomes more tender as distance from hoof and horn increases. Different countries and cuisines have different cuts and names, and sometimes use the same name for a different cut; for example, the cut described as "brisket" in the United States is from a significantly different part of the carcass than British brisket. To improve tenderness of beef, it is often aged (i.e., stored refrigerated) to allow endogenous proteolytic enzymes to weaken structural and myofibrillar proteins. Wet aging is accomplished using vacuum packaging to reduce spoilage and yield loss. Dry aging involves hanging primals (usually ribs or loins) in humidity-controlled coolers. Outer surfaces dry out and can support growth of molds (and spoilage bacteria, if too humid), resulting in trim and evaporative losses. Evaporation concentrates the remaining proteins and increases flavor intensity; the molds can contribute a nut-like flavor. After two to three days there are significant effects. The majority of the tenderizing effect occurs in the first 10 days. Boxed beef, stored and distributed in vacuum packaging, is, in effect, wet aged during distribution. Premium steakhouses dry age for 21 to 28 days or wet age up to 45 days for maximum effect on flavor and tenderness. Meat from less tender cuts or older cattle can be mechanically tenderized by forcing small, sharp blades through the cuts to disrupt the proteins. Also, solutions of exogenous proteolytic enzymes (papain, bromelin or ficin) can be injected to augment the endogenous enzymes. Similarly, solutions of salt and sodium phosphates can be injected to soften and swell the myofibrillar proteins. This improves juiciness and tenderness. Salt can improve the flavor, but phosphate can contribute a soapy flavor. These methods are applicable to all types of meat and some other foodstuffs. Beef can be cooked to various degrees, from very rare to well done. The degree of cooking corresponds to the temperature in the approximate center of the meat, which can be measured with a meat thermometer. Beef can be cooked using the sous-vide method, which cooks the entire steak to the same temperature, but when cooked using a method such as broiling or roasting it is typically cooked such that it has a "bulls eye" of doneness, with the least done (coolest) at the center and the most done (warmest) at the outside. Meat can be cooked in boiling oil, typically by shallow frying, although deep frying may be used, often for meat enrobed with breadcrumbs as in milanesas or finger steaks. Larger pieces such as steaks may be cooked this way, or meat may be cut smaller as in stir frying, typically an Asian way of cooking: cooking oil with flavorings such as garlic, ginger and onions is put in a very hot wok. Then small pieces of meat are added, followed by ingredients which cook more quickly, such as mixed vegetables. The dish is ready when the ingredients are 'just cooked'. Moist heat cooking methods include braising, pot roasting, stewing and sous-vide. These techniques are often used for cuts of beef that are tougher, as these longer, lower-temperature cooking methods have time to dissolve connecting tissue which otherwise makes meat remain tough after cooking. Meat has usually been cooked in water which is just simmering, such as in stewing; higher temperatures make meat tougher by causing the proteins to contract. Since thermostatic temperature control became available, cooking at temperatures well below boiling, 52 °C (126 °F) (sous-vide) to 90 °C (194 °F) (slow cooking), for prolonged periods has become possible; this is just hot enough to convert the tough collagen in connective tissue into gelatin through hydrolysis, with minimal toughening. With the adequate combination of temperature and cooking time, pathogens, such as bacteria will be killed, and pasteurization can be achieved. Because browning (Maillard reactions) can only occur at higher temperatures (above the boiling point of water), these moist techniques do not develop the flavors associated with browning. Meat will often undergo searing in a very hot pan, grilling or browning with a torch before moist cooking (though sometimes after). Thermostatically controlled methods, such as sous-vide, can also prevent overcooking by bringing the meat to the exact degree of doneness desired, and holding it at that temperature indefinitely. The combination of precise temperature control and long cooking duration makes it possible to be assured that pasteurization has been achieved, both on the surface and the interior of even very thick cuts of meat, which can not be assured with most other cooking techniques. (Although extremely long-duration cooking can break down the texture of the meat to an undesirable degree.) Beef can be cooked quickly at the table through several techniques. In hot pot cooking, such as shabu-shabu, very thinly sliced meat is cooked by the diners at the table by immersing it in a heated pot of water or stock with vegetables. In fondue bourguignonne, diners dip small pieces of beef into a pot of hot oil at the table. Both techniques typically feature accompanying flavorful sauces to complement the meat. Steak tartare is a French dish made from finely chopped or ground (minced) raw meat (often beef). More accurately, it is scraped so as not to let even the slightest of the sinew fat get into the scraped meat. It is often served with onions, capers, seasonings such as fresh ground pepper and Worcestershire sauce, and sometimes raw egg yolk. The Belgian or Dutch dish filet américain is also made of finely chopped ground beef, though it is seasoned differently, and either eaten as a main dish or can be used as a dressing for a sandwich. Kibbeh nayyeh is a similar Lebanese and Syrian dish. And in Ethiopia, a ground raw meat dish called tire siga or kitfo is eaten (upon availability). Carpaccio of beef is a thin slice of raw beef dressed with olive oil, lemon juice and seasoning. Often, the beef is partially frozen before slicing to allow very thin slices to be cut. Yukhoe is a variety of hoe, raw dishes in Korean cuisine which is usually made from raw ground beef seasoned with various spices or sauces. The beef part used for yukhoe is tender rump steak. For the seasoning, soy sauce, sugar, salt, sesame oil, green onion, and ground garlic, sesame seed, black pepper and juice of bae (Korean pear) are used. The beef is mostly topped with the yolk of a raw egg. Bresaola is an air-dried, salted beef that has been aged about two to three months until it becomes hard and a dark red, almost purple, colour. It is lean, has a sweet, musty smell and is tender. It originated in Valtellina, a valley in the Alps of northern Italy's Lombardy region. Bündnerfleisch is a similar product from neighbouring Switzerland. Chipped beef is an American industrially produced air-dried beef product, described by one of its manufacturers as being "similar to bresaola, but not as tasty." Beef jerky is dried, salted, smoked beef popular in the United States. Biltong is a cured, salted, air dried beef popular in South Africa. Pastrami is often made from beef; raw beef is salted, then partly dried and seasoned with various herbs and spices, and smoked. Corned beef is a cut of beef cured or pickled in a seasoned brine. The corn in corned beef refers to the grains of coarse salts (known as corns) used to cure it. The term corned beef can denote different styles of brine-cured beef, depending on the region. Some, like American-style corned beef, are highly seasoned and often considered delicatessen fare. Spiced beef is a cured and salted joint of round, topside, or silverside, traditionally served at Christmas in Ireland. It is a form of salt beef, cured with spices and saltpetre, intended to be boiled or broiled in Guinness or a similar stout, and then optionally roasted for a period after. There are various other recipes for pickled beef. Sauerbraten is a German variant. Beef is the third most widely consumed meat in the world, accounting for about 25% of meat production worldwide, after pork and poultry at 38% and 30% respectively. Beef is a source of complete protein and it is a rich source (20% or more of the Daily Value, DV) of Niacin, Vitamin B12, iron and zinc. Red meat is the most significant dietary source of carnitine and, like any other meat (pork, fish, veal, lamb etc.), is a source of creatine. Creatine is converted to creatinine during cooking. Consumption of red meat, and especially processed red meat, is known to increase the risk of bowel cancer and some other cancers. A 2010 meta-analysis found that processed red meat (and all processed meat) was correlated with a higher risk of coronary heart disease, although based on studies that separated the two, this meta-analysis found that red meat intake was not associated with higher incidence of coronary heart disease. As of 2020, there is substantial evidence for a link between high consumption of red meat and coronary heart disease. Some cattle raised in the United States feed on pastures fertilized with sewage sludge. Elevated dioxins may be present in meat from these cattle. Ground beef has been subject to recalls in the United States, due to Escherichia coli (E. coli) contamination: In 1984, the use of meat and bone meal in cattle feed resulted in the world's first outbreak of bovine spongiform encephalopathy (BSE or, colloquially, mad cow disease) in the United Kingdom. Since then, other countries have had outbreaks of BSE: In 2010, the EU, through the European Food Safety Authority (EFSA), proposed a roadmap to gradually lift the restrictions on the feed ban. In 2013, the ban on feeding mammal-based products to cattle, was amended to allow for certain milk, fish, eggs, and plant-fed farm animal products to be used. Most Indic religions reject the killing and eating of cows. Hinduism prohibits cow beef known as Go-Maans in Hindi. Bovines have a sacred status in India especially the cow, due to their provision of sustenance for families. Bovines are generally considered to be integral to the landscape. However, they do not consider the cow to be a god. Many of India's rural economies depend on cattle farming; hence they have been revered in society. Since the Vedic period, cattle, especially cows, were venerated as a source of milk, and dairy products, and their relative importance in transport services and farming like ploughing, row planting, ridging. Veneration grew with the advent of Jainism and the Gupta period. In medieval India, Maharaja Ranjit Singh issued a proclamation on stopping cow slaughter. Conflicts over cow slaughter often have sparked religious riots that have led to loss of human life and in one 1893 riot alone, more than 100 people were killed for the cause. For religious reasons, the ancient Egyptian priests also refrained from consuming beef. Buddhists and Sikhs are also against wrongful slaughtering of animals, but they do not have a wrongful eating doctrine. In ancient China, the killing of cattle and consumption of beef was prohibited, as they were valued for their role in agriculture. This custom is still followed by a few Chinese families across the world. During the season of Lent, Orthodox Christians and Catholics periodically give up meat and poultry (and sometimes dairy products and eggs) as a religious act. Observant Jews and Muslims may not eat any meat or poultry which has not been slaughtered and treated in conformance with religious laws. Most of the North Indian states prohibit the killing of cow and consumption of beef for religious reasons. Certain Hindu castes and sects continue to avoid beef from their diets. Article 48 of the Constitution of India mandates the state may take steps for preserving and improving the bovine breeds, and prohibit the slaughter, of cows and calves and other milch and draught cattle. Article 47 of the Constitution of India provides states must raise the level of nutrition and the standard of living and to improve public health as among its primary duties, based on this a reasonableness in slaughter of common cattle was instituted, if the animals ceased to be capable of breeding, providing milk, or serving as draught animals. The overall mismanagement of India's common cattle is dubbed in academic fields as "India's bovine burden." In 2017, a rule against the slaughter of cattle and the eating of beef was signed into law by presidential assent as a modified version of Prevention of Cruelty to Animals Act, 1960. The original act, however, did permit the humane slaughter of animals for use as food. Existing meat export policy in India prohibits the export of beef (meat of cow, oxen and calf). Bone-in meat, a carcass, or half carcass of buffalo is also prohibited from export. Only the boneless meat of buffalo, meat of goat and sheep and birds is permitted for export. In 2017, India sought a total "beef ban" and Australian market analysts predicted that this would create market opportunities for leather traders and meat producers there and elsewhere. Their prediction estimated a twenty percent shortage of beef and a thirteen percent shortage of leather in the world market. The cow is the national animal of Nepal, and slaughter of cattle is prohibited by law. In 2003, Cuba banned cow slaughter due to severe shortage of milk and milk products. On 14 April 2021, the ban was loosened, allowing ranchers to do as they wish as long as state quotas were met and the health of the herd could be ensured.
[[File:Court ladies pounding silk from a painting (捣练图) by Emperor Huizong.jpg|thumb|Court ladies pounding silk]] [[File:Meyers b14 s0826a.jpg|thumb|Four of the most important domesticated silk worms, together with their adult moth forms, [[Meyers Konversations-Lexikon]] (1885-1892)]] '''Silk''' is a natural [[fibre]] made by the [[silk worm]] cocoon. Silk fibres are very strong and are often used to make [[cloth]]. The cloth from silk can be made into [[rug]]s, [[bed]]ding, or can be used to [[writing|write]] or [[painting|paint]] on. In the past, silk was used to make [[parachute]]s. The practice of growing silkworms for silk production is called ''sericulture''. Most [[spider]]s make a natural fibre of their own that is also called silk. == History == The practice of breeding silkworms, known as [[sericulture]], dates back to around 5000 BC during the time of the Yellow Emperor, Huang Di. Legend says that his wife, [[Leizu|Lei Zu]], introduced silkworm rearing and invented the loom, earning her the title [[Goddess of Silk.]]<ref>{{Cite web|title=History of Silk {{!}} China Collection|url=https://china.lu/en/silk-30|access-date=2024-10-02|website=china.lu}}</ref> ==Chemical properties== Silk that is made by the silk worm is made up of two main [[protein]]s, [[sericin]] and [[fibroin]]. Fibroin is the structural center of the silk and gives it its [[Strength of materials|strength]]. It is made up of [[amino acid]]s, which make the [[fiber]]s strong and hard to break. The [[tensile strength]] is there because of [[hydrogen bond]]s. When silk is stretched, there is a strong [[Force (physics)|force]] on these many bonds, so they do not break. Serecin is the sticky material which surrounds the fibroin and gives it protection from the outside. Silk is resistant to most [[mineral acid]]s, except for [[sulfuric acid]], which dissolves it. [[Perspiration]] gives the silk a [[yellow]]ish colour. ==Related pages== * [[Silk Road]] ==References== {{commonscat|Silk}} {{reflist}} [[Category:Basic English 850 words]] [[Category:Fibers]] [[Category:Cloth]] {{stub}}
Silk is a natural protein fiber, some forms of which can be woven into textiles. The protein fiber of silk is composed mainly of fibroin and is produced by certain insect larvae to form cocoons. The best-known silk is obtained from the cocoons of the larvae of the mulberry silkworm Bombyx mori reared in captivity (sericulture). The shimmering appearance of silk is due to the triangular prism-like structure of the silk fibre, which allows silk cloth to refract incoming light at different angles, thus producing different colors. Silk is produced by several insects; but, generally, only the silk of moth caterpillars has been used for textile manufacturing. There has been some research into other types of silk, which differ at the molecular level. Silk is mainly produced by the larvae of insects undergoing complete metamorphosis, but some insects, such as webspinners and raspy crickets, produce silk throughout their lives. Silk production also occurs in hymenoptera (bees, wasps, and ants), silverfish, caddisflies, mayflies, thrips, leafhoppers, beetles, lacewings, fleas, flies, and midges. Other types of arthropods produce silk, most notably various arachnids, such as spiders. The word silk comes from Old English: sioloc, from Latin: sericum and Ancient Greek: σηρικός, romanized: sērikós, "silken", ultimately from the Chinese word "sī" and other Asian sources—compare Mandarin sī "silk", Manchurian sirghe, Mongolian sirkek. The production of silk originated in China in the Neolithic period, although it would eventually reach other places of the world (Yangshao culture, 4th millennium BC). Silk production remained confined to China until the Silk Road opened at some point during the latter part of the 1st millennium BC, though China maintained its virtual monopoly over silk production for another thousand years. Several kinds of wild silk, produced by caterpillars other than the mulberry silkworm, have been known and spun in China, South Asia, and Europe since ancient times, e.g. the production of Eri silk in Assam, India. However, the scale of production was always far smaller than for cultivated silks. There are several reasons for this: first, they differ from the domesticated varieties in colour and texture and are therefore less uniform; second, cocoons gathered in the wild have usually had the pupa emerge from them before being discovered so the silk thread that makes up the cocoon has been torn into shorter lengths; and third, many wild cocoons are covered in a mineral layer that prevents attempts to reel from them long strands of silk. Thus, the only way to obtain silk suitable for spinning into textiles in areas where commercial silks are not cultivated was by tedious and labor-intensive carding. Some natural silk structures have been used without being unwound or spun. Spider webs were used as a wound dressing in ancient Greece and Rome, and as a base for painting from the 16th century. Caterpillar nests were pasted together to make a fabric in the Aztec Empire. Commercial silks originate from reared silkworm pupae, which are bred to produce a white-colored silk thread with no mineral on the surface. The pupae are killed by either dipping them in boiling water before the adult moths emerge or by piercing them with a needle. These factors all contribute to the ability of the whole cocoon to be unravelled as one continuous thread, permitting a much stronger cloth to be woven from the silk. Wild silks also tend to be more difficult to dye than silk from the cultivated silkworm. A technique known as demineralizing allows the mineral layer around the cocoon of wild silk moths to be removed, leaving only variability in color as a barrier to creating a commercial silk industry based on wild silks in the parts of the world where wild silk moths thrive, such as in Africa and South America. Silk use in fabric was first developed in ancient China. The earliest evidence for silk is the presence of the silk protein fibroin in soil samples from two tombs at the neolithic site Jiahu in Henan, which date back about 8,500 years. The earliest surviving example of silk fabric dates from about 3630 BC, and was used as the wrapping for the body of a child at a Yangshao culture site in Qingtaicun near Xingyang, Henan. Legend gives credit for developing silk to a Chinese empress, Leizu (Hsi-Ling-Shih, Lei-Tzu). Silks were originally reserved for the Emperors of China for their own use and gifts to others, but spread gradually through Chinese culture and trade both geographically and socially, and then to many regions of Asia. Because of its texture and lustre, silk rapidly became a popular luxury fabric in the many areas accessible to Chinese merchants. Silk was in great demand, and became a staple of pre-industrial international trade. Silk was also used as a surface for writing, especially during the Warring States period (475-221 BCE). The fabric was light, it survived the damp climate of the Yangtze region, absorbed ink well, and provided a white background for the text. In July 2007, archaeologists discovered intricately woven and dyed silk textiles in a tomb in Jiangxi province, dated to the Eastern Zhou dynasty roughly 2,500 years ago. Although historians have suspected a long history of a formative textile industry in ancient China, this find of silk textiles employing "complicated techniques" of weaving and dyeing provides direct evidence for silks dating before the Mawangdui-discovery and other silks dating to the Han dynasty (202 BC – 220 AD). Silk is described in a chapter of the Fan Shengzhi shu from the Western Han (202 BC – 9 AD). There is a surviving calendar for silk production in an Eastern Han (25–220 AD) document. The two other known works on silk from the Han period are lost. The first evidence of the long distance silk trade is the finding of silk in the hair of an Egyptian mummy of the 21st dynasty, c.1070 BC. The silk trade reached as far as the Indian subcontinent, the Middle East, Europe, and North Africa. This trade was so extensive that the major set of trade routes between Europe and Asia came to be known as the Silk Road. The Emperors of China strove to keep knowledge of sericulture secret to maintain the Chinese monopoly. Nonetheless, sericulture reached Korea with technological aid from China around 200 BC, the ancient Kingdom of Khotan by AD 50, and India by AD 140. In the ancient era, silk from China was the most lucrative and sought-after luxury item traded across the Eurasian continent, and many civilizations, such as the ancient Persians, benefited economically from trade. Silk has a long history in India. It is known as Resham in eastern and north India, and Pattu in southern parts of India. Recent archaeological discoveries in Harappa and Chanhu-daro suggest that sericulture, employing wild silk threads from native silkworm species, existed in South Asia during the time of the Indus Valley civilisation (now in Pakistan and India) dating between 2450 BC and 2000 BC. Shelagh Vainker, a silk expert at the Ashmolean Museum in Oxford, who sees evidence for silk production in China "significantly earlier" than 2500–2000 BC, suggests, "people of the Indus civilization either harvested silkworm cocoons or traded with people who did, and that they knew a considerable amount about silk." India is the second largest producer of silk in the world after China. About 97% of the raw mulberry silk comes from six Indian states, namely, Andhra Pradesh, Karnataka, Jammu and Kashmir, Tamil Nadu, Bihar, and West Bengal. North Bangalore, the upcoming site of a $20 million "Silk City" Ramanagara and Mysore, contribute to a majority of silk production in Karnataka. In Tamil Nadu, mulberry cultivation is concentrated in the Coimbatore, Erode, Bhagalpuri, Tiruppur, Salem, and Dharmapuri districts. Hyderabad, Andhra Pradesh, and Gobichettipalayam, Tamil Nadu, were the first locations to have automated silk reeling units in India. In the northeastern state of Assam, three different types of indigenous variety of silk are produced, collectively called Assam silk: Muga silk, Eri silk and Pat silk. Muga, the golden silk, and Eri are produced by silkworms that are native only to Assam. They have been reared since ancient times similar to other East and South-East Asian countries. Silk is produced year-round in Thailand by two types of silkworms, the cultured Bombycidae and wild Saturniidae. Most production is after the rice harvest in the southern and northeastern parts of the country. Women traditionally weave silk on hand looms and pass the skill on to their daughters, as weaving is considered to be a sign of maturity and eligibility for marriage. Thai silk textiles often use complicated patterns in various colours and styles. Most regions of Thailand have their own typical silks. A single thread filament is too thin to use on its own so women combine many threads to produce a thicker, usable fiber. They do this by hand-reeling the threads onto a wooden spindle to produce a uniform strand of raw silk. The process takes around 40 hours to produce a half kilogram of silk. Many local operations use a reeling machine for this task, but some silk threads are still hand-reeled. The difference is that hand-reeled threads produce three grades of silk: two fine grades that are ideal for lightweight fabrics, and a thick grade for heavier material. The silk fabric is soaked in extremely cold water and bleached before dyeing to remove the natural yellow coloring of Thai silk yarn. To do this, skeins of silk thread are immersed in large tubs of hydrogen peroxide. Once washed and dried, the silk is woven on a traditional hand-operated loom. The Rajshahi Division of northern Bangladesh is the hub of the country's silk industry. There are three types of silk produced in the region: mulberry, endi, and tassar. Bengali silk was a major item of international trade for centuries. It was known as Ganges silk in medieval Europe. Bengal was the leading exporter of silk between the 16th and 19th centuries. The 7th century CE murals of Afrasiyab in Samarkand, Sogdiana, show a Chinese Embassy carrying silk and a string of silkworm cocoons to the local Sogdian ruler. In the Torah, a scarlet cloth item called in Hebrew "sheni tola'at" שני תולעת – literally "crimson of the worm" – is described as being used in purification ceremonies, such as those following a leprosy outbreak (Leviticus 14), alongside cedar wood and hyssop (za'atar). Eminent scholar and leading medieval translator of Jewish sources and books of the Bible into Arabic, Rabbi Saadia Gaon, translates this phrase explicitly as "crimson silk" – חריר קרמז حرير قرمز. In Islamic teachings, Muslim men are forbidden to wear silk. Many religious jurists believe the reasoning behind the prohibition lies in avoiding clothing for men that can be considered feminine or extravagant. There are disputes regarding the amount of silk a fabric can consist of (e.g., whether a small decorative silk piece on a cotton caftan is permissible or not) for it to be lawful for men to wear, but the dominant opinion of most Muslim scholars is that the wearing of silk by men is forbidden. Modern attire has raised a number of issues, including, for instance, the permissibility of wearing silk neckties, which are masculine articles of clothing. In the Odyssey, 19.233, when Odysseus, while pretending to be someone else, is questioned by Penelope about her husband's clothing, he says that he wore a shirt "gleaming like the skin of a dried onion" (varies with translations, literal translation here) which could refer to the lustrous quality of silk fabric. Aristotle wrote of Coa vestis, a wild silk textile from Kos. Sea silk from certain large sea shells was also valued. The Roman Empire knew of and traded in silk, and Chinese silk was the most highly priced luxury good imported by them. During the reign of emperor Tiberius, sumptuary laws were passed that forbade men from wearing silk garments, but these proved ineffectual. The Historia Augusta mentions that the third-century emperor Elagabalus was the first Roman to wear garments of pure silk, whereas it had been customary to wear fabrics of silk/cotton or silk/linen blends. Despite the popularity of silk, the secret of silk-making only reached Europe around AD 550, via the Byzantine Empire. Contemporary accounts state that monks working for the emperor Justinian I smuggled silkworm eggs to Constantinople from China inside hollow canes. All top-quality looms and weavers were located inside the Great Palace complex in Constantinople, and the cloth produced was used in imperial robes or in diplomacy, as gifts to foreign dignitaries. The remainder was sold at very high prices. Italy was the most important producer of silk during the Medieval age. The first center to introduce silk production to Italy was the city of Catanzaro during the 11th century in the region of Calabria. The silk of Catanzaro supplied almost all of Europe and was sold in a large market fair in the port of Reggio Calabria, to Spanish, Venetian, Genovese, and Dutch merchants. Catanzaro became the lace capital of the world with a large silkworm breeding facility that produced all the laces and linens used in the Vatican. The city was world-famous for its fine fabrication of silks, velvets, damasks, and brocades. Another notable center was the Italian city-state of Lucca which largely financed itself through silk-production and silk-trading, beginning in the 12th century. Other Italian cities involved in silk production were Genoa, Venice, and Florence. The Piedmont area of Northern Italy became a major silk producing area when water-powered silk throwing machines were developed. The Silk Exchange in Valencia from the 15th century—where previously in 1348 also perxal (percale) was traded as some kind of silk—illustrates the power and wealth of one of the great Mediterranean mercantile cities. Silk was produced in and exported from the province of Granada, Spain, especially the Alpujarras region, until the Moriscos, whose industry it was, were expelled from Granada in 1571. Since the 15th century, silk production in France has been centered around the city of Lyon where many mechanic tools for mass production were first introduced in the 17th century. James I attempted to establish silk production in England, purchasing and planting 100,000 mulberry trees, some on land adjacent to Hampton Court Palace, but they were of a species unsuited to the silk worms, and the attempt failed. In 1732 John Guardivaglio set up a silk throwing enterprise at Logwood mill in Stockport; in 1744, Burton Mill was erected in Macclesfield; and in 1753 Old Mill was built in Congleton. These three towns remained the centre of the English silk throwing industry until silk throwing was replaced by silk waste spinning. British enterprise also established silk filature in Cyprus in 1928. In England in the mid-20th century, raw silk was produced at Lullingstone Castle in Kent. Silkworms were raised and reeled under the direction of Zoe Lady Hart Dyke, later moving to Ayot St Lawrence in Hertfordshire in 1956. During World War II, supplies of silk for UK parachute manufacture were secured from the Middle East by Peter Gaddum. Wild silk taken from the nests of native caterpillars was used by the Aztecs to make containers and as paper. Silkworms were introduced to Oaxaca from Spain in the 1530s and the region profited from silk production until the early 17th century, when the king of Spain banned export to protect Spain's silk industry. Silk production for local consumption has continued until the present day, sometimes spinning wild silk. King James I introduced silk-growing to the British colonies in America around 1619, ostensibly to discourage tobacco planting. The Shakers in Kentucky adopted the practice. The history of industrial silk in the United States is largely tied to several smaller urban centers in the Northeast region. Beginning in the 1830s, Manchester, Connecticut emerged as the early center of the silk industry in America, when the Cheney Brothers became the first in the United States to properly raise silkworms on an industrial scale; today the Cheney Brothers Historic District showcases their former mills. With the mulberry tree craze of that decade, other smaller producers began raising silkworms. This economy particularly gained traction in the vicinity of Northampton, Massachusetts and its neighboring Williamsburg, where a number of small firms and cooperatives emerged. Among the most prominent of these was the cooperative utopian Northampton Association for Education and Industry, of which Sojourner Truth was a member. Following the destructive Mill River Flood of 1874, one manufacturer, William Skinner, relocated his mill from Williamsburg to the then-new city of Holyoke. Over the next 50 years he and his sons would maintain relations between the American silk industry and its counterparts in Japan, and expanded their business to the point that by 1911, the Skinner Mill complex contained the largest silk mill under one roof in the world, and the brand Skinner Fabrics had become the largest manufacturer of silk satins internationally. Other efforts later in the 19th century would also bring the new silk industry to Paterson, New Jersey, with several firms hiring European-born textile workers and granting it the nickname "Silk City" as another major center of production in the United States. World War II interrupted the silk trade from Asia, and silk prices increased dramatically. U.S. industry began to look for substitutes, which led to the use of synthetics such as nylon. Synthetic silks have also been made from lyocell, a type of cellulose fiber, and are often difficult to distinguish from real silk (see spider silk for more on synthetic silks). In Terengganu, which is now part of Malaysia, a second generation of silkworm was being imported as early as 1764 for the country's silk textile industry, especially songket. However, since the 1980s, Malaysia is no longer engaged in sericulture but does plant mulberry trees. In Vietnamese legend, silk appeared in the first millennium AD and is still being woven today. The process of silk production is known as sericulture. The entire production process of silk can be divided into several steps which are typically handled by different entities. Extracting raw silk starts by cultivating the silkworms on mulberry leaves. Once the worms start pupating in their cocoons, these are dissolved in boiling water in order for individual long fibres to be extracted and fed into the spinning reel. To produce 1 kg of silk, 104 kg of mulberry leaves must be eaten by 3000 silkworms. It takes about 5000 silkworms to make a pure silk kimono. The major silk producers are China (54%) and India (14%). Other statistics: The environmental impact of silk production is potentially large when compared with other natural fibers. A life-cycle assessment of Indian silk production shows that the production process has a large carbon and water footprint, mainly due to the fact that it is an animal-derived fiber and more inputs such as fertilizer and water are needed per unit of fiber produced. Silk fibers from the Bombyx mori silkworm have a triangular cross section with rounded corners, 5–10 μm wide. The fibroin-heavy chain is composed mostly of beta-sheets, due to a 59-mer amino acid repeat sequence with some variations. The flat surfaces of the fibrils reflect light at many angles, giving silk a natural sheen. The cross-section from other silkworms can vary in shape and diameter: crescent-like for Anaphe and elongated wedge for tussah. Silkworm fibers are naturally extruded from two silkworm glands as a pair of primary filaments (brin), which are stuck together, with sericin proteins that act like glue, to form a bave. Bave diameters for tussah silk can reach 65 μm. See cited reference for cross-sectional SEM photographs. Silk has a smooth, soft texture that is not slippery, unlike many synthetic fibers. Silk is one of the strongest natural fibers, but it loses up to 20% of its strength when wet. It has a good moisture regain of 11%. Its elasticity is moderate to poor: if elongated even a small amount, it remains stretched. It can be weakened if exposed to too much sunlight. It may also be attacked by insects, especially if left dirty. One example of the durable nature of silk over other fabrics is demonstrated by the recovery in 1840 of silk garments from a wreck of 1782: 'The most durable article found has been silk; for besides pieces of cloaks and lace, a pair of black satin breeches, and a large satin waistcoat with flaps, were got up, of which the silk was perfect, but the lining entirely gone ... from the thread giving way ... No articles of dress of woollen cloth have yet been found.' Silk is a poor conductor of electricity and thus susceptible to static cling. Silk has a high emissivity for infrared light, making it feel cool to the touch. Unwashed silk chiffon may shrink up to 8% due to a relaxation of the fiber macrostructure, so silk should either be washed prior to garment construction, or dry cleaned. Dry cleaning may still shrink the chiffon up to 4%. Occasionally, this shrinkage can be reversed by a gentle steaming with a press cloth. There is almost no gradual shrinkage nor shrinkage due to molecular-level deformation. Natural and synthetic silk is known to manifest piezoelectric properties in proteins, probably due to its molecular structure. Silkworm silk was used as the standard for the denier, a measurement of linear density in fibers. Silkworm silk therefore has a linear density of approximately 1 den, or 1.1 dtex. Silk emitted by the silkworm consists of two main proteins, sericin and fibroin, fibroin being the structural center of the silk, and sericin being the sticky material surrounding it. Fibroin is made up of the amino acids Gly-Ser-Gly-Ala-Gly-Ala and forms beta pleated sheets. Hydrogen bonds form between chains, and side chains form above and below the plane of the hydrogen bond network. The high proportion (50%) of glycine allows tight packing. This is because glycine's R group is only a hydrogen and so is not as sterically constrained. The addition of alanine and serine makes the fibres strong and resistant to breaking. This tensile strength is due to the many interceded hydrogen bonds, and when stretched the force is applied to these numerous bonds and they do not break. Silk resists most mineral acids, except for sulfuric acid, which dissolves it. It is yellowed by perspiration. Chlorine bleach will also destroy silk fabrics. RSF is produced by chemically dissolving silkworm cocoons, leaving their molecular structure intact. The silk fibers dissolve into tiny thread-like structures known as microfibrils. The resulting solution is extruded through a small opening, causing the microfibrils to reassemble into a single fiber. The resulting material is reportedly twice as stiff as silk. Silk's absorbency makes it comfortable to wear in warm weather and while active. Its low conductivity keeps warm air close to the skin during cold weather. It is often used for clothing such as shirts, ties, blouses, formal dresses, high-fashion clothes, lining, lingerie, pajamas, robes, dress suits, sun dresses, and traditional Asian clothing. Silk is also excellent for insect-proof clothing, protecting the wearer from mosquitoes and horseflies. Fabrics that are often made from silk include satin, charmeuse, habutai, chiffon, taffeta, crêpe de chine, dupioni, noil, tussah, and shantung, among others. Silk's attractive lustre and drape makes it suitable for many furnishing applications. It is used for upholstery, wall coverings, window treatments (if blended with another fiber), rugs, bedding, and wall hangings. Silk had many industrial and commercial uses, such as in parachutes, bicycle tires, comforter filling, and artillery gunpowder bags. A special manufacturing process removes the outer sericin coating of the silk, which makes it suitable as non-absorbable surgical sutures. This process has also recently led to the introduction of specialist silk underclothing, which has been used for skin conditions including eczema. New uses and manufacturing techniques have been found for silk for making everything from disposable cups to drug delivery systems and holograms. Silk began to serve as a biomedical material for sutures in surgeries as early as the second century CE. In the past 30 years, it has been widely studied and used as a biomaterial due to its mechanical strength, biocompatibility, tunable degradation rate, ease to load cellular growth factors (for example, BMP-2), and its ability to be processed into several other formats such as films, gels, particles, and scaffolds. Silks from Bombyx mori, a kind of cultivated silkworm, are the most widely investigated silks. Silks derived from Bombyx mori are generally made of two parts: the silk fibroin fiber which contains a light chain of 25kDa and a heavy chain of 350kDa (or 390kDa) linked by a single disulfide bond and a glue-like protein, sericin, comprising 25 to 30 percentage by weight. Silk fibroin contains hydrophobic beta sheet blocks, interrupted by small hydrophilic groups. And the beta-sheets contribute much to the high mechanical strength of silk fibers, which achieves 740 MPa, tens of times that of poly(lactic acid) and hundreds of times that of collagen. This impressive mechanical strength has made silk fibroin very competitive for applications in biomaterials. Indeed, silk fibers have found their way into tendon tissue engineering, where mechanical properties matter greatly. In addition, mechanical properties of silks from various kinds of silkworms vary widely, which provides more choices for their use in tissue engineering. Most products fabricated from regenerated silk are weak and brittle, with only ≈1–2% of the mechanical strength of native silk fibers due to the absence of appropriate secondary and hierarchical structure, Biocompatibility, i.e., to what level the silk will cause an immune response, is a critical issue for biomaterials. The issue arose during its increasing clinical use. Wax or silicone is usually used as a coating to avoid fraying and potential immune responses when silk fibers serve as suture materials. Although the lack of detailed characterization of silk fibers, such as the extent of the removal of sericin, the surface chemical properties of coating material, and the process used, make it difficult to determine the real immune response of silk fibers in literature, it is generally believed that sericin is the major cause of immune response. Thus, the removal of sericin is an essential step to assure biocompatibility in biomaterial applications of silk. However, further research fails to prove clearly the contribution of sericin to inflammatory responses based on isolated sericin and sericin based biomaterials. In addition, silk fibroin exhibits an inflammatory response similar to that of tissue culture plastic in vitro when assessed with human mesenchymal stem cells (hMSCs) or lower than collagen and PLA when implant rat MSCs with silk fibroin films in vivo. Thus, appropriate degumming and sterilization will assure the biocompatibility of silk fibroin, which is further validated by in vivo experiments on rats and pigs. There are still concerns about the long-term safety of silk-based biomaterials in the human body in contrast to these promising results. Even though silk sutures serve well, they exist and interact within a limited period depending on the recovery of wounds (several weeks), much shorter than that in tissue engineering. Another concern arises from biodegradation because the biocompatibility of silk fibroin does not necessarily assure the biocompatibility of the decomposed products. In fact, different levels of immune responses and diseases have been triggered by the degraded products of silk fibroin. Biodegradability (also known as biodegradation)—the ability to be disintegrated by biological approaches, including bacteria, fungi, and cells—is another significant property of biomaterials. Biodegradable materials can minimize the pain of patients from surgeries, especially in tissue engineering, since there is no need for surgery in order to remove the implanted scaffold.. Wang et al. showed the in vivo degradation of silk via aqueous 3-D scaffolds implanted into Lewis rats. Enzymes are the means used to achieve degradation of silk in vitro. Protease XIV from Streptomyces griseus and α-chymotrypsin from bovine pancreases are two popular enzymes for silk degradation. In addition, gamma radiation, as well as cell metabolism, can also regulate the degradation of silk. Compared with synthetic biomaterials such as polyglycolides and polylactides, silk is advantageous in some aspects of biodegradation. The acidic degraded products of polyglycolides and polylactides will decrease the pH of the ambient environment and thus adversely influence the metabolism of cells, which is not an issue for silk. In addition, silk materials can retain strength over a desired period from weeks to months on an as-needed basis, by mediating the content of beta sheets. Genetic modification of domesticated silkworms has been used to alter the composition of the silk. As well as possibly facilitating the production of more useful types of silk, this may allow other industrially or therapeutically useful proteins to be made by silkworms. Silk moths lay eggs on specially prepared paper. The eggs hatch and the caterpillars (silkworms) are fed fresh mulberry leaves. After about 35 days and 4 moltings, the caterpillars are 10,000 times heavier than when hatched and are ready to begin spinning a cocoon. A straw frame is placed over the tray of caterpillars, and each caterpillar begins spinning a cocoon by moving its head in a pattern. Two glands produce liquid silk and force it through openings in the head called spinnerets. Liquid silk is coated in sericin, a water-soluble protective gum, and solidifies on contact with the air. Within 2–3 days, the caterpillar spins about one mile (1.6 km) of filament and is completely encased in a cocoon. The silk farmers then heat the cocoons to kill them, leaving some to metamorphose into moths to breed the next generation of caterpillars. Harvested cocoons are then soaked in boiling water to soften the sericin holding the silk fibers together in a cocoon shape. The fibers are then unwound to produce a continuous thread. Since a single thread is too fine and fragile for commercial use, anywhere from three to ten strands are spun together to form a single thread of silk. As the process of harvesting the silk from the cocoon kills the larvae by boiling, sericulture has been criticized by animal welfare activists, including People for the Ethical Treatment of Animals (PETA), who urge people not to buy silk items. Mahatma Gandhi was critical of silk production because of his Ahimsa (non-violent) philosophy, which led to the promotion of cotton and Ahimsa silk, a type of wild silk made from the cocoons of wild and semi-wild silk moths.
{{Infobox musical artist | name = The Beatles | image = The Beatles members at New York City in 1964.jpg | caption = The Beatles in 1964. Clockwise from top left: [[John Lennon]], [[Paul McCartney]], [[Ringo Starr]] and [[George Harrison]] | alt = A square quartered into four head shots of young men with moptop haircuts. All four wear white shirts and dark coats. | background = group_or_band | origin = [[Liverpool]], England | genre = {{hlist|<!-- Current consensus is to list ONLY Rock and pop. Please do not add other genres or sub-genres without first reaching consensus with other editors on the talk page. Thank you. -->[[Rock music|Rock]]| [[Pop music|pop]]|beat|[[psychedelic rock|psychedelia]]}} | years_active = 1960–1970<!--Please discuss on talk page before changing.--> | label = {{flatlist| * [[Parlophone]] * [[Apple Records|Apple]] * [[Capitol Records|Capitol]] }} | associated_acts = {{flatlist| * [[The Quarrymen]] * [[Tony Sheridan]] * [[Billy Preston]] * [[Plastic Ono Band]]<!-- please discuss on Talk page before adding other acts here --> }} | website = {{URL|thebeatles.com}} | past_members = * [[John Lennon]] * [[Paul McCartney]] * [[George Harrison]] * [[Ringo Starr]] See [[#Members|members section]] for others }} [[File:The Beatles and the Beatlemania.webm|thumb|261x261px|A simple video summary about The Beatles]] '''The Beatles''' were an English pop/rock band, started in [[Liverpool]], [[England]] in 1960. The members were [[John Lennon]], [[Paul McCartney]], [[George Harrison]], and [[Ringo Starr]]. Most people say they are the most [[wikt:success|successful]] and [[wikt:Influence|influential]] band in the history of [[popular music]].{{sfn|Hasted|2017|p=425}} The group were a main part of the creation of [[counterculture of the 1960s|1960s counterculture]]. They began as a skiffle band, and were influenced by 1950s American [[rock and roll]]. In their later years, the band was experimental with genres such as different types of rock, [[classical music|classical]], and Indian music. Their main [[songwriter]]s were [[Lennon–McCartney|Lennon and McCartney]]. Before The Beatles became popular, they played in clubs in Liverpool and [[Hamburg]] over three years between 1960 and 1963, with [[Stuart Sutcliffe]] playing [[bass guitar|bass]]. They went through many [[drum]]mers, including [[Pete Best]], before finally asking Ringo Starr to join in 1962. Sutcliffe also quit, meaning Paul McCartney started playing the bass instead. [[Brian Epstein]] was their manager and [[George Martin]] produced most of their music. Their first single was "[[Love Me Do]]", in late 1962. It was a hit and they became popular in the [[United Kingdom]]. As they became more popular, their popularity was named "[[Beatlemania]]". By 1964, the Beatles were worldwide stars and led the "[[British Invasion]]" of the USA. They made some of the best-selling albums of the 1960s, with twelve [[Album|studio albums]]. They broke up in 1970. John Lennon was later [[Murder of John Lennon|murdered]] in [[New York City]] in 1980 and George Harrison died of [[lung cancer]] in 2001. McCartney and Starr still make music. In 2023, they released their final single "[[Now and Then (Beatles song)|Now and Then]]".<ref>{{Cite web |last=Venta |first=Lance |date=27 October 2023 |title=740 iHeartMedia Stations To Simultaneously Debut "Last Beatles Song" |url=https://radioinsight.com/headlines/260345/740-iheartmedia-stations-to-simultaneously-debut-last-beatles-song/ |url-status=live |archive-url=https://web.archive.org/web/20231029003236/https://radioinsight.com/headlines/260345/740-iheartmedia-stations-to-simultaneously-debut-last-beatles-song/ |archive-date=29 October 2023 |access-date=29 October 2023 |website=RadioInsight}}</ref> ==History of the Beatles== Starting in 1956, John Lennon and several of his friends played in a British band called [[the Quarrymen]]. Over the next few years, the members of the band changed, and by 1960, the band was called the Beatles. They did not have their first hit until 1962. In February 1963 their song, "[[Please Please Me (song)|Please Please Me]]", reached the number 1 position on the British charts. This was the first of a record 19 British number 1 singles. They first came to the [[United States]] in 1964. They were met at the [[John F. Kennedy International Airport]] in New York City by thousands of screaming American teenagers. The Beatles were so popular that they were attacked by screaming fans everywhere they went around the world. The effect they had on their fans was known as 'Beatlemania'. The Beatles made their first live American television appearance on [[The Ed Sullivan Show]] on 9 February 1964. About 74 million viewers—about half of the American population—watched the group perform on the show. Beatles songs soon filled the top 5 places on the American top 40 chart - a record that has never been matched. After the Beatles became so popular in the US, other British bands, such as [[The Rolling Stones]], [[The Animals]], [[The Kinks]] and [[Gerry and the Pacemakers]] had songs become hits there as well. So many British bands became popular after the Beatles' success that this time became known in America as the "[[British Invasion]]". Toward the mid-1960s, The Beatles became bolder with their style of music. This largely started in 1965, with the release of the album ''[[Rubber Soul]]'', and hit a peak in 1967 with the release of ''[[Sgt. Pepper's Lonely Hearts Club Band]]'', which was named as the greatest album of all time by [[Rolling Stone]] Magazine.<ref name='"list"'>{{cite web|title=500 Greatest Albums of All Time|date=31 May 2009|url=https://www.rollingstone.com/music/lists/500-greatest-albums-of-all-time-20120531|publisher=Rolling Stone|accessdate=8 June 2013|archive-date=16 September 2013|archive-url=https://www.webcitation.org/6JfM5Dlce?url=http://www.rollingstone.com/music/lists/500-greatest-albums-of-all-time-20120531|url-status=live}}</ref> They were also named the most influential artists of all time by Rolling Stone magazine, because their music, clothing style, and attitudes shaped much of what was popular among young people in the 1960s. They were all awarded an [[Order of the British Empire|OBE]] in 1965 by [[Harold Wilson]]. He was accused of having debased and cheapened the [[honours system]]. <ref>{{Cite book|title=Harold Wilson the winner|url=https://archive.org/details/haroldwilsonwinn0000nick|last=Thomas-Symonds|first=Nick|publisher=Weidenfield & Nicolson|year=2023|isbn=9781474611961|location=London|pages=[https://archive.org/details/haroldwilsonwinn0000nick/page/n210 201]}}</ref> The Beatles became so popular that no regular concert venue was big enough for their concerts. This led to them playing the first ever stadium rock concert at Shea Stadium in America, to around 50,000 people. The Beatles stopped touring and playing live music in 1966 because they were sick of audiences screaming so loudly that their music could not be heard. They were also tired of the pressures of touring. Among other things, they were so popular that thousands of people would gather outside the hotels they stayed in day and night meaning that they could never leave their rooms unless they were playing a concert. The Beatles broke up in 1970 because of the pressures of fame and each member becoming more independent both in their personal lives and musically. In 1973 the two-disc sets "1962-1966" (the "Red Album") and "1967-1970" "(the "Blue Album") were released. These were both re-released on [[Compact disc|CD]] in 1993. ==After breaking up== The band was still very popular all over the world after they broke up. They are the best-selling music acts of all time, with sales between 600 million to 1 billion records. Their [[music]] is still important and still influences many musicians. Musicians today perform [[cover version]]s of Beatles songs, and people everywhere still listen to their music. Their song 'Yesterday' has been recorded by more artists than any other song.<ref name="cnn list">{{cite news|title=Rolling Stone's top 10 Beatles songs of all time|url=http://www.cnn.com/2010/SHOWBIZ/Music/08/25/beatles.songs.roll/index.html?_s=PM:SHOWBIZ|publisher=CNN Entertainment|accessdate=8 June 2013|date=26 August 2010|archive-date=6 March 2016|archive-url=https://web.archive.org/web/20160306103307/http://www.cnn.com/2010/SHOWBIZ/Music/08/25/beatles.songs.roll/index.html?_s=PM:SHOWBIZ|url-status=live}}</ref> It is also the song that has been played the most on radio ever. The Beatles made thirteen albums and twenty-six [[Single (music)|singles]] together. They also started their own [[record label]], [[Apple Records]]. They made two [[movie]]s, ''[[A Hard Day's Night]]'' and ''[[Help!]]'', where they appeared as [[actor]]s. Later they made ''Magical Mystery Tour'', a [[television]] special. ''[[Yellow Submarine]]'' was a [[cartoon]] movie based on their music. ''[[Let It Be]]'' showed them working on a new album. After the Beatles broke up in 1969, all four members started their own solo [[career]]s. John Lennon became a famous peace activist in the late 1960s and early 1970s. He wrote successful songs including "Give Peace a Chance", "[[Imagine (John Lennon song)|Imagine]]", and "Happy Xmas (War Is Over)". John Lennon was murdered on December 8, 1980 outside his home in New York. Part of Central Park in New York and an airport in Liverpool are named in his honor. The other three Beatles got together in the 1990s to make two new records. They used [[demo (music)|demo]] recordings of two John Lennon songs and added their own new parts. Producer [[Jeff Lynne]] helped them so all four members could appear on the songs. The songs were "Free as a Bird" and "Real Love". Both were top 5 hits in the UK in 1995 and 1996. Paul McCartney started the band [[Paul McCartney and Wings|Wings]] with his wife [[Linda McCartney|Linda]]. In 1977 his song, "Mull of Kintyre", became the biggest selling single in British history. It sold even more copies than the Beatles' singles. Paul McCartney was [[knight]]ed in 1997. George Harrison and Ringo Starr had early success as solo artists but were less successful later on. Harrison formed the group the [[Traveling Wilburys]] in the 80s with other rock legends [[Bob Dylan]], [[Roy Orbison]], [[Tom Petty]], and Jeff Lynne. Harrison died of [[lung cancer]] on November 29, 2001. Starr still tours the world with his "All Starr Band" and was [[knighted]] in 2018. ==Albums== [[File:The White Album.svg|thumb|right|The cover of ''[[The Beatles (album)|The Beatles]]'' (better known as the "White Album") ]]The Beatles made 13 very successful albums during their active years from 1960 to 1970. Listed below are the albums made during their career. * ''[[Please Please Me]]'' (1963) * ''[[With the Beatles]]'' (1963) * ''[[A Hard Day's Night]]'' (1964) * ''[[Beatles for Sale]]'' (1964) * ''[[Help!]]'' (1965) * ''[[Rubber Soul]]'' (1965) * ''[[Revolver (Beatles album)|Revolver]]'' (1966) * ''[[Sgt. Pepper's Lonely Hearts Club Band]]'' (1967) * ''[[Magical Mystery Tour]]'' (1967) (Double EP) * ''[[Yellow Submarine]]'' (1968) * ''[[The Beatles (album)|The Beatles]]'' (better known as the "White Album") (1968) * ''[[Abbey Road]]'' (1969) * ''[[Let it Be]]'' (1970) ;Compilations * ''The Beatles 1962-1966'' (1973) * ''The Beatles 1967-1970'' (1973) * ''Rock and Roll Music'' (1976) * ''Love Songs'' (1978) * ''Rarities'' (1980) * ''Reel Music'' (1982) * ''20 Greatest Hits'' (1983) * ''Past Masters'' (1988, two volumes) * ''Anthology'' (1995, three volumes) * ''1'' (2000) * ''Love'' (2006) == Movies == * ''[[A Hard Day's Night]]'' (1964), directed by [[Richard Lester]], co-starred [[Wilfrid Brambell]], [[Norman Rossington]], [[John Junkin]], [[Lionel Blair]], [[Victor Spinetti]], and [[Derek Nimmo]] * ''[[Help!]]'' (1965), directed by Lester, co-starred [[Leo McKern]], [[Eleanor Bron]], Victor Spinetti and [[Roy Kinnear]] * ''[[Magical Mystery Tour]]'' (1967), hour-long TV special, with [[Ivor Cutler]] and [[Jessie Robins]] * ''[[Yellow Submarine]]'', (1968) [[Animation|cartoon]] * ''[[Let It Be]]'' (1969), documentary film about the making of the album with the same name; producer [[George Martin]] and [[road manager]] [[Mal Evans]] appear briefly on camera, as do [[Yoko Ono]] and [[Heather McCartney]]. == Children of the Beatles == * Paul McCartney's daughter Stella McCartney has had a successful career as a clothing designer.<ref>{{Cite web |url=http://www.vogue.co.uk/biographies/080422-stella-mccartney-biography.aspx |title=Vogue UK; Stella McCartney Biography |access-date=2009-02-22 |archive-date=2010-03-04 |archive-url=https://web.archive.org/web/20100304110947/http://www.vogue.co.uk/biographies/080422-stella-mccartney-biography.aspx |url-status=live }}</ref> * John Lennon's sons [[Sean Lennon]] and [[Julian Lennon]] have had successful musical careers. Sean has been involved in a number of bands: Cibo Matto, and Dopo Yume.<ref>[http://www.fashionfollower.com/dopoyume.asp ''Hopeful Romantics:Dopo Yume Wear their Hearts on Their Sleeves''] {{Webarchive|url=https://web.archive.org/web/20070804170831/http://www.fashionfollower.com/dopoyume.asp |date=2007-08-04 }} ''www.fashionfollower.com''</ref> ==References== {{Reflist}} ==Other websites== * [http://www.beatles.com/ The Beatles homepage] * [http://www.beatles-stuff.com/ Beatles Website] {{Webarchive|url=https://web.archive.org/web/20121108051541/http://www.beatles-stuff.com/ |date=2012-11-08 }} * {{Pop Chronicles|27}} {{The Beatles}} {{Authority control}} {{DEFAULTSORT:Beatles, The}} [[Category:Musical groups disestablished in 1970]] [[Category:1960s British music groups]] [[Category:1970s disestablishments in the United Kingdom]] [[Category:1960 establishments in England]] [[Category:1970s British music groups]] [[Category:British boy bands]] [[Category:British pop rock bands]] [[Category:English pop music groups]] [[Category:English rock bands]] [[Category:Musical groups established in 1960]] [[Category:Musical groups from Liverpool]] [[Category:Musical quartets]] [[Category:The Beatles| ]] [[Category:1970 disestablishments in Europe]]
The Beatles were an English rock band formed in Liverpool in 1960, comprising John Lennon, Paul McCartney, George Harrison and (joining in 1962) Ringo Starr. They are regarded as the most influential band of all time and were integral to the development of 1960s counterculture and the recognition of popular music as an art form. Rooted in skiffle, beat, and 1950s rock 'n' roll, their sound incorporated elements of classical music and traditional pop in innovative ways. The band also explored music styles ranging from folk and Indian music to psychedelia and hard rock. As pioneers in recording, songwriting and artistic presentation, the Beatles revolutionised many aspects of the music industry and were often publicised as leaders of the era's youth and sociocultural movements. Led by primary songwriters Lennon and McCartney, the Beatles evolved from Lennon's previous group, the Quarrymen, and built their reputation by playing clubs in Liverpool and Hamburg over three years from 1960, initially with Stuart Sutcliffe playing bass. The core trio of Lennon, McCartney and Harrison, together since 1958, went through a succession of drummers, including Pete Best, before inviting Starr to join them in 1962. Manager Brian Epstein moulded them into a professional act, and producer George Martin guided and developed their recordings, greatly expanding their domestic success after they signed with EMI Records and achieved their first hit, "Love Me Do", in late 1962. As their popularity grew into the intense fan frenzy dubbed "Beatlemania", the band acquired the nickname "the Fab Four". Epstein, Martin or another member of the band's entourage was sometimes informally referred to as a "fifth Beatle". By early 1964, the Beatles were international stars and had achieved unprecedented levels of critical and commercial success. They became a leading force in Britain's cultural resurgence, ushering in the British Invasion of the United States pop market. They soon made their film debut with A Hard Day's Night (1964). A growing desire to refine their studio efforts, coupled with the challenging nature of their concert tours, led to the band's retirement from live performances in 1966. During this time, they produced records of greater sophistication, including the albums Rubber Soul (1965), Revolver (1966) and Sgt. Pepper's Lonely Hearts Club Band (1967). They also enjoyed further commercial success with The Beatles (also known as "the White Album", 1968) and Abbey Road (1969). The success of these records heralded the album era, as albums became the dominant form of record use over singles. These records also increased public interest in psychedelic drugs and Eastern spirituality and furthered advancements in electronic music, album art and music videos. In 1968, they founded Apple Corps, a multi-armed multimedia corporation that continues to oversee projects related to the band's legacy. After the group's break-up in 1970, all principal former members enjoyed success as solo artists, and some partial reunions have occurred. Lennon was murdered in 1980, and Harrison died of lung cancer in 2001. McCartney and Starr remain musically active. The Beatles are the best-selling music act of all time, with estimated sales of 600 million units worldwide. They are the most successful act in the history of the US Billboard charts, holding the record for most number-one albums on the UK Albums Chart (15), most number-one hits on the US Billboard Hot 100 chart (20), and most singles sold in the UK (21.9 million). The band received many accolades, including seven Grammy Awards, four Brit Awards, an Academy Award (for Best Original Song Score for the 1970 documentary film Let It Be) and fifteen Ivor Novello Awards. They were inducted into the Rock and Roll Hall of Fame in 1988, and each principal member was individually inducted between 1994 and 2015. In 2004 and 2011, the group topped Rolling Stone's lists of the greatest artists in history. Time magazine named them among the 20th century's 100 most important people. In November 1956, sixteen-year-old John Lennon formed a skiffle group with several friends from Quarry Bank High School in Liverpool. They briefly called themselves the Blackjacks, before changing their name to the Quarrymen after discovering that another local group were already using the name. Fifteen-year-old Paul McCartney met Lennon on 6 July 1957, and joined as a rhythm guitarist shortly after. In February 1958, McCartney invited his friend George Harrison, then aged fifteen, to watch the band. Harrison auditioned for Lennon, impressing him with his playing, but Lennon initially thought Harrison was too young. After a month's persistence, during a second meeting (arranged by McCartney), Harrison performed the lead guitar part of the instrumental song "Raunchy" on the upper deck of a Liverpool bus, and they enlisted him as lead guitarist. By January 1959, Lennon's Quarry Bank friends had left the group, and he began his studies at the Liverpool College of Art. The three guitarists, billing themselves as Johnny and the Moondogs, were playing rock and roll whenever they could find a drummer. Lennon's art school friend Stuart Sutcliffe, who had just sold one of his paintings and was persuaded to purchase a bass guitar with the proceeds, joined in January 1960. He suggested changing the band's name to Beatals, as a tribute to Buddy Holly and the Crickets. They used this name until May, when they became the Silver Beetles, before undertaking a brief tour of Scotland as the backing group for pop singer and fellow Liverpudlian Johnny Gentle. By early July, they had refashioned themselves as the Silver Beatles, and by the middle of August simply the Beatles. Allan Williams, the Beatles' unofficial manager, arranged a residency for them in Hamburg. They auditioned and hired drummer Pete Best in mid-August 1960. The band, now a five-piece, departed Liverpool for Hamburg four days later, contracted to club owner Bruno Koschmider for what would be a 3+1⁄2-month residency. Beatles historian Mark Lewisohn writes: "They pulled into Hamburg at dusk on 17 August, the time when the red-light area comes to life ... flashing neon lights screamed out the various entertainment on offer, while scantily clad women sat unabashed in shop windows waiting for business opportunities." Koschmider had converted a couple of strip clubs in the district into music venues, and he initially placed the Beatles at the Indra Club. After closing Indra due to noise complaints, he moved them to the Kaiserkeller in October. When he learned they had been performing at the rival Top Ten Club in breach of their contract, he gave them one month's termination notice, and reported the underage Harrison, who had obtained permission to stay in Hamburg by lying to the German authorities about his age. The authorities arranged for Harrison's deportation in late November. One week later, Koschmider had McCartney and Best arrested for arson after they set fire to a condom in a concrete corridor; the authorities deported them. Lennon returned to Liverpool in early December, while Sutcliffe remained in Hamburg until late February with his German fiancée Astrid Kirchherr, who took the first semi-professional photos of the Beatles. During the next two years, the Beatles were resident for periods in Hamburg, where they used Preludin both recreationally and to maintain their energy through all-night performances. In 1961, during their second Hamburg engagement, Kirchherr cut Sutcliffe's hair in the "exi" (existentialist) style, later adopted by the other Beatles. Later on, Sutcliffe decided to leave the band early that year and resume his art studies in Germany. McCartney took over bass. Producer Bert Kaempfert contracted what was now a four-piece group until June 1962, and he used them as Tony Sheridan's backing band on a series of recordings for Polydor Records. As part of the sessions, the Beatles were signed to Polydor for one year. Credited to "Tony Sheridan & the Beat Brothers", the single "My Bonnie", recorded in June 1961 and released four months later, reached number 32 on the Musikmarkt chart. After the Beatles completed their second Hamburg residency, they enjoyed increasing popularity in Liverpool with the growing Merseybeat movement. However, they were growing tired of the monotony of numerous appearances at the same clubs night after night. In November 1961, during one of the group's frequent performances at the Cavern Club, they encountered Brian Epstein, a local record-store owner and music columnist. He later recalled: "I immediately liked what I heard. They were fresh, and they were honest, and they had what I thought was a sort of presence ... [a] star quality." Epstein courted the band over the next couple of months, and they appointed him as their manager in January 1962. Throughout early and mid-1962, Epstein sought to free the Beatles from their contractual obligations to Bert Kaempfert Productions. He eventually negotiated a one-month early release in exchange for one last recording session in Hamburg. On their return to Germany in April, a distraught Kirchherr met them at the airport with news of Sutcliffe's death the previous day from a brain haemorrhage. Epstein began negotiations with record labels for a recording contract. To secure a UK record contract, Epstein negotiated an early end to the band's contract with Polydor, in exchange for more recordings backing Tony Sheridan. After a New Year's Day audition, Decca Records rejected the band, saying, "Guitar groups are on the way out, Mr. Epstein." However, three months later, producer George Martin signed the Beatles to EMI's Parlophone label. Martin's first recording session with the Beatles took place at EMI Recording Studios (later Abbey Road Studios) in London on 6 June 1962. He immediately complained to Epstein about Best's drumming and suggested they use a session drummer in his place. Already contemplating Best's dismissal, the Beatles replaced him in mid-August with Ringo Starr, who left Rory Storm and the Hurricanes to join them. A 4 September session at EMI yielded a recording of "Love Me Do" featuring Starr on drums, but a dissatisfied Martin hired drummer Andy White for the band's third session a week later, which produced recordings of "Love Me Do", "Please Please Me" and "P.S. I Love You". Martin initially selected the Starr version of "Love Me Do" for the band's first single, though subsequent re-pressings featured the White version, with Starr on tambourine. Released in early October, "Love Me Do" peaked at number seventeen on the Record Retailer chart. Their television debut came later that month with a live performance on the regional news programme People and Places. After Martin suggested rerecording "Please Please Me" at a faster tempo, a studio session in late November yielded that recording, of which Martin accurately predicted, "You've just made your first No. 1." In December 1962, the Beatles concluded their fifth and final Hamburg residency. By 1963, they had agreed that all four band members would contribute vocals to their albums – including Starr, despite his restricted vocal range, to validate his standing in the group. Lennon and McCartney had established a songwriting partnership, and as the band's success grew, their dominant collaboration limited Harrison's opportunities as a lead vocalist. Epstein, to maximise the Beatles' commercial potential, encouraged them to adopt a professional approach to performing. Lennon recalled him saying, "Look, if you really want to get in these bigger places, you're going to have to change – stop eating on stage, stop swearing, stop smoking ...." On 11 February 1963, the Beatles recorded ten songs during a single studio session for their debut LP, Please Please Me. It was supplemented by the four tracks already released on their first two singles. Martin considered recording the LP live at The Cavern Club, but after deciding that the building's acoustics were inadequate, he elected to simulate a "live" album with minimal production in "a single marathon session at Abbey Road". After the moderate success of "Love Me Do", the single "Please Please Me" was released in January 1963, two months ahead of the album. It reached number one on every UK chart except Record Retailer, where it peaked at number two. Recalling how the Beatles "rushed to deliver a debut album, bashing out Please Please Me in a day", AllMusic critic Stephen Thomas Erlewine wrote: "Decades after its release, the album still sounds fresh, precisely because of its intense origins." Lennon said little thought went into composition at the time; he and McCartney were "just writing songs à la Everly Brothers, à la Buddy Holly, pop songs with no more thought of them than that – to create a sound. And the words were almost irrelevant." Released in March 1963, Please Please Me was the first of eleven consecutive Beatles albums released in the United Kingdom to reach number one. The band's third single, "From Me to You", came out in April and began an almost unbroken string of seventeen British number-one singles, including all but one of the eighteen they released over the next six years. Issued in August, their fourth single, "She Loves You", achieved the fastest sales of any record in the UK up to that time, selling three-quarters of a million copies in under four weeks. It became their first single to sell a million copies, and remained the biggest-selling record in the UK until 1978. The success brought increased media exposure, to which the Beatles responded with an irreverent and comical attitude that defied the expectations of pop musicians at the time, inspiring even more interest. The band toured the UK three times in the first half of the year: a four-week tour that began in February, the Beatles' first nationwide, preceded three-week tours in March and May–June. As their popularity spread, a frenzied adulation of the group took hold. On 13 October, the Beatles starred on Sunday Night at the London Palladium, the UK's top variety show. Their performance was televised live and watched by 15 million viewers. One national paper's headlines in the following days coined the term "Beatlemania" to describe the riotous enthusiasm by screaming fans who greeted the band – and it stuck. Although not billed as tour leaders, the Beatles overshadowed American acts Tommy Roe and Chris Montez during the February engagements and assumed top billing "by audience demand", something no British act had previously accomplished while touring with artists from the US. A similar situation arose during their May–June tour with Roy Orbison. In late October, the Beatles began a five-day tour of Sweden, their first time abroad since the final Hamburg engagement of December 1962. On their return to the UK on 31 October, several hundred screaming fans greeted them in heavy rain at Heathrow Airport. Around 50 to 100 journalists and photographers, as well as representatives from the BBC, also joined the airport reception, the first of more than 100 such events. The next day, the band began its fourth tour of Britain within nine months, this one scheduled for six weeks. In mid-November, as Beatlemania intensified, police resorted to using high-pressure water hoses to control the crowd before a concert in Plymouth. Please Please Me maintained the top position on the Record Retailer chart for 30 weeks, only to be displaced by its follow-up, With the Beatles, which EMI released on 22 November to record advance orders of 270,000 copies. The LP topped a half-million albums sold in one week. Recorded between July and October, With the Beatles made better use of studio production techniques than its predecessor. It held the top spot for 21 weeks with a chart life of 40 weeks. Erlewine described the LP as "a sequel of the highest order – one that betters the original". In a reversal of then standard practice, EMI released the album ahead of the impending single "I Want to Hold Your Hand", with the song excluded to maximise the single's sales. The album caught the attention of music critic William Mann of The Times, who suggested that Lennon and McCartney were "the outstanding English composers of 1963". The newspaper published a series of articles in which Mann offered detailed analyses of the music, lending it respectability. With the Beatles became the second album in UK chart history to sell a million copies, a figure previously reached only by the 1958 South Pacific soundtrack. When writing the sleeve notes for the album, the band's press officer, Tony Barrow, used the superlative the "fabulous foursome", which the media widely adopted as "the Fab Four". EMI's American subsidiary, Capitol Records, hindered the Beatles' releases in the United States for more than a year by initially declining to issue their music, including their first three singles. Concurrent negotiations with the independent US label Vee-Jay led to the release of some, but not all, of the songs in 1963. Vee-Jay finished preparation for the album Introducing... The Beatles, comprising most of the songs of Parlophone's Please Please Me, but a management shake-up led to the album not being released. After it emerged that the label did not report royalties on their sales, the licence that Vee-Jay had signed with EMI was voided. A new licence was granted to the Swan label for the single "She Loves You". The record received some airplay in the Tidewater area of Virginia from Gene Loving of radio station WGH and was featured on the "Rate-a-Record" segment of American Bandstand, but it failed to catch on nationally. Epstein brought a demo copy of "I Want to Hold Your Hand" to Capitol's Brown Meggs, who signed the band and arranged for a $40,000 US marketing campaign. American chart success began after disc jockey Carroll James of AM radio station WWDC, in Washington, DC, obtained a copy of the British single "I Want to Hold Your Hand" in mid-December 1963 and began playing it on-air. Taped copies of the song soon circulated among other radio stations throughout the US. This caused an increase in demand, leading Capitol to bring forward the release of "I Want to Hold Your Hand" by three weeks. Issued on 26 December, with the band's previously scheduled debut there just weeks away, "I Want to Hold Your Hand" sold a million copies, becoming a number-one hit in the US by mid-January. In its wake Vee-Jay released Introducing... The Beatles along with Capitol's debut album, Meet the Beatles!, while Swan reactivated production of "She Loves You". On 7 February 1964, the Beatles departed from Heathrow with an estimated 4,000 fans waving and screaming as the aircraft took off. Upon landing at New York's John F. Kennedy Airport, an uproarious crowd estimated at 3,000 greeted them. They gave their first live US television performance two days later on The Ed Sullivan Show, watched by approximately 73 million viewers in over 23 million households, or 34 per cent of the American population. Biographer Jonathan Gould writes that, according to the Nielsen rating service, it was "the largest audience that had ever been recorded for an American television program". The next morning, the Beatles awoke to a largely negative critical consensus in the US, but a day later at their first US concert, Beatlemania erupted at the Washington Coliseum. Back in New York the following day, the Beatles met with another strong reception during two shows at Carnegie Hall. The band flew to Florida, where they appeared on The Ed Sullivan Show a second time, again before 70 million viewers, before returning to the UK on 22 February. The Beatles' first visit to the US took place when the nation was still mourning the assassination of President John F. Kennedy the previous November. Commentators often suggest that for many, particularly the young, the Beatles' performances reignited the sense of excitement and possibility that momentarily faded in the wake of the assassination, and helped pave the way for the revolutionary social changes to come later in the decade. Their hairstyle, unusually long for the era and mocked by many adults, became an emblem of rebellion to the burgeoning youth culture. The group's popularity generated unprecedented interest in British music, and many other UK acts subsequently made their American debuts, successfully touring over the next three years in what was termed the British Invasion. The Beatles' success in the US opened the door for a successive string of British beat groups and pop acts such as the Dave Clark Five, the Animals, Petula Clark, the Kinks, and the Rolling Stones to achieve success in America. During the week of 4 April 1964, the Beatles held twelve positions on the Billboard Hot 100 singles chart, including the top five. Capitol Records' lack of interest throughout 1963 did not go unnoticed, and a competitor, United Artists Records, encouraged its film division to offer the Beatles a three-motion-picture deal, primarily for the commercial potential of the soundtracks in the US. Directed by Richard Lester, A Hard Day's Night involved the band for six weeks in March–April 1964 as they played themselves in a musical comedy. The film premiered in London and New York in July and August, respectively, and was an international success, with some critics drawing a comparison with the Marx Brothers. United Artists released a full soundtrack album for the North American market, combining Beatles songs and Martin's orchestral score; elsewhere, the group's third studio LP, A Hard Day's Night, contained songs from the film on side one and other new recordings on side two. According to Erlewine, the album saw them "truly coming into their own as a band. All of the disparate influences on their first two albums coalesced into a bright, joyous, original sound, filled with ringing guitars and irresistible melodies." That "ringing guitar" sound was primarily the product of Harrison's 12-string electric Rickenbacker, a prototype given to him by the manufacturer, which made its debut on the record. Touring internationally in June and July, the Beatles staged 37 shows over 27 days in Denmark, the Netherlands, Hong Kong, Australia and New Zealand. In August and September, they returned to the US, with a 30-concert tour of 23 cities. Generating intense interest once again, the month-long tour attracted between 10,000 and 20,000 fans to each 30-minute performance in cities from San Francisco to New York. In August, journalist Al Aronowitz arranged for the Beatles to meet Bob Dylan. Visiting the band in their New York hotel suite, Dylan introduced them to cannabis. Gould points out the musical and cultural significance of this meeting, before which the musicians' respective fanbases were "perceived as inhabiting two separate subcultural worlds": Dylan's audience of "college kids with artistic or intellectual leanings, a dawning political and social idealism, and a mildly bohemian style" contrasted with their fans, "veritable 'teenyboppers' – kids in high school or grade school whose lives were totally wrapped up in the commercialised popular culture of television, radio, pop records, fan magazines, and teen fashion. To many of Dylan's followers in the folk music scene, the Beatles were seen as idolaters, not idealists." Within six months of the meeting, according to Gould, "Lennon would be making records on which he openly imitated Dylan's nasal drone, brittle strum, and introspective vocal persona"; and six months after that, Dylan began performing with a backing band and electric instrumentation, and "dressed in the height of Mod fashion". As a result, Gould continues, the traditional division between folk and rock enthusiasts "nearly evaporated", as the Beatles' fans began to mature in their outlook and Dylan's audience embraced the new, youth-driven pop culture. During the 1964 US tour, the group were confronted with racial segregation in the country at the time. When informed that the venue for their 11 September concert, the Gator Bowl in Jacksonville, Florida, was segregated, the Beatles said they would refuse to perform unless the audience was integrated. Lennon stated: "We never play to segregated audiences and we aren't going to start now ... I'd sooner lose our appearance money." City officials relented and agreed to allow an integrated show. The group also cancelled their reservations at the whites-only Hotel George Washington in Jacksonville. For their subsequent US tours in 1965 and 1966, the Beatles included clauses in contracts stipulating that shows be integrated. According to Gould, the Beatles' fourth studio LP, Beatles for Sale, evidenced a growing conflict between the commercial pressures of their global success and their creative ambitions. They had intended the album, recorded between August and October 1964, to continue the format established by A Hard Day's Night which, unlike their first two LPs, contained only original songs. They had nearly exhausted their backlog of songs on the previous album, however, and given the challenges constant international touring posed to their songwriting efforts, Lennon admitted, "Material's becoming a hell of a problem". As a result, six covers from their extensive repertoire were chosen to complete the album. Released in early December, its eight original compositions stood out, demonstrating the growing maturity of the Lennon–McCartney songwriting partnership. In early 1965, following a dinner with Lennon, Harrison and their wives, Harrison's dentist, John Riley, secretly added LSD to their coffee. Lennon described the experience: "It was just terrifying, but it was fantastic. I was pretty stunned for a month or two." He and Harrison subsequently became regular users of the drug, joined by Starr on at least one occasion. Harrison's use of psychedelic drugs encouraged his path to meditation and Hinduism. He commented: "For me, it was like a flash. The first time I had acid, it just opened up something in my head that was inside of me, and I realised a lot of things. I didn't learn them because I already knew them, but that happened to be the key that opened the door to reveal them. From the moment I had that, I wanted to have it all the time – these thoughts about the yogis and the Himalayas, and Ravi's music." McCartney was initially reluctant to try it, but eventually did so in late 1966. He became the first Beatle to discuss LSD publicly, declaring in a magazine interview that "it opened my eyes" and "made me a better, more honest, more tolerant member of society". Controversy erupted in June 1965 when Queen Elizabeth II appointed all four Beatles Members of the Order of the British Empire (MBE) after Prime Minister Harold Wilson nominated them for the award. In protest – the honour was at that time primarily bestowed upon military veterans and civic leaders – some conservative MBE recipients returned their insignia. In July, the Beatles' second film, Help!, was released, again directed by Lester. Described as "mainly a relentless spoof of Bond", it inspired a mixed response among both reviewers and the band. McCartney said: "Help! was great but it wasn't our film – we were sort of guest stars. It was fun, but basically, as an idea for a film, it was a bit wrong." The soundtrack was dominated by Lennon, who wrote and sang lead on most of its songs, including the two singles: "Help!" and "Ticket to Ride". The Help! album, the group's fifth studio LP, mirrored A Hard Day's Night by featuring soundtrack songs on side one and additional songs from the same sessions on side two. The LP contained all original material save for two covers, "Act Naturally" and "Dizzy Miss Lizzy"; they were the last covers the band would include on an album, except for Let It Be's brief rendition of the traditional Liverpool folk song "Maggie Mae". The band expanded their use of vocal overdubs on Help! and incorporated classical instruments into some arrangements, including a string quartet on the pop ballad "Yesterday". Composed by and sung by McCartney – none of the other Beatles perform on the recording – "Yesterday" has inspired the most cover versions of any song ever written. With Help!, the Beatles became the first rock group to be nominated for a Grammy Award for Album of the Year. The group's third US tour opened with a performance before a world-record crowd of 55,600 at New York's Shea Stadium on 15 August – "perhaps the most famous of all Beatles' concerts", in Lewisohn's description. A further nine successful concerts followed in other American cities. At a show in Atlanta, the Beatles gave one of the first live performances ever to make use of a foldback system of on-stage monitor speakers. Towards the end of the tour, they met with Elvis Presley, a foundational musical influence on the band, who invited them to his home in Beverly Hills. September 1965 saw the launch of an American Saturday-morning cartoon series, The Beatles, that echoed A Hard Day's Night's slapstick antics over its two-year original run. The series was a historical milestone as the first weekly television series to feature animated versions of real, living people. In mid-October, the Beatles entered the recording studio; for the first time when making an album, they had an extended period without other major commitments. Until this time, according to George Martin, "we had been making albums rather like a collection of singles. Now we were really beginning to think about albums as a bit of art on their own." Released in December, Rubber Soul was hailed by critics as a major step forward in the maturity and complexity of the band's music. Their thematic reach was beginning to expand as they embraced deeper aspects of romance and philosophy, a development that NEMS executive Peter Brown attributed to the band members' "now habitual use of marijuana". Lennon referred to Rubber Soul as "the pot album" and Starr said: "Grass was really influential in a lot of our changes, especially with the writers. And because they were writing different material, we were playing differently." After Help!'s foray into classical music with flutes and strings, Harrison's introduction of a sitar on "Norwegian Wood (This Bird Has Flown)" marked a further progression outside the traditional boundaries of popular music. As the lyrics grew more artful, fans began to study them for deeper meaning. While some of Rubber Soul's songs were the product of Lennon and McCartney's collaborative songwriting, the album also included distinct compositions from each, though they continued to share official credit. "In My Life", of which each later claimed lead authorship, is considered a highlight of the entire Lennon–McCartney catalogue. Harrison called Rubber Soul his "favourite album", and Starr referred to it as "the departure record". McCartney has said, "We'd had our cute period, and now it was time to expand." However, recording engineer Norman Smith later stated that the studio sessions revealed signs of growing conflict within the group – "the clash between John and Paul was becoming obvious", he wrote, and "as far as Paul was concerned, George could do no right". In 2003, Rolling Stone ranked Rubber Soul fifth among "The 500 Greatest Albums of All Time", and AllMusic's Richie Unterberger describes it as "one of the classic folk-rock records". Capitol Records, from December 1963 when it began issuing Beatles recordings for the US market, exercised complete control over format, compiling distinct US albums from the band's recordings and issuing songs of their choosing as singles. In June 1966, the Capitol LP Yesterday and Today caused an uproar with its cover, which portrayed the grinning Beatles dressed in butcher's overalls, accompanied by raw meat and mutilated plastic baby dolls. According to Beatles biographer Bill Harry, it has been incorrectly suggested that this was meant as a satirical response to the way Capitol had "butchered" the US versions of the band's albums. Thousands of copies of the LP had a new cover pasted over the original; an unpeeled "first-state" copy fetched $10,500 at a December 2005 auction. In England, meanwhile, Harrison met sitar maestro Ravi Shankar, who agreed to train him on the instrument. During a tour of the Philippines the month after the Yesterday and Today furore, the Beatles unintentionally snubbed the nation's first lady, Imelda Marcos, who had expected them to attend a breakfast reception at the Presidential Palace. When presented with the invitation, Epstein politely declined on the band members' behalf, as it had never been his policy to accept such official invitations. They soon found that the Marcos regime was unaccustomed to taking no for an answer. The resulting riots endangered the group and they escaped the country with difficulty. Immediately afterwards, the band members visited India for the first time. We're more popular than Jesus now; I don't know which will go first – rock 'n' roll or Christianity. – John Lennon, 1966 Almost as soon as they returned home, the Beatles faced a fierce backlash from US religious and social conservatives (as well as the Ku Klux Klan) over a comment Lennon had made in a March interview with British reporter Maureen Cleave. "Christianity will go", Lennon had said. "It will vanish and shrink. I needn't argue about that; I'm right and I will be proved right ... Jesus was alright but his disciples were thick and ordinary. It's them twisting it that ruins it for me." His comments went virtually unnoticed in England, but when US teenage fan magazine Datebook printed them five months later, it sparked a controversy with Christians in America's conservative Bible Belt region. The Vatican issued a protest, and bans on Beatles' records were imposed by Spanish and Dutch stations and South Africa's national broadcasting service. Epstein accused Datebook of having taken Lennon's words out of context. At a press conference, Lennon pointed out, "If I'd said television was more popular than Jesus, I might have got away with it." He claimed that he was referring to how other people viewed their success, but at the prompting of reporters, he concluded: "If you want me to apologise, if that will make you happy, then okay, I'm sorry." Released in August 1966, a week before the Beatles' final tour, Revolver marked another artistic step forward for the group. The album featured sophisticated songwriting, studio experimentation, and a greatly expanded repertoire of musical styles, ranging from innovative classical string arrangements to psychedelia. Abandoning the customary group photograph, its Aubrey Beardsley-inspired cover – designed by Klaus Voormann, a friend of the band since their Hamburg days – was a monochrome collage and line drawing caricature of the group. The album was preceded by the single "Paperback Writer", backed by "Rain". Short promotional films were made for both songs; described by cultural historian Saul Austerlitz as "among the first true music videos", they aired on The Ed Sullivan Show and Top of the Pops in June. Among the experimental songs on Revolver was "Tomorrow Never Knows", the lyrics for which Lennon drew from Timothy Leary's The Psychedelic Experience: A Manual Based on the Tibetan Book of the Dead. Its creation involved eight tape decks distributed about the EMI building, each staffed by an engineer or band member, who randomly varied the movement of a tape loop while Martin created a composite recording by sampling the incoming data. McCartney's "Eleanor Rigby" made prominent use of a string octet; Gould describes it as "a true hybrid, conforming to no recognisable style or genre of song". Harrison's emergence as a songwriter was reflected in three of his compositions appearing on the record. Among these, "Taxman", which opened the album, marked the first example of the Beatles making a political statement through their music. In 2020, Rolling Stone ranked Revolver at #11 on their list of "The 500 Greatest Albums of All Time". As preparations were made for a tour of the US, the Beatles knew that their music would hardly be heard. Having originally used Vox AC30 amplifiers, they later acquired more powerful 100-watt amplifiers, specially designed for them by Vox, as they moved into larger venues in 1964; however, these were still inadequate. Struggling to compete with the volume of sound generated by screaming fans, the band had grown increasingly bored with the routine of performing live. Recognising that their shows were no longer about the music, they decided to make the August tour their last. The band performed none of their new songs on the tour. In Chris Ingham's description, they were very much "studio creations ... and there was no way a four-piece rock 'n' roll group could do them justice, particularly through the desensitising wall of the fans' screams. 'Live Beatles' and 'Studio Beatles' had become entirely different beasts." The band's concert at San Francisco's Candlestick Park on 29 August was their last commercial concert. It marked the end of four years dominated by almost non-stop touring that included over 1,400 concert appearances internationally. Freed from the burden of touring, the Beatles embraced an increasingly experimental approach as they recorded Sgt. Pepper's Lonely Hearts Club Band, beginning in late November 1966. According to engineer Geoff Emerick, the album's recording took over 700 hours. He recalled the band's insistence "that everything on Sgt. Pepper had to be different. We had microphones right down in the bells of brass instruments and headphones turned into microphones attached to violins. We used giant primitive oscillators to vary the speed of instruments and vocals and we had tapes chopped to pieces and stuck together upside down and the wrong way around." Parts of "A Day in the Life" featured a 40-piece orchestra. The sessions initially yielded the non-album double A-side single "Strawberry Fields Forever"/"Penny Lane" in February 1967; the Sgt. Pepper LP followed with a rush-release in May. The musical complexity of the records, created using relatively primitive four-track recording technology, astounded contemporary artists. Among music critics, acclaim for the album was virtually universal. Gould writes: The overwhelming consensus is that the Beatles had created a popular masterpiece: a rich, sustained, and overflowing work of collaborative genius whose bold ambition and startling originality dramatically enlarged the possibilities and raised the expectations of what the experience of listening to popular music on record could be. On the basis of this perception, Sgt. Pepper became the catalyst for an explosion of mass enthusiasm for album-formatted rock that would revolutionise both the aesthetics and the economics of the record business in ways that far outstripped the earlier pop explosions triggered by the Elvis phenomenon of 1956 and the Beatlemania phenomenon of 1963. In the wake of Sgt. Pepper, the underground and mainstream press widely publicised the Beatles as leaders of youth culture, as well as "lifestyle revolutionaries". The album was the first major pop/rock LP to include its complete lyrics, which appeared on the back cover. Those lyrics were the subject of critical analysis; for instance, in late 1967 the album was the subject of a scholarly inquiry by American literary critic and professor of English Richard Poirier, who observed that his students were "listening to the group's music with a degree of engagement that he, as a teacher of literature, could only envy". The elaborate cover also attracted considerable interest and study. A collage designed by pop artists Peter Blake and Jann Haworth, it depicted the group as the fictional band referred to in the album's title track standing in front of a crowd of famous people. The heavy moustaches worn by the group reflected the growing influence of the hippie movement, while cultural historian Jonathan Harris describes their "brightly coloured parodies of military uniforms" as a knowingly "anti-authoritarian and anti-establishment" display. Sgt. Pepper topped the UK charts for 23 consecutive weeks, with a further four weeks at number one in the period through to February 1968. With 2.5 million copies sold within three months of its release, Sgt. Pepper's initial commercial success exceeded that of all previous Beatles albums. It sustained its immense popularity into the 21st century while breaking numerous sales records. In 2003, Rolling Stone ranked Sgt. Pepper at number one on its list of the greatest albums of all time. Two Beatles film projects were conceived within weeks of completing Sgt. Pepper: Magical Mystery Tour, a one-hour television film, and Yellow Submarine, an animated feature-length film produced by United Artists. The group began recording music for the former in late April 1967, but the project then lay dormant as they focused on recording songs for the latter. On 25 June, the Beatles performed their forthcoming single "All You Need Is Love" to an estimated 350 million viewers on Our World, the first live global television link. Released a week later, during the Summer of Love, the song was adopted as a flower power anthem. The Beatles' use of psychedelic drugs was at its height during that summer. In July and August, the group pursued interests related to similar utopian-based ideology, including a week-long investigation into the possibility of starting an island-based commune off the coast of Greece. On 24 August, the group were introduced to Maharishi Mahesh Yogi in London. The next day, they travelled to Bangor for his Transcendental Meditation retreat. On 27 August, their manager's assistant, Peter Brown, phoned to inform them that Epstein had died. The coroner ruled the death an accidental carbitol overdose, although it was widely rumoured to be a suicide. His death left the group disoriented and fearful about the future. Lennon recalled: "We collapsed. I knew that we were in trouble then. I didn't really have any misconceptions about our ability to do anything other than play music, and I was scared. I thought, 'We've fuckin' had it now.'" Harrison's then-wife Pattie Boyd remembered that "Paul and George were in complete shock. I don't think it could have been worse if they had heard that their own fathers had dropped dead." During a band meeting in September, McCartney recommended that the band proceed with Magical Mystery Tour. The Magical Mystery Tour soundtrack was released in the UK as a six-track double extended play (EP) in early December 1967. It was the first example of a double EP in the UK. The record carried on the psychedelic vein of Sgt. Pepper, though in line with the band's wishes, the packaging reinforced the idea that the release was a film soundtrack rather than a follow-up to Sgt. Pepper. In the US, the soundtrack appeared as an identically titled LP that also included five tracks from the band's recent singles. In its first three weeks, the album set a record for the highest initial sales of any Capitol LP, and it is the only Capitol compilation later to be adopted in the band's official canon of studio albums. Magical Mystery Tour first aired on Boxing Day to an audience of approximately 15 million. Largely directed by McCartney, the film was the band's first critical failure in the UK. It was dismissed as "blatant rubbish" by the Daily Express; the Daily Mail called it "a colossal conceit"; and The Guardian labelled the film "a kind of fantasy morality play about the grossness and warmth and stupidity of the audience". Gould describes it as "a great deal of raw footage showing a group of people getting on, getting off, and riding on a bus". Although the viewership figures were respectable, its slating in the press led US television networks to lose interest in broadcasting the film. The group were less involved with Yellow Submarine, which featured the band appearing as themselves for only a short live-action segment. Premiering in July 1968, the film featured cartoon versions of the band members and a soundtrack with eleven of their songs, including four unreleased studio recordings that made their debut in the film. Critics praised the film for its music, humour and innovative visual style. A soundtrack LP was issued seven months later; it contained those four new songs, the title track (already issued on Revolver), "All You Need Is Love" (already issued as a single and on the US Magical Mystery Tour LP) and seven instrumental pieces composed by Martin. In February 1968, the Beatles travelled to Maharishi Mahesh Yogi's ashram in Rishikesh, India, to take part in a three-month meditation "Guide Course". Their time in India marked one of the band's most prolific periods, yielding numerous songs, including a majority of those on their next album. However, Starr left after only ten days, unable to stomach the food, and McCartney eventually grew bored and departed a month later. For Lennon and Harrison, creativity turned to question when an electronics technician known as Magic Alex suggested that the Maharishi was attempting to manipulate them. When he alleged that the Maharishi had made sexual advances to women attendees, a persuaded Lennon left abruptly just two months into the course, bringing an unconvinced Harrison and the remainder of the group's entourage with him. In anger, Lennon wrote a scathing song titled "Maharishi", renamed "Sexy Sadie" to avoid potential legal issues. McCartney said, "We made a mistake. We thought there was more to him than there was." In May, Lennon and McCartney travelled to New York for the public unveiling of the Beatles' new business venture, Apple Corps. It was initially formed several months earlier as part of a plan to create a tax-effective business structure, but the band then desired to extend the corporation to other pursuits, including record distribution, peace activism, and education. McCartney described Apple as "rather like a Western communism". The enterprise drained the group financially with a series of unsuccessful projects handled largely by members of the Beatles' entourage, who were given their jobs regardless of talent and experience. Among its numerous subsidiaries were Apple Electronics, established to foster technological innovations with Magic Alex at the head, and Apple Retailing, which opened the short-lived Apple Boutique in London. Harrison later said, "Basically, it was chaos ... John and Paul got carried away with the idea and blew millions, and Ringo and I just had to go along with it." From late May to mid-October 1968, the group recorded what became The Beatles, a double LP commonly known as "the White Album" for its virtually featureless cover. During this time, relations between the members grew openly divisive. Starr quit for two weeks, leaving his bandmates to record "Back in the U.S.S.R." and "Dear Prudence" as a trio, with McCartney filling in on drums. Lennon had lost interest in collaborating with McCartney, whose contribution "Ob-La-Di, Ob-La-Da" he scorned as "granny music shit". Tensions were further aggravated by Lennon's romantic preoccupation with avant-garde artist Yoko Ono, whom he insisted on bringing to the sessions despite the group's well-established understanding that girlfriends were not allowed in the studio. McCartney has recalled that the album "wasn't a pleasant one to make". He and Lennon identified the sessions as the start of the band's break-up. With the record, the band executed a wider range of musical styles and broke with their recent tradition of incorporating several musical styles in one song by keeping each piece of music consistently faithful to a select genre. During the sessions, the group upgraded to an eight-track tape console, which made it easier for them to layer tracks piecemeal, while the members often recorded independently of each other, affording the album a reputation as a collection of solo recordings rather than a unified group effort. Describing the double album, Lennon later said: "Every track is an individual track; there isn't any Beatle music on it. [It's] John and the band, Paul and the band, George and the band." The sessions also produced the Beatles' longest song yet, "Hey Jude", released in August as a non-album single with "Revolution". Issued in November, the White Album was the band's first Apple Records album release, although EMI continued to own their recordings. The record attracted more than 2 million advance orders, selling nearly 4 million copies in the US in little over a month, and its tracks dominated the playlists of American radio stations. Its lyric content was the focus of much analysis by the counterculture. Despite its popularity, reviewers were largely confused by the album's content, and it failed to inspire the level of critical writing that Sgt. Pepper had. General critical opinion eventually turned in favour of the White Album, and in 2003, Rolling Stone ranked it as the tenth-greatest album of all time. Although Let It Be was the Beatles' final album release, it was largely recorded before Abbey Road. The project's impetus came from an idea Martin attributes to McCartney, who suggested they "record an album of new material and rehearse it, then perform it before a live audience for the very first time – on record and on film". Originally intended for a one-hour television programme to be called Beatles at Work, in the event much of the album's content came from studio work beginning in January 1969, many hours of which were captured on film by director Michael Lindsay-Hogg. Martin said that the project was "not at all a happy recording experience. It was a time when relations between the Beatles were at their lowest ebb." Lennon described the largely impromptu sessions as "hell ... the most miserable ... on Earth", and Harrison, "the low of all-time". Irritated by McCartney and Lennon, Harrison walked out for five days. Upon returning, he threatened to leave the band unless they "abandon[ed] all talk of live performance" and instead focused on finishing a new album, initially titled Get Back, using songs recorded for the TV special. He also demanded they cease work at Twickenham Film Studios, where the sessions had begun, and relocate to the newly finished Apple Studio. His bandmates agreed, and it was decided to salvage the footage shot for the TV production for use in a feature film. To alleviate tensions within the band and improve the quality of their live sound, Harrison invited keyboardist Billy Preston to participate in the last nine days of sessions. Preston received label billing on the "Get Back" single – the only musician ever to receive that acknowledgment on an official Beatles release. After the rehearsals, the band could not agree on a location to film a concert, rejecting several ideas, including a boat at sea, a lunatic asylum, the Libyan desert, and the Colosseum. Ultimately, what would be their final live performance was filmed on the rooftop of the Apple Corps building at 3 Savile Row, London, on 30 January 1969. Five weeks later, engineer Glyn Johns, whom Lewisohn describes as Get Back's "uncredited producer", began work assembling an album, given "free rein" as the band "all but washed their hands of the entire project". New strains developed between the band members regarding the appointment of a financial adviser, the need for which had become evident without Epstein to manage business affairs. Lennon, Harrison and Starr favoured Allen Klein, who had managed the Rolling Stones and Sam Cooke; McCartney wanted Lee and John Eastman – father and brother, respectively, of Linda Eastman, whom McCartney married on 12 March. Agreement could not be reached, so both Klein and the Eastmans were temporarily appointed: Klein as the Beatles' business manager and the Eastmans as their lawyers. Further conflict ensued, however, and financial opportunities were lost. On 8 May, Klein was named sole manager of the band, the Eastmans having previously been dismissed as the Beatles' lawyers. McCartney refused to sign the management contract with Klein, but he was out-voted by the other Beatles. Martin stated that he was surprised when McCartney asked him to produce another album, as the Get Back sessions had been "a miserable experience" and he had "thought it was the end of the road for all of us". The primary recording sessions for Abbey Road began on 2 July. Lennon, who rejected Martin's proposed format of a "continuously moving piece of music", wanted his and McCartney's songs to occupy separate sides of the album. The eventual format, with individually composed songs on the first side and the second consisting largely of a medley, was McCartney's suggested compromise. Emerick noted that the replacement of the studio's valve-based mixing console with a transistorised one yielded a less punchy sound, leaving the group frustrated at the thinner tone and lack of impact and contributing to its "kinder, gentler" feel relative to their previous albums. On 4 July, the first solo single by a Beatle was released: Lennon's "Give Peace a Chance", credited to the Plastic Ono Band. The completion and mixing of "I Want You (She's So Heavy)" on 20 August was the last occasion on which all four Beatles were together in the same studio. On 8 September, while Starr was in hospital, the other band members met to discuss recording a new album. They considered a different approach to songwriting by ending the Lennon–McCartney pretence and having four compositions apiece from Lennon, McCartney and Harrison, with two from Starr and a lead single around Christmas. On 20 September, Lennon announced his departure to the rest of the group but agreed to withhold a public announcement to avoid undermining sales of the forthcoming album. Released on 26 September, Abbey Road sold four million copies within three months and topped the UK charts for a total of seventeen weeks. Its second track, the ballad "Something", was issued as a single – the only Harrison composition that appeared as a Beatles A-side. Abbey Road received mixed reviews, although the medley met with general acclaim. Unterberger considers it "a fitting swan song for the group", containing "some of the greatest harmonies to be heard on any rock record". Musicologist and author Ian MacDonald calls the album "erratic and often hollow", despite the "semblance of unity and coherence" offered by the medley. Martin singled it out as his favourite Beatles album; Lennon said it was "competent" but had "no life in it". For the still unfinished Get Back album, one last song, Harrison's "I Me Mine", was recorded on 3 January 1970. Lennon, in Denmark at the time, did not participate. In March, rejecting the work Johns had done on the project, now retitled Let It Be, Klein gave the session tapes to American producer Phil Spector, who had recently produced Lennon's solo single "Instant Karma!" In addition to remixing the material, Spector edited, spliced and overdubbed several of the recordings that had been intended as "live". McCartney was unhappy with the producer's approach and particularly dissatisfied with the lavish orchestration on "The Long and Winding Road", which involved a fourteen-voice choir and 36-piece instrumental ensemble. McCartney's demands that the alterations to the song be reverted were ignored, and he publicly announced his departure from the band on 10 April, a week before the release of his first self-titled solo album. On 8 May 1970, Let It Be was released. Its accompanying single, "The Long and Winding Road", was expected to be the Beatles' last; it was released in the US, but not in the UK. The Let It Be documentary film followed later that month, and would win the 1970 Academy Award for Best Original Song Score. Sunday Telegraph critic Penelope Gilliatt called it "a very bad film and a touching one ... about the breaking apart of this reassuring, geometrically perfect, once apparently ageless family of siblings". Several reviewers stated that some of the performances in the film sounded better than their analogous album tracks. Describing Let It Be as the "only Beatles album to occasion negative, even hostile reviews", Unterberger calls it "on the whole underrated"; he singles out "some good moments of straight hard rock in 'I've Got a Feeling' and 'Dig a Pony'", and praises "Let It Be", "Get Back", and "the folky 'Two of Us', with John and Paul harmonising together". McCartney filed suit for the dissolution of the Beatles' contractual partnership on 31 December 1970. Legal disputes continued long after their break-up, and the dissolution was not formalised until 29 December 1974, when Lennon signed the paperwork terminating the partnership while on vacation with his family at Walt Disney World Resort in Florida. Lennon, McCartney, Harrison and Starr all released solo albums in 1970. Their solo records sometimes involved one or more of the other members; Starr's Ringo (1973) was the only album to include compositions and performances by all four ex-Beatles, albeit on separate songs. With Starr's participation, Harrison staged the Concert for Bangladesh in New York City in August 1971. Other than an unreleased jam session in 1974, later bootlegged as A Toot and a Snore in '74, Lennon and McCartney never recorded together again. Two double-LP sets of the Beatles' greatest hits, compiled by Klein, 1962–1966 and 1967–1970, were released in 1973, at first under the Apple Records imprint. Commonly known as the "Red Album" and "Blue Album", respectively, each has earned a Multi-Platinum certification in the US and a Platinum certification in the UK. Between 1976 and 1982, EMI/Capitol released a wave of compilation albums without input from the ex-Beatles, starting with the double-disc compilation Rock 'n' Roll Music. The only one to feature previously unreleased material was The Beatles at the Hollywood Bowl (1977); the first officially issued concert recordings by the group, it contained selections from two shows they played during their 1964 and 1965 US tours. The music and enduring fame of the Beatles were commercially exploited in various other ways, again often outside their creative control. In April 1974, the musical John, Paul, George, Ringo ... and Bert, written by Willy Russell and featuring singer Barbara Dickson, opened in London. It included, with permission from Northern Songs, eleven Lennon-McCartney compositions and one by Harrison, "Here Comes the Sun". Displeased with the production's use of his song, Harrison withdrew his permission to use it. Later that year, the off-Broadway musical Sgt. Pepper's Lonely Hearts Club Band on the Road opened. All This and World War II (1976) was an unorthodox nonfiction film that combined newsreel footage with covers of Beatles songs by performers ranging from Elton John and Keith Moon to the London Symphony Orchestra. The Broadway musical Beatlemania, an unauthorised nostalgia revue, opened in early 1977 and proved popular, spinning off five separate touring productions. In 1979, the band sued the producers, settling for several million dollars in damages. Sgt. Pepper's Lonely Hearts Club Band (1978), a musical film starring the Bee Gees and Peter Frampton, was a commercial failure and an "artistic fiasco", according to Ingham. Accompanying the wave of Beatles nostalgia and persistent reunion rumours in the US during the 1970s, several entrepreneurs made public offers to the Beatles for a reunion concert. Promoter Bill Sargent first offered the Beatles $10 million for a reunion concert in 1974. He raised his offer to $30 million in January 1976 and then to $50 million the following month. On 24 April 1976, during a broadcast of Saturday Night Live, producer Lorne Michaels jokingly offered the Beatles $3,000 to reunite on the show. Lennon and McCartney were watching the live broadcast at Lennon's apartment at the Dakota in New York, which was within driving distance of the NBC studio where the show was being broadcast. The former bandmates briefly entertained the idea of going to the studio and surprising Michaels by accepting his offer, but decided not to. In December 1980, Lennon was shot and killed outside his New York City apartment. Harrison rewrote the lyrics of his song "All Those Years Ago" in Lennon's honour. With Starr on drums and McCartney and his wife, Linda, contributing backing vocals, the song was released as a single in May 1981. McCartney's own tribute, "Here Today", appeared on his Tug of War album in April 1982. In 1984, Starr co-starred in McCartney's film Give My Regards to Broad Street, and played with McCartney on several of the songs on the soundtrack. In 1987, Harrison's Cloud Nine album included "When We Was Fab", a song about the Beatlemania era. When the Beatles' studio albums were released on CD by EMI and Apple Corps in 1987, their catalogue was standardised throughout the world, establishing a canon of the twelve original studio LPs as issued in the UK plus the US LP version of Magical Mystery Tour. All the remaining material from the singles and EPs that had not appeared on these thirteen studio albums was gathered on the two-volume compilation Past Masters (1988). Except for the Red and Blue albums, EMI deleted all its other Beatles compilations – including the Hollywood Bowl record – from its catalogue. In 1988, the Beatles were inducted into the Rock and Roll Hall of Fame, their first year of eligibility. Harrison and Starr attended the ceremony with Lennon's widow, Yoko Ono, and his two sons, Julian and Sean. McCartney declined to attend, citing unresolved "business differences" that would make him "feel like a complete hypocrite waving and smiling with them at a fake reunion". The following year, EMI/Capitol settled a decade-long lawsuit filed by the band over royalties, clearing the way to commercially package previously unreleased material. Live at the BBC, the first official release of unissued Beatles performances in seventeen years, appeared in 1994. That same year McCartney, Harrison and Starr collaborated on the Anthology project. Anthology was the culmination of work begun in 1970, when Apple Corps director Neil Aspinall, their former road manager and personal assistant, had started to gather material for a documentary with the working title The Long and Winding Road. During 1995–96, the project yielded a television miniseries, an eight-volume video set, and three two-CD/three-LP box sets featuring artwork by Klaus Voormann. Documenting their history in the band's own words, the Anthology project included the release of several unissued Beatles recordings. Alongside producer Jeff Lynne, McCartney, Harrison and Starr also added new instrumental and vocal parts to songs recorded as demos by Lennon in the late 1970s, resulting in the release of two "new" Beatles singles, "Free as a Bird" and "Real Love". A third Lennon demo, "Now and Then", was also attempted, but abandoned due to the low quality of the recording. The Anthology releases were commercially successful and the television series was viewed by an estimated 400 million people. A book of the same name followed in October 2000. In 1999, to coincide with the re-release of the 1968 film Yellow Submarine, an expanded soundtrack album, Yellow Submarine Songtrack, was issued. The Beatles' 1, a compilation album of the band's British and American number-one hits, was released on 13 November 2000. It became the fastest-selling album of all time, with 3.6 million sold in its first week and 13 million within a month. It topped albums charts in at least 28 countries. The compilation had sold 31 million copies globally by April 2009. Harrison died from metastatic lung cancer in November 2001. McCartney and Starr were among the musicians who performed at the Concert for George, organised by Eric Clapton and Harrison's widow, Olivia. The tribute event took place at the Royal Albert Hall on the first anniversary of Harrison's death. In 2003, Let It Be... Naked, a reconceived version of the Let It Be album, with McCartney supervising production, was released. One of the main differences from the Spector-produced version was the omission of the original string arrangements. It was a top-ten hit in both Britain and America. The US album configurations from 1964 to 1965 were released as box sets in 2004 and 2006; The Capitol Albums, Volume 1 and Volume 2 included both stereo and mono versions based on the mixes that were prepared for vinyl at the time of the music's original American release. As a soundtrack for Cirque du Soleil's Las Vegas Beatles stage revue, Love, George Martin and his son Giles remixed and blended 130 of the band's recordings to create what Martin called "a way of re-living the whole Beatles musical lifespan in a very condensed period". The show premiered in June 2006, and the Love album was released that November. In April 2009, Starr performed three songs with McCartney at a benefit concert held at New York's Radio City Music Hall and organised by McCartney. On 9 September 2009, the Beatles' entire back catalogue was reissued following an extensive digital remastering process that lasted four years. Stereo editions of all twelve original UK studio albums, along with Magical Mystery Tour and the Past Masters compilation, were released on compact disc both individually and as a box set. A second collection, The Beatles in Mono, included remastered versions of every Beatles album released in true mono along with the original 1965 stereo mixes of Help! and Rubber Soul (both of which Martin had remixed for the 1987 editions). The Beatles: Rock Band, a music video game in the Rock Band series, was issued on the same day. In December 2009, the band's catalogue was officially released in FLAC and MP3 format in a limited edition of 30,000 USB flash drives. Owing to a long-running royalty disagreement, the Beatles were among the last major artists to sign deals with online music services. Residual disagreement emanating from Apple Corps' dispute with Apple, Inc., iTunes' owners, over the use of the name "Apple" was also partly responsible for the delay, although in 2008, McCartney stated that the main obstacle to making the Beatles' catalogue available online was that EMI "want[s] something we're not prepared to give them". In 2010, the official canon of thirteen Beatles studio albums, Past Masters, and the "Red" and "Blue" greatest-hits albums were made available on iTunes. In 2012, EMI's recorded music operations were sold to Universal Music Group. In order for Universal Music to acquire EMI, the European Union, for antitrust reasons, forced EMI to spin off assets including Parlophone. Universal was allowed to keep the Beatles' recorded music catalogue, managed by Capitol Records under its Capitol Music Group division. The entire original Beatles album catalogue was also reissued on vinyl in 2012; available either individually or as a box set. In 2013, a second volume of BBC recordings, On Air – Live at the BBC Volume 2, was released. That December saw the release of another 59 Beatles recordings on iTunes. The set, titled The Beatles Bootleg Recordings 1963, had the opportunity to gain a 70-year copyright extension conditional on the songs being published at least once before the end of 2013. Apple Records released the recordings on 17 December to prevent them from going into the public domain and had them taken down from iTunes later that same day. Fan reactions to the release were mixed, with one blogger saying "the hardcore Beatles collectors who are trying to obtain everything will already have these." On 26 January 2014, McCartney and Starr performed together at the 56th Annual Grammy Awards, held at the Staples Center in Los Angeles. The following day, The Night That Changed America: A Grammy Salute to The Beatles television special was taped in the Los Angeles Convention Center's West Hall. It aired on 9 February, the exact date of – and at the same time, and on the same network as – the original broadcast of the Beatles' first US television appearance on The Ed Sullivan Show, 50 years earlier. The special included performances of Beatles songs by current artists as well as by McCartney and Starr, archival footage, and interviews with the two surviving ex-Beatles carried out by David Letterman at the Ed Sullivan Theater. In December 2015, the Beatles released their catalogue for streaming on various streaming music services including Spotify and Apple Music. In September 2016, the documentary film The Beatles: Eight Days a Week was released. Directed by Ron Howard, it chronicled the Beatles' career during their touring years from 1961 to 1966, from their performances in Liverpool's the Cavern Club in 1961 to their final concert in San Francisco in 1966. The film was released theatrically on 15 September in the UK and the US, and started streaming on Hulu on 17 September. It received several awards and nominations, including for Best Documentary at the 70th British Academy Film Awards and the Outstanding Documentary or Nonfiction Special at the 69th Primetime Creative Arts Emmy Awards. An expanded, remixed and remastered version of The Beatles at the Hollywood Bowl was released on 9 September, to coincide with the release of the film. On 18 May 2017, Sirius XM Radio launched a 24/7 radio channel, The Beatles Channel. A week later, Sgt. Pepper's Lonely Hearts Club Band was reissued with new stereo mixes and unreleased material for the album's 50th anniversary. Similar box sets were released for The Beatles in November 2018, and Abbey Road in September 2019. On the first week of October 2019, Abbey Road returned to number one on the UK Albums Chart. The Beatles broke their own record for the album with the longest gap between topping the charts as Abbey Road hit the top spot 50 years after its original release. In November 2021, The Beatles: Get Back, a documentary directed by Peter Jackson using footage captured for the Let It Be film, was released on Disney+ as a three-part miniseries. A book also titled The Beatles: Get Back was released on 12 October, ahead of the documentary. A super deluxe version of the Let It Be album was released on 15 October. In January 2022, an album titled Get Back (Rooftop Performance), consisting of newly mixed audio of the Beatles' rooftop performance, was released on streaming services. In 2022, McCartney and Starr collaborated on a new recording of "Let It Be" with Dolly Parton, Peter Frampton and Mick Fleetwood, set to be released on Parton's album Rockstar in November 2023. In October, a special edition of Revolver was released, featuring unreleased demos, studio outtakes, the original mono mix and a new stereo remix using AI de-mixing technology developed by Peter Jackson's WingNut Films, which had previously been used to restore audio for the documentary Get Back. In June 2023, McCartney announced plans to release "the final Beatles record" later in the year, using Jackson's de-mixing technology to extract Lennon's voice from an old demo of a song that he had written as a solo artist. In October 2023, the song was revealed to be "Now and Then", with a physical and digital release date of 2 November 2023. The official music video for "Now and Then" was released the following day, garnering upwards of 8 million views in its first 12 hours, as the song arrived on Spotify's rankings as one of the most-streamed current songs. "Now and Then" debuted simultaneously across music, alternative, news/talk and sports stations. The song's premiere achieved the record for the most radio stations to simulcast a music track. The song became their first UK number one single since 1969. In Icons of Rock: An Encyclopedia of the Legends Who Changed Music Forever, Scott Schinder and Andy Schwartz describe the Beatles' musical evolution: In their initial incarnation as cheerful, wisecracking moptops, the Fab Four revolutionised the sound, style, and attitude of popular music and opened rock and roll's doors to a tidal wave of British rock acts. Their initial impact would have been enough to establish the Beatles as one of their era's most influential cultural forces, but they didn't stop there. Although their initial style was a highly original, irresistibly catchy synthesis of early American rock and roll and R&B, the Beatles spent the rest of the 1960s expanding rock's stylistic frontiers, consistently staking out new musical territory on each release. The band's increasingly sophisticated experimentation encompassed a variety of genres, including folk-rock, country, psychedelia, and baroque pop, without sacrificing the effortless mass appeal of their early work. In The Beatles as Musicians, Walter Everett describes Lennon and McCartney's contrasting motivations and approaches to composition: "McCartney may be said to have constantly developed – as a means to entertain – a focused musical talent with an ear for counterpoint and other aspects of craft in the demonstration of a universally agreed-upon common language that he did much to enrich. Conversely, Lennon's mature music is best appreciated as the daring product of a largely unconscious, searching but undisciplined artistic sensibility." Ian MacDonald describes McCartney as "a natural melodist – a creator of tunes capable of existing apart from their harmony". His melody lines are characterised as primarily "vertical", employing wide, consonant intervals which express his "extrovert energy and optimism". Conversely, Lennon's "sedentary, ironic personality" is reflected in a "horizontal" approach featuring minimal, dissonant intervals and repetitive melodies which rely on their harmonic accompaniment for interest: "Basically a realist, he instinctively kept his melodies close to the rhythms and cadences of speech, colouring his lyrics with bluesy tone and harmony rather than creating tunes that made striking shapes of their own." MacDonald praises Harrison's lead guitar work for the role his "characterful lines and textural colourings" play in supporting Lennon and McCartney's parts, and describes Starr as "the father of modern pop/rock drumming". The Beatles' earliest influences include Elvis Presley, Carl Perkins, Little Richard, Chuck Berry and Gene Vincent. During the Beatles' co-residency with Little Richard at the Star-Club in Hamburg, from April to May 1962, he advised them on the proper technique for performing his songs. Of Presley, Lennon said, "Nothing really affected me until I heard Elvis. If there hadn't been Elvis, there would not have been the Beatles." Chuck Berry was particularly influential in terms of songwriting and lyrics. Lennon noted, "He was well advanced of his time lyric-wise. We all owe a lot to him." Other early influences include Buddy Holly, Eddie Cochran, Roy Orbison, the Everly Brothers and Jerry Lee Lewis. The Beatles continued to absorb influences long after their initial success, often finding new musical and lyrical avenues by listening to their contemporaries, including Bob Dylan, Smokey Robinson and the Miracles, the Who, Frank Zappa, the Lovin' Spoonful, the Byrds and the Beach Boys, whose 1966 album Pet Sounds amazed and inspired McCartney. Referring to the Beach Boys' creative leader, Martin later stated: "No one made a greater impact on the Beatles than Brian [Wilson]." Ravi Shankar, with whom Harrison studied for six weeks in India in late 1966, had a significant effect on his musical development during the band's later years. Originating as a skiffle group, the Beatles quickly embraced 1950s rock and roll and helped pioneer the Merseybeat genre, and their repertoire ultimately expanded to include a broad variety of pop music. Reflecting the range of styles they explored, Lennon said of Beatles for Sale, "You could call our new one a Beatles country-and-western LP", while Gould credits Rubber Soul as "the instrument by which legions of folk-music enthusiasts were coaxed into the camp of pop". Although the 1965 song "Yesterday" was not the first pop record to employ orchestral strings, it marked the group's first recorded use of classical music elements. Gould observes: "The more traditional sound of strings allowed for a fresh appreciation of their talent as composers by listeners who were otherwise allergic to the din of drums and electric guitars." They continued to experiment with string arrangements to various effect; Sgt. Pepper's "She's Leaving Home", for instance, is "cast in the mold of a sentimental Victorian ballad", Gould writes, "its words and music filled with the clichés of musical melodrama". The band's stylistic range expanded in another direction with their 1966 B-side "Rain", described by Martin Strong as "the first overtly psychedelic Beatles record". Other psychedelic numbers followed, such as "Tomorrow Never Knows" (recorded before "Rain"), "Strawberry Fields Forever", "Lucy in the Sky with Diamonds" and "I Am the Walrus". The influence of Indian classical music was evident in Harrison's "The Inner Light", "Love You To" and "Within You Without You" – Gould describes the latter two as attempts "to replicate the raga form in miniature". Innovation was the most striking feature of their creative evolution, according to music historian and pianist Michael Campbell: "'A Day in the Life' encapsulates the art and achievement of the Beatles as well as any single track can. It highlights key features of their music: the sound imagination, the persistence of tuneful melody, and the close coordination between words and music. It represents a new category of song – more sophisticated than pop ... and uniquely innovative. There literally had never before been a song – classical or vernacular – that had blended so many disparate elements so imaginatively." Philosophy professor Bruce Ellis Benson agrees: "the Beatles ... give us a wonderful example of how such far-ranging influences as Celtic music, rhythm and blues, and country and western could be put together in a new way." Author Dominic Pedler describes the way they crossed musical styles: "Far from moving sequentially from one genre to another (as is sometimes conveniently suggested) the group maintained in parallel their mastery of the traditional, catchy chart hit while simultaneously forging rock and dabbling with a wide range of peripheral influences from country to vaudeville. One of these threads was their take on folk music, which would form such essential groundwork for their later collisions with Indian music and philosophy." As the personal relationships between the band members grew increasingly strained, their individual tastes became more apparent. The minimalistic cover artwork for the White Album contrasted with the complexity and diversity of its music, which encompassed Lennon's "Revolution 9" (whose musique concrète approach was influenced by Yoko Ono), Starr's country song "Don't Pass Me By", Harrison's rock ballad "While My Guitar Gently Weeps", and the "proto-metal roar" of McCartney's "Helter Skelter". George Martin's close involvement in his role as producer made him one of the leading candidates for the informal title of the "fifth Beatle". He applied his classical musical training in various ways, and functioned as "an informal music teacher" to the progressing songwriters, according to Gould. Martin suggested to a sceptical McCartney that the arrangement of "Yesterday" should feature a string quartet accompaniment, thereby introducing the Beatles to a "hitherto unsuspected world of classical instrumental colour", in MacDonald's description. Their creative development was also facilitated by Martin's willingness to experiment in response to their suggestions, such as adding "something baroque" to a particular recording. In addition to scoring orchestral arrangements for recordings, Martin often performed on them, playing instruments including piano, organ and brass. Collaborating with Lennon and McCartney required Martin to adapt to their different approaches to songwriting and recording. MacDonald comments, "while [he] worked more naturally with the conventionally articulate McCartney, the challenge of catering to Lennon's intuitive approach generally spurred him to his more original arrangements, of which "Being for the Benefit of Mr. Kite!" is an outstanding example." Martin said of the two composers' distinct songwriting styles and his stabilising influence: Compared with Paul's songs, all of which seemed to keep in some sort of touch with reality, John's had a psychedelic, almost mystical quality ... John's imagery is one of the best things about his work – 'tangerine trees', 'marmalade skies', 'cellophane flowers' ... I always saw him as an aural Salvador Dalí, rather than some drug-ridden record artist. On the other hand, I would be stupid to pretend that drugs didn't figure quite heavily in the Beatles' lives at that time ... they knew that I, in my schoolmasterly role, didn't approve ... Not only was I not into it myself, I couldn't see the need for it; and there's no doubt that, if I too had been on dope, Pepper would never have been the album it was. Perhaps it was the combination of dope and no dope that worked, who knows? Harrison echoed Martin's description of his stabilising role: "I think we just grew through those years together, him as the straight man and us as the loonies; but he was always there for us to interpret our madness – we used to be slightly avant-garde on certain days of the week, and he would be there as the anchor person, to communicate that through the engineers and on to the tape." Making innovative use of technology while expanding the possibilities of recorded music, the Beatles urged experimentation by Martin and his recording engineers. Seeking ways to put chance occurrences to creative use, accidental guitar feedback, a resonating glass bottle, a tape loaded the wrong way round so that it played backwards – any of these might be incorporated into their music. Their desire to create new sounds on every new recording, combined with Martin's arranging abilities and the studio expertise of EMI staff engineers Norman Smith, Ken Townsend and Geoff Emerick, all contributed significantly to their records from Rubber Soul and, especially, Revolver onwards. Along with innovative studio techniques such as sound effects, unconventional microphone placements, tape loops, double tracking and vari-speed recording, the Beatles augmented their songs with instruments that were unconventional in rock music at the time. These included string and brass ensembles as well as Indian instruments such as the sitar in "Norwegian Wood" and the swarmandal in "Strawberry Fields Forever". They also used novel electronic instruments such as the Mellotron, with which McCartney supplied the flute voices on the "Strawberry Fields Forever" intro, and the clavioline, an electronic keyboard that created the unusual oboe-like sound on "Baby, You're a Rich Man". Former Rolling Stone associate editor Robert Greenfield compared the Beatles to Picasso, as "artists who broke through the constraints of their time period to come up with something that was unique and original ... [I]n the form of popular music, no one will ever be more revolutionary, more creative and more distinctive ..." The British poet Philip Larkin described their work as "an enchanting and intoxicating hybrid of Negro rock-and-roll with their own adolescent romanticism", and "the first advance in popular music since the War". The Beatles' 1964 arrival in the US is credited with initiating the album era; the music historian Joel Whitburn says that LP sales soon "exploded and eventually outpaced the sales and releases of singles" in the music industry. They not only sparked the British Invasion of the US, they became a globally influential phenomenon as well. From the 1920s, the US had dominated popular entertainment culture throughout much of the world, via Hollywood films, jazz, the music of Broadway and Tin Pan Alley and, later, the rock and roll that first emerged in Memphis, Tennessee. The Beatles are regarded as British cultural icons, with young adults from abroad naming the band among a group of people whom they most associated with UK culture. Their musical innovations and commercial success inspired musicians worldwide. Many artists have acknowledged the Beatles' influence and enjoyed chart success with covers of their songs. On radio, their arrival marked the beginning of a new era; in 1968 the programme director of New York's WABC radio station forbade his DJs from playing any "pre-Beatles" music, marking the defining line of what would be considered oldies on American radio. They helped to redefine the album as something more than just a few hits padded out with "filler", and they were primary innovators of the modern music video. The Shea Stadium show with which they opened their 1965 North American tour attracted an estimated 55,600 people, then the largest audience in concert history; Spitz describes the event as a "major breakthrough ... a giant step toward reshaping the concert business". Emulation of their clothing and especially their hairstyles, which became a mark of rebellion, had a global impact on fashion. According to Gould, the Beatles changed the way people listened to popular music and experienced its role in their lives. From what began as the Beatlemania fad, the group's popularity grew into what was seen as an embodiment of sociocultural movements of the decade. As icons of the 1960s counterculture, Gould continues, they became a catalyst for bohemianism and activism in various social and political arenas, fuelling movements such as women's liberation, gay liberation and environmentalism. According to Peter Lavezzoli, after the "more popular than Jesus" controversy in 1966, the Beatles felt considerable pressure to say the right things and "began a concerted effort to spread a message of wisdom and higher consciousness". Other commentators such as Mikal Gilmore and Todd Leopold have traced the inception of their socio-cultural impact earlier, interpreting even the Beatlemania period, particularly on their first visit to the US, as a key moment in the development of generational awareness. Referring to their appearance on the Ed Sullivan Show Leopold states: "In many ways, the Sullivan appearance marked the beginning of a cultural revolution ... The Beatles were like aliens dropped into the United States of 1964." According to Gilmore: Elvis Presley had shown us how rebellion could be fashioned into eye-opening style; the Beatles were showing us how style could have the impact of cultural revelation – or at least how a pop vision might be forged into an unimpeachable consensus. Established in 2009, Global Beatles Day is an annual holiday on 25 June each year that honours and celebrates the ideals of the Beatles. The date was chosen to commemorate the date the group participated in the BBC programme Our World in 1967, performing "All You Need Is Love" broadcast to an international audience. In 1965, Queen Elizabeth II appointed Lennon, McCartney, Harrison and Starr Members of the Order of the British Empire (MBE). The Beatles won the 1971 Academy Award for Best Original Song Score for the film Let It Be (1970). The recipients of seven Grammy Awards and fifteen Ivor Novello Awards, the Beatles have six Diamond albums, as well as 20 Multi-Platinum albums, 16 Platinum albums and six Gold albums in the US. In the UK, the Beatles have four Multi-Platinum albums, four Platinum albums, eight Gold albums and one Silver album. They were inducted into the Rock and Roll Hall of Fame in 1988. The best-selling band in history, the Beatles have sold more than 600 million units as of 2012. From 1991 to 2009 the Beatles sold 57 million albums in United States, according to Nielsen Soundscan. They have had more number-one albums on the UK charts, fifteen, and sold more singles in the UK, 21.9 million, than any other act. In 2004, Rolling Stone ranked the Beatles as the most significant and influential rock music artists of the last 50 years. They ranked number one on Billboard magazine's list of the all-time most successful Hot 100 artists, released in 2008 to celebrate the US singles chart's 50th anniversary. As of 2017, they hold the record for most number-one hits on the Billboard Hot 100, with twenty. The Recording Industry Association of America certifies that the Beatles have sold 183 million units in the US, more than any other artist. They were collectively included in Time magazine's compilation of the 20th century's 100 most influential people. In 2014, they received the Grammy Lifetime Achievement Award. On 16 January each year, beginning in 2001, people celebrate World Beatles Day under UNESCO. This date has direct relation to the opening of the Cavern Club in 1957. In 2007, the Beatles became the first band to feature on a series of UK postage stamps issued by the Royal Mail. Earlier in 1999, the United States Postal Service issued a stamp dedicated to the Beatles and Yellow Submarine. The Beatles have a core catalogue consisting of thirteen studio albums and a compilation of UK singles and EP tracks: Until 1969, the Beatles' catalogue was published almost exclusively by Northern Songs Ltd, a company formed in February 1963 by music publisher Dick James specifically for Lennon and McCartney, though it later acquired songs by other artists. The company was organised with James and his partner, Emmanuel Silver, owning a controlling interest, variously described as 51% or 50% plus one share. McCartney had 20%. Reports again vary concerning Lennon's portion – 19 or 20% – and Brian Epstein's – 9 or 10% – which he received in lieu of a 25% band management fee. In 1965, the company went public. Five million shares were created, of which the original principals retained 3.75 million. James and Silver each received 937,500 shares (18.75% of 5 million); Lennon and McCartney each received 750,000 shares (15%); and Epstein's management company, NEMS Enterprises, received 375,000 shares (7.5%). Of the 1.25 million shares put up for sale, Harrison and Starr each acquired 40,000. At the time of the stock offering, Lennon and McCartney renewed their three-year publishing contracts, binding them to Northern Songs until 1973. Harrison created Harrisongs to represent his Beatles compositions, but signed a three-year contract with Northern Songs that gave it the copyright to his work through March 1968, which included "Taxman" and "Within You Without You". The songs on which Starr received co-writing credit before 1968, such as "What Goes On" and "Flying", were also Northern Songs copyrights. Harrison did not renew his contract with Northern Songs when it ended, signing instead with Apple Publishing while retaining the copyright to his work from that point on. Harrison thus owns the rights to his later Beatles songs such as "While My Guitar Gently Weeps" and "Something". That year, as well, Starr created Startling Music, which holds the rights to his Beatles compositions, "Don't Pass Me By" and "Octopus's Garden". In March 1969, James arranged to sell his and his partner's shares of Northern Songs to the British broadcasting company Associated Television (ATV), founded by impresario Lew Grade, without first informing the Beatles. The band then made a bid to gain a controlling interest by attempting to work out a deal with a consortium of London brokerage firms that had accumulated a 14% holding. The deal collapsed over the objections of Lennon, who declared, "I'm sick of being fucked about by men in suits sitting on their fat arses in the City." By the end of May, ATV had acquired a majority stake in Northern Songs, controlling nearly the entire Lennon–McCartney catalogue, as well as any future material until 1973. In frustration, Lennon and McCartney sold their shares to ATV in late October 1969. In 1981, financial losses by ATV's parent company, Associated Communications Corporation (ACC), led it to attempt to sell its music division. According to authors Brian Southall and Rupert Perry, Grade contacted McCartney, offering ATV Music and Northern Songs for $30 million. According to an account McCartney gave in 1995, he met with Grade and explained he was interested solely in the Northern Songs catalogue if Grade were ever willing to "separate off" that portion of ATV Music. Soon afterwards, Grade offered to sell him Northern Songs for £20 million, giving the ex-Beatle "a week or so" to decide. By McCartney's account, he and Ono countered with a £5 million bid that was rejected. According to reports at the time, Grade refused to separate Northern Songs and turned down an offer of £21–25 million from McCartney and Ono for Northern Songs. In 1982, ACC was acquired in a takeover by Australian business magnate Robert Holmes à Court for £60 million. In 1985, Michael Jackson purchased ATV for a reported $47.5 million. The acquisition gave him control over the publishing rights to more than 200 Beatles songs, as well as 40,000 other copyrights. In 1995, in a deal that earned him a reported $110 million, Jackson merged his music publishing business with Sony, creating a new company, Sony/ATV Music Publishing, in which he held a 50% stake. The merger made the new company, then valued at over half a billion dollars, the third-largest music publisher in the world. In 2016, Sony acquired Jackson's share of Sony/ATV from the Jackson estate for $750 million. Despite the lack of publishing rights to most of their songs, Lennon's estate and McCartney continue to receive their respective shares of the writers' royalties, which together are 331⁄3% of total commercial proceeds in the US and which vary elsewhere around the world between 50 and 55%. Two of Lennon and McCartney's earliest songs – "Love Me Do" and "P.S. I Love You" – were published by an EMI subsidiary, Ardmore & Beechwood, before they signed with James. McCartney acquired their publishing rights from Ardmore in 1978, and they are the only two Beatles songs owned by McCartney's company MPL Communications. On 18 January 2017, McCartney filed a suit in the United States district court against Sony/ATV Music Publishing seeking to reclaim ownership of his share of the Lennon–McCartney song catalogue beginning in 2018. Under US copyright law, for works published before 1978 the author can reclaim copyrights assigned to a publisher after 56 years. McCartney and Sony agreed to a confidential settlement in June 2017. Fictionalised Documentaries and filmed performances
{{redirect|Rebirth|the Lil Wayne album|Rebirth (album)}} '''Reincarnation''' is the name of the idea that people are born again in another body after they [[death|die]] and this cycle continues over many lifetimes. [[Rebirth]] or "transmigration" is the preferred term for those believers who do not believe in eternal [[soul]]s. Many [[Hindu]]s, [[Jainism|Jainists]], Celtic pagans, [[Buddhism|Buddhists]], and people who follow some African religions believe in reincarnation or rebirth/ transmigration. Carnate means “of flesh”, and reincarnate means to “reenter the flesh". ==Belief in reincarnation or rebirth/transmigration== The belief is held in many religions except [[Islam]] and [[Christianity]], although 20 to 30 percent of Christians in western countries also believe in reincarnation.<ref name=book/> Ethnologists have documented the belief among nearly all the traditional religions of ethnic groups in Africa, North and South America, and Australia/Oceania. The tribes of north-west North America continue to believe in reincarnation despite negative attitudes towards it on the part of Christian missionaries and churches.<ref name = "book">{{cite book| last = Henry| first = Jane| title = Parapsychology: Research on Exceptional Experiences| url = https://books.google.com/books?id=EaIhapm-4UgC| year = 2005| publisher = Psychology Press| isbn = 978-0-415-21360-8| page = 224 }}</ref> == Reincarnation in Hinduism == [[File:Reincarnation AS.jpg|190px|thumb|Drawing of reincarnation in Hindu art]] [[Hindu]]s believe in reincarnation, the process where the [[soul]] repeatedly takes on a physical body through being born on [[Earth]]. Ancient scriptures of [[Hinduism]] starting around 700&nbsp;BC teach that the [[soul]], or immortal “self”, takes [[birth]] time and time again. The soul survives and continues its long journey until it is one with [[God]]. Hindus believe that the soul never dies, but inhabits one [[body]] after another during its evolutionary journey guided by karma. [[Karma]] (literally: [[action]]) is the sum of one's actions, and the [[force]] that determines one's next reincarnation. The soul evolves from immaturity to spiritual [[illumination]]. Therefore, each reincarnating soul chooses a home and a family which can best fulfill its next step of learning and maturation. Each life on [[Earth]] is similar to a class in school. Maturation of the [[soul]] on Earth means fulfilling its worldly desires, which can only be experienced through a body.<ref>See Bhagavad Gita XVI.8-20</ref> At death the [[soul]] leaves the physical body. But the soul does not die. It lives on in a subtle body called the astral body. The astral body exists in a nonphysical [[dimension]] called the astral plane. Here the [[soul]] continues to have experiences until it is born again in another physical body as a baby.<ref>Satguru [[Sivaya Subramuniyaswami]], "''Ten Questions people ask About Hinduism …and ten terrific answers!''" (p. 4) [http://www.himalayanacademy.com/basics/tenq/hindu10questions.pdf] {{Webarchive|url=https://web.archive.org/web/20070411180040/http://www.himalayanacademy.com/basics/tenq/hindu10questions.pdf |date=2007-04-11 }}</ref> After many lifetimes of following [[dharma]] (right way of living), the soul is fully matured in love, wisdom and knowledge of [[God]]. There is no longer a need for physical [[birth]], for all lessons have been learned, all karmas fulfilled. When all desire has vanished, the person will not be born again anymore.<ref>{{cite book| last = Rinehart| first = Robin| title = Contemporary Hinduism: Ritual, Culture, and Practice| url = https://archive.org/details/contemporaryhind0000unse_x1k0| year = 2004| publisher = ABC-CLIO| isbn = 1-57607-905-8 }}</ref> ==Reincarnation research== {{Main|Reincarnation research}} [[Ian Stevenson]] found that the best research evidence supporting the belief in reincarnation comes from the cases of young children who, typically between the ages of 2 and 5, make statements about a previous life they claim to have had before being born.<ref name = "st225">{{cite book| last = Henry| first = Jane| title = Parapsychology: Research on Exceptional Experiences| url = https://books.google.com/books?id=EaIhapm-4UgC| year = 2005| publisher = Psychology Press| isbn = 978-0-415-21360-8| page = 225 }}</ref> Some 35 per cent of the children also have birthmarks or birth defects which often correspond to injuries or illness experienced by the deceased person who the subject remembers. Medical documents have confirmed this correspondence in more than forty cases.<ref name="book"/> ==Related pages== * [[Resurrection]] * [[Life Before Life]] * [[Ian Stevenson]] == References == {{reflist}} ==Other websites== *[https://psi-encyclopedia.spr.ac.uk/articles/criticisms-reincarnation-case-studies#Reincarnation_Case_Studies Criticisms of Reincarnation Case Studies] PSI Encyclopedia [[Category:Theology]] [[Category:Cultural studies]] [[Category:Afterlife]]
Reincarnation, also known as rebirth or transmigration, is the philosophical or religious concept that the non-physical essence of a living being begins a new life in a different physical form or body after biological death. In most beliefs involving reincarnation, the soul of a human being is immortal and does not disperse after the physical body has perished. Upon death, the soul merely becomes transmigrated into a newborn baby or an animal to continue its immortality. The term transmigration means the passing of a soul from one body to another after death. Reincarnation (punarjanma) is a central tenet of the Indian religions such as Hinduism, Buddhism, Jainism, and Sikhism; although there are Hindu groups who do not believe in reincarnation, instead believing in an afterlife. In various forms, it occurs as an esoteric belief in many streams of Judaism, certain pagan religions including Wicca, and some beliefs of the Indigenous peoples of the Americas and Indigenous Australians (though most believe in an afterlife or spirit world). A belief in the soul's rebirth or migration (metempsychosis) was expressed by certain Ancient Greek historical figures, such as Pythagoras, Socrates, and Plato. Although the majority of denominations within Christianity and Islam do not believe that individuals reincarnate, particular groups within these religions do refer to reincarnation; these groups include the mainstream historical and contemporary followers of Cathars, Alawites, the Druze, and the Rosicrucians. The historical relations between these sects and the beliefs about reincarnation that were characteristic of Neoplatonism, Orphism, Hermeticism, Manichaenism, and Gnosticism of the Roman era as well as the Indian religions have been the subject of recent scholarly research. In recent decades, many Europeans and North Americans have developed an interest in reincarnation, and many contemporary works mention it. The word reincarnation derives from a Latin term that literally means 'entering the flesh again'. Reincarnation refers to the belief that an aspect of every human being (or all living beings in some cultures) continues to exist after death. This aspect may be the soul or mind or consciousness or something transcendent which is reborn in an interconnected cycle of existence; the transmigration belief varies by culture, and is envisioned to be in the form of a newly born human being, or animal, or plant, or spirit, or as a being in some other non-human realm of existence. An alternative term is transmigration, implying migration from one life (body) to another. The term has been used by modern philosophers such as Kurt Gödel and has entered the English language. The Greek equivalent to reincarnation, metempsychosis (μετεμψύχωσις), derives from meta ('change') and empsykhoun ('to put a soul into'), a term attributed to Pythagoras. Another Greek term sometimes used synonymously is palingenesis, 'being born again'. Rebirth is a key concept found in major Indian religions, and discussed using various terms. Reincarnation, or Punarjanman (Sanskrit: पुनर्जन्मन्, 'rebirth, transmigration'), is discussed in the ancient Sanskrit texts of Hinduism, Buddhism, and Jainism, with many alternate terms such as punarāvṛtti (पुनरावृत्ति), punarājāti (पुनराजाति), punarjīvātu (पुनर्जीवातु), punarbhava (पुनर्भव), āgati-gati (आगति-गति, common in Buddhist Pali text), nibbattin (निब्बत्तिन्), upapatti (उपपत्ति), and uppajjana (उप्पज्जन). These religions believe that this reincarnation is cyclic and an endless Saṃsāra, unless one gains spiritual insights that ends this cycle leading to liberation. The reincarnation concept is considered in Indian religions as a step that starts each "cycle of aimless drifting, wandering or mundane existence", but one that is an opportunity to seek spiritual liberation through ethical living and a variety of meditative, yogic (marga), or other spiritual practices. They consider the release from the cycle of reincarnations as the ultimate spiritual goal, and call the liberation by terms such as moksha, nirvana, mukti and kaivalya. However, the Buddhist, Hindu and Jain traditions have differed, since ancient times, in their assumptions and in their details on what reincarnates, how reincarnation occurs and what leads to liberation. Gilgul, Gilgul neshamot, or Gilgulei Ha Neshamot (Hebrew: גלגול הנשמות) is the concept of reincarnation in Kabbalistic Judaism, found in much Yiddish literature among Ashkenazi Jews. Gilgul means 'cycle' and neshamot is 'souls'. Kabbalistic reincarnation says that humans reincarnate only to humans unless YHWH/Ein Sof/God chooses. The origins of the notion of reincarnation are obscure. Discussion of the subject appears in the philosophical traditions of India. The Greek Pre-Socratics discussed reincarnation, and the Celtic druids are also reported to have taught a doctrine of reincarnation. The concepts of the cycle of birth and death, saṁsāra, and liberation partly derive from ascetic traditions that arose in India around the middle of the first millennium BCE. The first textual references to the idea of reincarnation appear in the Rigveda, Yajurveda and Upanishads of the late Vedic period (c. 1100 – c. 500 BCE), predating the Buddha and Mahavira. Though no direct evidence of this has been found, the tribes of the Ganges valley or the Dravidian traditions of South India have been proposed as another early source of reincarnation beliefs. The idea of reincarnation, saṁsāra, did exist in the early Vedic religions. The early Vedas does mention the doctrine of karma and rebirth. It is in the early Upanishads, which are pre-Buddha and pre-Mahavira, where these ideas are developed and described in a general way. Detailed descriptions first appear around the mid-1st millennium BCE in diverse traditions, including Buddhism, Jainism and various schools of Hindu philosophy, each of which gave unique expression to the general principle. Sangam literature connotes the ancient Tamil literature and is the earliest known literature of South India. The Tamil tradition and legends link it to three literary gatherings around Madurai. According to Kamil Zvelebil, a Tamil literature and history scholar, the most acceptable range for the Sangam literature is 100 BCE to 250 CE, based on the linguistic, prosodic and quasi-historic allusions within the texts and the colophons. There are several mentions of rebirth and moksha in the Purananuru. The text explains Hindu rituals surrounding death such as making riceballs called pinda and cremation. The text states that good souls get a place in Indraloka where Indra welcomes them. The texts of ancient Jainism that have survived into the modern era are post-Mahavira, likely from the last centuries of the first millennium BCE, and extensively mention rebirth and karma doctrines. The Jaina philosophy assumes that the soul (jiva in Jainism; atman in Hinduism) exists and is eternal, passing through cycles of transmigration and rebirth. After death, reincarnation into a new body is asserted to be instantaneous in early Jaina texts. Depending upon the accumulated karma, rebirth occurs into a higher or lower bodily form, either in heaven or hell or earthly realm. No bodily form is permanent: everyone dies and reincarnates further. Liberation (kevalya) from reincarnation is possible, however, through removing and ending karmic accumulations to one's soul. From the early stages of Jainism on, a human being was considered the highest mortal being, with the potential to achieve liberation, particularly through asceticism. The early Buddhist texts discuss rebirth as part of the doctrine of saṃsāra. This asserts that the nature of existence is a "suffering-laden cycle of life, death, and rebirth, without beginning or end". Also referred to as the wheel of existence (Bhavacakra), it is often mentioned in Buddhist texts with the term punarbhava (rebirth, re-becoming). Liberation from this cycle of existence, Nirvana, is the foundation and the most important purpose of Buddhism. Buddhist texts also assert that an enlightened person knows his previous births, a knowledge achieved through high levels of meditative concentration. Tibetan Buddhism discusses death, bardo (an intermediate state), and rebirth in texts such as the Tibetan Book of the Dead. While Nirvana is taught as the ultimate goal in the Theravadin Buddhism, and is essential to Mahayana Buddhism, the vast majority of contemporary lay Buddhists focus on accumulating good karma and acquiring merit to achieve a better reincarnation in the next life. In early Buddhist traditions, saṃsāra cosmology consisted of five realms through which the wheel of existence cycled. This included hells (niraya), hungry ghosts (pretas), animals (tiryaka), humans (manushya), and gods (devas, heavenly). In latter Buddhist traditions, this list grew to a list of six realms of rebirth, adding demigods (asuras). The earliest layers of Vedic text incorporate the concept of life, followed by an afterlife in heaven and hell based on cumulative virtues (merit) or vices (demerit). However, the ancient Vedic rishis challenged this idea of afterlife as simplistic, because people do not live equally moral or immoral lives. Between generally virtuous lives, some are more virtuous; while evil too has degrees, and the texts assert that it would be unfair for people, with varying degrees of virtue or vices, to end up in heaven or hell, in "either or" and disproportionate manner irrespective of how virtuous or vicious their lives were. They introduced the idea of an afterlife in heaven or hell in proportion to one's merit. Early texts of Hinduism, Buddhism and Jainism share the concepts and terminology related to reincarnation. They also emphasize similar virtuous practices and karma as necessary for liberation and what influences future rebirths. For example, all three discuss various virtues—sometimes grouped as Yamas and Niyamas—such as non-violence, truthfulness, non-stealing, non-possessiveness, compassion for all living beings, charity and many others. Hinduism, Buddhism and Jainism disagree in their assumptions and theories about rebirth. Hinduism relies on its foundational assumption that 'soul, Self exists' (atman or attā), in contrast to Buddhist assumption that there is 'no soul, no Self' (anatta or anatman). Hindu traditions consider soul to be the unchanging eternal essence of a living being, and what journeys across reincarnations until it attains self-knowledge. Buddhism, in contrast, asserts a rebirth theory without a Self, and considers realization of non-Self or Emptiness as Nirvana (nibbana). Thus Buddhism and Hinduism have a very different view on whether a self or soul exists, which impacts the details of their respective rebirth theories. The reincarnation doctrine in Jainism differs from those in Buddhism, even though both are non-theistic Sramana traditions. Jainism, in contrast to Buddhism, accepts the foundational assumption that soul exists (Jiva) and asserts this soul is involved in the rebirth mechanism. Further, Jainism considers asceticism as an important means to spiritual liberation that ends all reincarnation, while Buddhism does not. Early Greek discussion of the concept dates to the sixth century BCE. An early Greek thinker known to have considered rebirth is Pherecydes of Syros (fl. 540 BCE). His younger contemporary Pythagoras (c. 570–c. 495 BCE), its first famous exponent, instituted societies for its diffusion. Some authorities believe that Pythagoras was Pherecydes' pupil, others that Pythagoras took up the idea of reincarnation from the doctrine of Orphism, a Thracian religion, or brought the teaching from India. Plato (428/427–348/347 BCE) presented accounts of reincarnation in his works, particularly the Myth of Er, where Plato makes Socrates tell how Er, the son of Armenius, miraculously returned to life on the twelfth day after death and recounted the secrets of the other world. There are myths and theories to the same effect in other dialogues, in the Chariot allegory of the Phaedrus, in the Meno, Timaeus and Laws. The soul, once separated from the body, spends an indeterminate amount of time in the intelligible realm (see The Allegory of the Cave in The Republic) and then assumes another body. In the Timaeus, Plato believes that the soul moves from body to body without any distinct reward-or-punishment phase between lives, because the reincarnation is itself a punishment or reward for how a person has lived. In Phaedo, Plato has his teacher Socrates, prior to his death, state: "I am confident that there truly is such a thing as living again, and that the living spring from the dead." However, Xenophon does not mention Socrates as believing in reincarnation, and Plato may have systematized Socrates' thought with concepts he took directly from Pythagoreanism or Orphism. Recent scholars have come to see that Plato has multiple reasons for the belief in reincarnation. One argument concerns the theory of reincarnation's usefulness for explaining why non-human animals exist: they are former humans, being punished for their vices; Plato gives this argument at the end of the Timaeus. The Orphic religion, which taught reincarnation, about the sixth century BCE, produced a copious literature. Orpheus, its legendary founder, is said to have taught that the immortal soul aspires to freedom while the body holds it prisoner. The wheel of birth revolves, the soul alternates between freedom and captivity round the wide circle of necessity. Orpheus proclaimed the need of the grace of the gods, Dionysus in particular, and of self-purification until the soul has completed the spiral ascent of destiny to live forever. An association between Pythagorean philosophy and reincarnation was routinely accepted throughout antiquity, as Pythagoras also taught about reincarnation. However, unlike the Orphics, who considered metempsychosis a cycle of grief that could be escaped by attaining liberation from it, Pythagoras seems to postulate an eternal, neutral reincarnation where subsequent lives would not be conditioned by any action done in the previous. In later Greek literature the doctrine is mentioned in a fragment of Menander and satirized by Lucian. In Roman literature it is found as early as Ennius, who, in a lost passage of his Annals, told how he had seen Homer in a dream, who had assured him that the same soul which had animated both the poets had once belonged to a peacock. Persius in his satires (vi. 9) laughs at this; it is referred to also by Lucretius and Horace. Virgil works the idea into his account of the Underworld in the sixth book of the Aeneid. It persists down to the late classic thinkers, Plotinus and the other Neoplatonists. In the Hermetica, a Graeco-Egyptian series of writings on cosmology and spirituality attributed to Hermes Trismegistus/Thoth, the doctrine of reincarnation is central. In the first century BCE Alexander Cornelius Polyhistor wrote: The Pythagorean doctrine prevails among the Gauls' teaching that the souls of men are immortal, and that after a fixed number of years they will enter into another body. Julius Caesar recorded that the druids of Gaul, Britain and Ireland had metempsychosis as one of their core doctrines: The principal point of their doctrine is that the soul does not die and that after death it passes from one body into another... the main object of all education is, in their opinion, to imbue their scholars with a firm belief in the indestructibility of the human soul, which, according to their belief, merely passes at death from one tenement to another; for by such doctrine alone, they say, which robs death of all its terrors, can the highest form of human courage be developed. Diodorus also recorded the Gaul belief that human souls were immortal, and that after a prescribed number of years they would commence upon a new life in another body. He added that Gauls had the custom of casting letters to their deceased upon the funeral pyres, through which the dead would be able to read them. Valerius Maximus also recounted they had the custom of lending sums of money to each other which would be repayable in the next world. This was mentioned by Pomponius Mela, who also recorded Gauls buried or burnt with them things they would need in a next life, to the point some would jump into the funeral piles of their relatives in order to cohabit in the new life with them. Hippolytus of Rome believed the Gauls had been taught the doctrine of reincarnation by a slave of Pythagoras named Zalmoxis. Conversely, Clement of Alexandria believed Pythagoras himself had learned it from the Celts and not the opposite, claiming he had been taught by Galatian Gauls, Hindu priests and Zoroastrians. However, author T. D. Kendrick rejected a real connection between Pythagoras and the Celtic idea reincarnation, noting their beliefs to have substantial differences, and any contact to be historically unlikely. Nonetheless, he proposed the possibility of an ancient common source, also related to the Orphic religion and Thracian systems of belief. Surviving texts indicate that there was a belief in rebirth in Germanic paganism. Examples include figures from eddic poetry and sagas, potentially by way of a process of naming and/or through the family line. Scholars have discussed the implications of these attestations and proposed theories regarding belief in reincarnation among the Germanic peoples prior to Christianization and potentially to some extent in folk belief thereafter. The belief in reincarnation developed among Jewish mystics in the Medieval World, among whom differing explanations were given of the afterlife, although with a universal belief in an immortal soul. It was explicitly rejected by Saadiah Gaon. Today, reincarnation is an esoteric belief within many streams of modern Judaism. Kabbalah teaches a belief in gilgul, transmigration of souls, and hence the belief in reincarnation is universal in Hasidic Judaism, which regards the Kabbalah as sacred and authoritative, and is also sometimes held as an esoteric belief within other strains of Orthodox Judaism. In Judaism, the Zohar, first published in the 13th century, discusses reincarnation at length, especially in the Torah portion "Balak." The most comprehensive kabbalistic work on reincarnation, Shaar HaGilgulim, was written by Chaim Vital, based on the teachings of his mentor, the 16th-century kabbalist Isaac Luria, who was said to know the past lives of each person through his semi-prophetic abilities. The 18th-century Lithuanian master scholar and kabbalist, Elijah of Vilna, known as the Vilna Gaon, authored a commentary on the biblical Book of Jonah as an allegory of reincarnation. The practice of conversion to Judaism is sometimes understood within Orthodox Judaism in terms of reincarnation. According to this school of thought in Judaism, when non-Jews are drawn to Judaism, it is because they had been Jews in a former life. Such souls may "wander among nations" through multiple lives, until they find their way back to Judaism, including through finding themselves born in a gentile family with a "lost" Jewish ancestor. There is an extensive literature of Jewish folk and traditional stories that refer to reincarnation. In Greco-Roman thought, the concept of metempsychosis disappeared with the rise of Early Christianity, reincarnation being incompatible with the Christian core doctrine of salvation of the faithful after death. It has been suggested that some of the early Church Fathers, especially Origen, still entertained a belief in the possibility of reincarnation, but evidence is tenuous, and the writings of Origen as they have come down to us speak explicitly against it. Hebrews 9:27 states that men "die once, but after this the judgement". Reincarnationism or biblical reincarnation is the belief that certain people are or can be reincarnations of biblical figures, such as Jesus Christ and the Virgin Mary. Some Christians believe that certain New Testament figures are reincarnations of Old Testament figures. For example, John the Baptist is believed by some to be a reincarnation of the prophet Elijah, and a few take this further by suggesting Jesus was the reincarnation of Elijah's disciple Elisha. Other Christians believe the Second Coming of Jesus was fulfilled by reincarnation. Sun Myung Moon, the founder of the Unification Church, considered himself to be the fulfillment of Jesus' return. The Catholic Church does not believe in reincarnation, which it regards as being incompatible with death. Nonetheless, the leaders of certain sects in the church have taught that they are reincarnations of Mary - for example, Marie-Paule Giguère of the Army of Mary and Maria Franciszka of the former Mariavites. The Congregation for the Doctrine of the Faith excommunicated the Army of Mary for teaching heresy, including reincarnationism. Several Gnostic sects professed reincarnation. The Sethians and followers of Valentinus believed in it. The followers of Bardaisan of Mesopotamia, a sect of the second century deemed heretical by the Catholic Church, drew upon Chaldean astrology, to which Bardaisan's son Harmonius, educated in Athens, added Greek ideas including a sort of metempsychosis. Another such teacher was Basilides (132–? CE/AD), known to us through the criticisms of Irenaeus and the work of Clement of Alexandria (see also Neoplatonism and Gnosticism and Buddhism and Gnosticism). In the third Christian century Manichaeism spread both east and west from Babylonia, then within the Sassanid Empire, where its founder Mani lived about 216–276. Manichaean monasteries existed in Rome in 312 AD. Noting Mani's early travels to the Kushan Empire and other Buddhist influences in Manichaeism, Richard Foltz attributes Mani's teaching of reincarnation to Buddhist influence. However the inter-relation of Manicheanism, Orphism, Gnosticism and neo-Platonism is far from clear. Taoist documents from as early as the Han Dynasty claimed that Lao Tzu appeared on earth as different persons in different times beginning in the legendary era of Three Sovereigns and Five Emperors. The (ca. third century BC) Chuang Tzu states: "Birth is not a beginning; death is not an end. There is existence without limitation; there is continuity without a starting-point. Existence without limitation is Space. Continuity without a starting point is Time. There is birth, there is death, there is issuing forth, there is entering in." Around the 11–12th century in Europe, several reincarnationist movements were persecuted as heresies, through the establishment of the Inquisition in the Latin west. These included the Cathar, Paterene or Albigensian church of western Europe, the Paulician movement, which arose in Armenia, and the Bogomils in Bulgaria. Christian sects such as the Bogomils and the Cathars, who professed reincarnation and other gnostic beliefs, were referred to as "Manichaean", and are today sometimes described by scholars as "Neo-Manichaean". As there is no known Manichaean mythology or terminology in the writings of these groups there has been some dispute among historians as to whether these groups truly were descendants of Manichaeism. While reincarnation has been a matter of faith in some communities from an early date it has also frequently been argued for on principle, as Plato does when he argues that the number of souls must be finite because souls are indestructible, Benjamin Franklin held a similar view. Sometimes such convictions, as in Socrates' case, arise from a more general personal faith, at other times from anecdotal evidence such as Plato makes Socrates offer in the Myth of Er. During the Renaissance translations of Plato, the Hermetica and other works fostered new European interest in reincarnation. Marsilio Ficino argued that Plato's references to reincarnation were intended allegorically, Shakespeare alluded to the doctrine of reincarnation but Giordano Bruno was burned at the stake by authorities after being found guilty of heresy by the Roman Inquisition for his teachings. But the Greek philosophical works remained available and, particularly in north Europe, were discussed by groups such as the Cambridge Platonists. Emanuel Swedenborg believed that we leave the physical world once, but then go through several lives in the spiritual world—a kind of hybrid of Christian tradition and the popular view of reincarnation. By the 19th century the philosophers Schopenhauer and Nietzsche could access the Indian scriptures for discussion of the doctrine of reincarnation, which recommended itself to the American Transcendentalists Henry David Thoreau, Walt Whitman and Ralph Waldo Emerson and was adapted by Francis Bowen into Christian Metempsychosis. By the early 20th century, interest in reincarnation had been introduced into the nascent discipline of psychology, largely due to the influence of William James, who raised aspects of the philosophy of mind, comparative religion, the psychology of religious experience and the nature of empiricism. James was influential in the founding of the American Society for Psychical Research (ASPR) in New York City in 1885, three years after the British Society for Psychical Research (SPR) was inaugurated in London, leading to systematic, critical investigation of paranormal phenomena. Famous World War II American General George Patton was a strong believer in reincarnation, believing, among other things, he was a reincarnation of the Carthaginian General Hannibal. At this time popular awareness of the idea of reincarnation was boosted by the Theosophical Society's dissemination of systematised and universalised Indian concepts and also by the influence of magical societies like The Golden Dawn. Notable personalities like Annie Besant, W. B. Yeats and Dion Fortune made the subject almost as familiar an element of the popular culture of the west as of the east. By 1924 the subject could be satirised in popular children's books. Humorist Don Marquis created a fictional cat named Mehitabel who claimed to be a reincarnation of Queen Cleopatra. Théodore Flournoy was among the first to study a claim of past-life recall in the course of his investigation of the medium Hélène Smith, published in 1900, in which he defined the possibility of cryptomnesia in such accounts. Carl Gustav Jung, like Flournoy based in Switzerland, also emulated him in his thesis based on a study of cryptomnesia in psychism. Later Jung would emphasise the importance of the persistence of memory and ego in psychological study of reincarnation: "This concept of rebirth necessarily implies the continuity of personality... (that) one is able, at least potentially, to remember that one has lived through previous existences, and that these existences were one's own...." Hypnosis, used in psychoanalysis for retrieving forgotten memories, was eventually tried as a means of studying the phenomenon of past life recall. More recently, many people in the West have developed an interest in and acceptance of reincarnation. Many new religious movements include reincarnation among their beliefs, e.g. modern Neopagans, Spiritism, Astara, Dianetics, and Scientology. Many esoteric philosophies also include reincarnation, e.g. Theosophy, Anthroposophy, Kabbalah, and Gnostic and Esoteric Christianity such as the works of Martinus Thomsen. Demographic survey data from 1999 to 2002 shows a significant minority of people from Europe (22%) and America (20%) believe in the existence of life before birth and after death, leading to a physical rebirth. The belief in reincarnation is particularly high in the Baltic countries, with Lithuania having the highest figure for the whole of Europe, 44%, while the lowest figure is in East Germany, 12%. A quarter of U.S. Christians, including 10% of all born again Christians, embrace the idea. Academic psychiatrist and believer in reincarnation, Ian Stevenson, reported that belief in reincarnation is held (with variations in details) by adherents of almost all major religions except Christianity and Islam. In addition, between 20 and 30 percent of persons in western countries who may be nominal Christians also believe in reincarnation. One 1999 study by Walter and Waterhouse reviewed the previous data on the level of reincarnation belief and performed a set of thirty in-depth interviews in Britain among people who did not belong to a religion advocating reincarnation. The authors reported that surveys have found about one fifth to one quarter of Europeans have some level of belief in reincarnation, with similar results found in the USA. In the interviewed group, the belief in the existence of this phenomenon appeared independent of their age, or the type of religion that these people belonged to, with most being Christians. The beliefs of this group also did not appear to contain any more than usual of "new age" ideas (broadly defined) and the authors interpreted their ideas on reincarnation as "one way of tackling issues of suffering", but noted that this seemed to have little effect on their private lives. Waterhouse also published a detailed discussion of beliefs expressed in the interviews. She noted that although most people "hold their belief in reincarnation quite lightly" and were unclear on the details of their ideas, personal experiences such as past-life memories and near-death experiences had influenced most believers, although only a few had direct experience of these phenomena. Waterhouse analyzed the influences of second-hand accounts of reincarnation, writing that most of the people in the survey had heard other people's accounts of past-lives from regression hypnosis and dreams and found these fascinating, feeling that there "must be something in it" if other people were having such experiences. Other influential contemporary figures that have written on reincarnation include Alice Ann Bailey, one of the first writers to use the terms New Age and age of Aquarius, Torkom Saraydarian, an Armenian-American musician and religious author, Dolores Cannon, Atul Gawande, Michael Newton, Bruce Greyson, Raymond Moody and Unity Church founder Charles Fillmore. Neale Donald Walsch, an American author of the series Conversations with God claims that he has reincarnated more than 600 times. The Indian spiritual teacher Meher Baba who had significant following in the West taught that reincarnation followed from human desire and ceased once a person was freed from desire. According to various Buddhist scriptures, Gautama Buddha believed in the existence of an afterlife in another world and in reincarnation, Since there actually is another world (any world other than the present human one, i.e. different rebirth realms), one who holds the view 'there is no other world' has wrong view... The Buddha also asserted that karma influences rebirth, and that the cycles of repeated births and deaths are endless. Before the birth of Buddha, ancient Indian scholars had developed competing theories of afterlife, including the materialistic school such as Charvaka, which posited that death is the end, there is no afterlife, no soul, no rebirth, no karma, and they described death to be a state where a living being is completely annihilated, dissolved. Buddha rejected this theory, adopted the alternate existing theories on rebirth, criticizing the materialistic schools that denied rebirth and karma, states Damien Keown. Such beliefs are inappropriate and dangerous, stated Buddha, because such annihilationism views encourage moral irresponsibility and material hedonism; he tied moral responsibility to rebirth. The Buddha introduced the concept that there is no permanent self (soul), and this central concept in Buddhism is called anattā. Major contemporary Buddhist traditions such as Theravada, Mahayana and Vajrayana traditions accept the teachings of Buddha. These teachings assert there is rebirth, there is no permanent self and no irreducible ātman (soul) moving from life to another and tying these lives together, there is impermanence, that all compounded things such as living beings are aggregates dissolve at death, but every being reincarnates. The rebirth cycles continue endlessly, states Buddhism, and it is a source of duhkha (suffering, pain), but this reincarnation and duhkha cycle can be stopped through nirvana. The anattā doctrine of Buddhism is a contrast to Hinduism, the latter asserting that "soul exists, it is involved in rebirth, and it is through this soul that everything is connected". Different traditions within Buddhism have offered different theories on what reincarnates and how reincarnation happens. One theory suggests that it occurs through consciousness (Sanskrit: vijñāna; Pali: samvattanika-viññana) or stream of consciousness (Sanskrit: citta-santāna, vijñāna-srotām, or vijñāna-santāna; Pali: viññana-sotam) upon death, which reincarnates into a new aggregation. This process, states this theory, is similar to the flame of a dying candle lighting up another. The consciousness in the newly born being is neither identical to nor entirely different from that in the deceased but the two form a causal continuum or stream in this Buddhist theory. Transmigration is influenced by a being's past karma (Pali: kamma). The root cause of rebirth, states Buddhism, is the abiding of consciousness in ignorance (Sanskrit: avidya; Pali: avijja) about the nature of reality, and when this ignorance is uprooted, rebirth ceases. Buddhist traditions also vary in their mechanistic details on rebirth. Most Theravada Buddhists assert that rebirth is immediate while the Tibetan and most Chinese and Japanese schools hold to the notion of a bardo (intermediate state) that can last up to 49 days. The bardo rebirth concept of Tibetan Buddhism, originally developed in India but spread to Tibet and other Buddhist countries, and involves 42 peaceful deities, and 58 wrathful deities. These ideas led to maps on karma and what form of rebirth one takes after death, discussed in texts such as The Tibetan Book of the Dead. The major Buddhist traditions accept that the reincarnation of a being depends on the past karma and merit (demerit) accumulated, and that there are six realms of existence in which the rebirth may occur after each death. Within Japanese Zen, reincarnation is accepted by some, but rejected by others. A distinction can be drawn between 'folk Zen', as in the Zen practiced by devotional lay people, and 'philosophical Zen'. Folk Zen generally accepts the various supernatural elements of Buddhism such as rebirth. Philosophical Zen, however, places more emphasis on the present moment. Some schools conclude that karma continues to exist and adhere to the person until it works out its consequences. For the Sautrantika school, each act "perfumes" the individual or "plants a seed" that later germinates. Tibetan Buddhism stresses the state of mind at the time of death. To die with a peaceful mind will stimulate a virtuous seed and a fortunate rebirth; a disturbed mind will stimulate a non-virtuous seed and an unfortunate rebirth. In the major Christian denominations, the concept of reincarnation is not present and it is nowhere explicitly referred to in the Bible. However, the impossibility of a second earthly death is stated by 1 Peter 3:18–20, where it affirms that the messiah, Jesus of Nazareth, died once forever for the sins of all the human kind. Matthew 14:1–2 mentions that king Herod Antipas took Jesus to be a risen John the Baptist, when introducing the story of John's execution at Herod's orders. In a survey by the Pew Forum in 2009, 22% of American Christians expressed a belief in reincarnation, and in a 1981 survey 31% of regular churchgoing European Catholics expressed a belief in reincarnation. Some Christian theologians interpret certain Biblical passages as referring to reincarnation. These passages include the questioning of Jesus as to whether he is Elijah, John the Baptist, Jeremiah, or another prophet (Matthew 16:13–15 and John 1:21–22) and, less clearly (while Elijah was said not to have died, but to have been taken up to heaven), John the Baptist being asked if he is not Elijah (John 1:25). Geddes MacGregor, an Episcopalian priest and professor of philosophy, has made a case for the compatibility of Christian doctrine and reincarnation. There is evidence that Origen, a Church father in early Christian times, taught reincarnation in his lifetime but that when his works were translated into Latin these references were concealed. One of the epistles written by St. Jerome, "To Avitus" (Letter 124; Ad Avitum. Epistula CXXIV), which asserts that Origen's On the First Principles (Latin: De Principiis; Greek: Περὶ Ἀρχῶν) was mistranscribed: About ten years ago that saintly man Pammachius sent me a copy of a certain person's [ Rufinus's ] rendering, or rather misrendering, of Origen's First Principles; with a request that in a Latin version I should give the true sense of the Greek and should set down the writer's words for good or for evil without bias in either direction. When I did as he wished and sent him the book, he was shocked to read it and locked it up in his desk lest being circulated it might wound the souls of many. Under the impression that Origen was a heretic like Arius, St. Jerome criticizes ideas described in On the First Principles. Further in "To Avitus" (Letter 124), St. Jerome writes about "convincing proof" that Origen teaches reincarnation in the original version of the book: The following passage is a convincing proof that he holds the transmigration of the souls and annihilation of bodies. 'If it can be shown that an incorporeal and reasonable being has life in itself independently of the body and that it is worse off in the body than out of it; then beyond a doubt bodies are only of secondary importance and arise from time to time to meet the varying conditions of reasonable creatures. Those who require bodies are clothed with them, and contrariwise, when fallen souls have lifted themselves up to better things, their bodies are once more annihilated. They are thus ever vanishing and ever reappearing.' The original text of On First Principles has almost completely disappeared. It remains extant as De Principiis in fragments faithfully translated into Latin by St. Jerome and in "the not very reliable Latin translation of Rufinus." Reincarnation was also taught by several gnostics such as Marcion of Sinope. Belief in reincarnation was rejected by Augustine of Hippo in The City of God. Reincarnation is a paramount tenet in the Druze faith. There is an eternal duality of the body and the soul and it is impossible for the soul to exist without the body. Therefore, reincarnations occur instantly at one's death. While in the Hindu and Buddhist belief system a soul can be transmitted to any living creature, in the Druze belief system this is not possible and a human soul will only transfer to a human body. Furthermore, souls cannot be divided into different or separate parts and the number of souls existing is finite. Few Druzes are able to recall their past but, if they are able to they are called a Nateq. Typically souls who have died violent deaths in their previous incarnation will be able to recall memories. Since death is seen as a quick transient state, mourning is discouraged. Unlike other Abrahamic faiths, heaven and hell are spiritual. Heaven is the ultimate happiness received when soul escapes the cycle of rebirths and reunites with the Creator, while hell is conceptualized as the bitterness of being unable to reunite with the Creator and escape from the cycle of rebirth. The body dies, assert the Hindu traditions, but not the soul, which they assume to be the eternal reality, indestructible and bliss. Everything and all existence is believed to be connected and cyclical in many Hinduism-sects, all living beings composed of two things, the soul and the body or matter. Ātman does not change and cannot change by its innate nature in the Hindu belief. Current Karma impacts the future circumstances in this life, as well as the future forms and realms of lives. Good intent and actions lead to good future, bad intent and actions lead to bad future, impacting how one reincarnates, in the Hindu view of existence. There is no permanent heaven or hell in most Hinduism-sects. In the afterlife, based on one's karma, the soul is reborn as another being in heaven, hell, or a living being on earth (human, animal). Gods, too, die once their past karmic merit runs out, as do those in hell, and they return getting another chance on earth. This reincarnation continues, endlessly in cycles, until one embarks on a spiritual pursuit, realizes self-knowledge, and thereby gains mokṣa, the final release out of the reincarnation cycles. This release is believed to be a state of utter bliss, which Hindu traditions believe is either related or identical to Brahman, the unchanging reality that existed before the creation of universe, continues to exist, and shall exist after the universe ends. The Upanishads, part of the scriptures of the Hindu traditions, primarily focus on the liberation from reincarnation. The Bhagavad Gita discusses various paths to liberation. The Upanishads, states Harold Coward, offer a "very optimistic view regarding the perfectibility of human nature", and the goal of human effort in these texts is a continuous journey to self-perfection and self-knowledge so as to end Saṃsāra—the endless cycle of rebirth and redeath. The aim of spiritual quest in the Upanishadic traditions is find the true self within and to know one's soul, a state that they assert leads to blissful state of freedom, moksha. The Bhagavad Gita states: Just as in the body childhood, adulthood and old age happen to an embodied being. So also he (the embodied being) acquires another body. The wise one is not deluded about this. (2:13) As, after casting away worn out garments, a man later takes new ones. So after casting away worn out bodies, the embodied Self encounters other new ones. (2:22) When an embodied being transcends, these three qualities which are the source of the body, Released from birth, death, old age and pain, he attains immortality. (14:20) There are internal differences within Hindu traditions on reincarnation and the state of moksha. For example, the dualistic devotional traditions such as Madhvacharya's Dvaita Vedanta tradition of Hinduism champion a theistic premise, assert that human soul and Brahman are different, loving devotion to Brahman (god Vishnu in Madhvacharya's theology) is the means to release from Samsara, it is the grace of God which leads to moksha, and spiritual liberation is achievable only in after-life (videhamukti). The non-dualistic traditions such as Adi Shankara's Advaita Vedanta tradition of Hinduism champion a monistic premise, asserting that the individual human soul and Brahman are identical, only ignorance, impulsiveness and inertia leads to suffering through Saṃsāra, in reality there are no dualities, meditation and self-knowledge is the path to liberation, the realization that one's soul is identical to Brahman is moksha, and spiritual liberation is achievable in this life (jivanmukti). Most Islamic schools of thought reject any idea of reincarnation of living beings. It teaches a linear concept of life, wherein a human being has only one life and upon death he or she is judged by God, then rewarded in heaven or punished in hell. Islam teaches final resurrection and Judgement Day, but there is no prospect for the reincarnation of a human being into a different body or being. During the early history of Islam, some of the Caliphs persecuted all reincarnation-believing people, such as Manichaeism, to the point of extinction in Mesopotamia and Persia (modern day Iraq and Iran). However, some Muslim minority sects such as those found among Sufis, and some Muslims in South Asia and Indonesia have retained their pre-Islamic Hindu and Buddhist beliefs in reincarnation. For instance, historically, South Asian Isma'ilis performed chantas yearly, one of which is for seeking forgiveness of sins committed in past lives. The idea of reincarnation is accepted by a few heterodox sects, particularly of the Ghulat. Alawites hold that they were originally stars or divine lights that were cast out of heaven through disobedience and must undergo repeated reincarnation (or metempsychosis) before returning to heaven. They can be reincarnated as Christians or others through sin and as animals if they become infidels. In Jainism, the reincarnation doctrine, along with its theories of Saṃsāra and Karma, are central to its theological foundations, as evidenced by the extensive literature on it in the major sects of Jainism, and their pioneering ideas on these topics from the earliest times of the Jaina tradition. Reincarnation in contemporary Jainism traditions is the belief that the worldly life is characterized by continuous rebirths and suffering in various realms of existence. Karma forms a central and fundamental part of Jain faith, being intricately connected to other of its philosophical concepts like transmigration, reincarnation, liberation, non-violence (ahiṃsā) and non-attachment, among others. Actions are seen to have consequences: some immediate, some delayed, even into future incarnations. So the doctrine of karma is not considered simply in relation to one life-time, but also in relation to both future incarnations and past lives. Uttarādhyayana Sūtra 3.3–4 states: "The jīva or the soul is sometimes born in the world of gods, sometimes in hell. Sometimes it acquires the body of a demon; all this happens on account of its karma. This jīva sometimes takes birth as a worm, as an insect or as an ant." The text further states (32.7): "Karma is the root of birth and death. The souls bound by karma go round and round in the cycle of existence." Actions and emotions in the current lifetime affect future incarnations depending on the nature of the particular karma. For example, a good and virtuous life indicates a latent desire to experience good and virtuous themes of life. Therefore, such a person attracts karma that ensures that their future births will allow them to experience and manifest their virtues and good feelings unhindered. In this case, they may take birth in heaven or in a prosperous and virtuous human family. On the other hand, a person who has indulged in immoral deeds, or with a cruel disposition, indicates a latent desire to experience cruel themes of life. As a natural consequence, they will attract karma which will ensure that they are reincarnated in hell, or in lower life forms, to enable their soul to experience the cruel themes of life. There is no retribution, judgment or reward involved but a natural consequences of the choices in life made either knowingly or unknowingly. Hence, whatever suffering or pleasure that a soul may be experiencing in its present life is on account of choices that it has made in the past. As a result of this doctrine, Jainism attributes supreme importance to pure thinking and moral behavior. The Jain texts postulate four gatis, that is states-of-existence or birth-categories, within which the soul transmigrates. The four gatis are: deva (demigods), manuṣya (humans), nāraki (hell beings), and tiryañca (animals, plants, and microorganisms). The four gatis have four corresponding realms or habitation levels in the vertically tiered Jain universe: deva occupy the higher levels where the heavens are situated; manuṣya and tiryañca occupy the middle levels; and nāraki occupy the lower levels where seven hells are situated. Single-sensed souls, however, called nigoda, and element-bodied souls pervade all tiers of this universe. Nigodas are souls at the bottom end of the existential hierarchy. They are so tiny and undifferentiated, that they lack even individual bodies, living in colonies. According to Jain texts, this infinity of nigodas can also be found in plant tissues, root vegetables and animal bodies. Depending on its karma, a soul transmigrates and reincarnates within the scope of this cosmology of destinies. The four main destinies are further divided into sub-categories and still smaller sub-sub-categories. In all, Jain texts speak of a cycle of 8.4 million birth destinies in which souls find themselves again and again as they cycle within samsara. In Jainism, God has no role to play in an individual's destiny; one's personal destiny is not seen as a consequence of any system of reward or punishment, but rather as a result of its own personal karma. A text from a volume of the ancient Jain canon, Bhagvati sūtra 8.9.9, links specific states of existence to specific karmas. Violent deeds, killing of creatures having five sense organs, eating fish, and so on, lead to rebirth in hell. Deception, fraud and falsehood lead to rebirth in the animal and vegetable world. Kindness, compassion and humble character result in human birth; while austerities and the making and keeping of vows lead to rebirth in heaven. Each soul is thus responsible for its own predicament, as well as its own salvation. Accumulated karma represent a sum total of all unfulfilled desires, attachments and aspirations of a soul. It enables the soul to experience the various themes of the lives that it desires to experience. Hence a soul may transmigrate from one life form to another for countless of years, taking with it the karma that it has earned, until it finds conditions that bring about the required fruits. In certain philosophies, heavens and hells are often viewed as places for eternal salvation or eternal damnation for good and bad deeds. But according to Jainism, such places, including the earth are simply the places which allow the soul to experience its unfulfilled karma. Jewish mystical texts (the Kabbalah), from their classic Medieval canon onward, teach a belief in Gilgul Neshamot (Hebrew for metempsychosis; literally 'soul cycle'; plural gilgulim). The Zohar and the Sefer HaBahir specifically discuss reincarnation. It is a common belief in contemporary Hasidic Judaism, which regards the Kabbalah as sacred and authoritative, though understood in light of a more innate psychological mysticism. Kabbalah also teaches that "The soul of Moses is reincarnated in every generation." Other, Non-Hasidic, Orthodox Jewish groups while not placing a heavy emphasis on reincarnation, do acknowledge it as a valid teaching. Its popularization entered modern secular Yiddish literature and folk motif. The 16th century mystical renaissance in communal Safed replaced scholastic Rationalism as mainstream traditional Jewish theology, both in scholarly circles and in the popular imagination. References to gilgul in former Kabbalah became systematized as part of the metaphysical purpose of creation. Isaac Luria (the Ari) brought the issue to the centre of his new mystical articulation, for the first time, and advocated identification of the reincarnations of historic Jewish figures that were compiled by Haim Vital in his Shaar HaGilgulim. Gilgul is contrasted with the other processes in Kabbalah of Ibbur ('pregnancy'), the attachment of a second soul to an individual for (or by) good means, and Dybuk ('possession'), the attachment of a spirit, demon, etc. to an individual for (or by) "bad" means. In Lurianic Kabbalah, reincarnation is not retributive or fatalistic, but an expression of Divine compassion, the microcosm of the doctrine of cosmic rectification of creation. Gilgul is a heavenly agreement with the individual soul, conditional upon circumstances. Luria's radical system focused on rectification of the Divine soul, played out through Creation. The true essence of anything is the divine spark within that gives it existence. Even a stone or leaf possesses such a soul that "came into this world to receive a rectification". A human soul may occasionally be exiled into lower inanimate, vegetative or animal creations. The most basic component of the soul, the nefesh, must leave at the cessation of blood production. There are four other soul components and different nations of the world possess different forms of souls with different purposes. Each Jewish soul is reincarnated in order to fulfill each of the 613 Mosaic commandments that elevate a particular spark of holiness associated with each commandment. Once all the Sparks are redeemed to their spiritual source, the Messianic Era begins. Non-Jewish observance of the 7 Laws of Noah assists the Jewish people, though Biblical adversaries of Israel reincarnate to oppose. Among the many rabbis who accepted reincarnation are Nahmanides (the Ramban) and Rabbenu Bahya ben Asher, Levi ibn Habib (the Ralbah), Shelomoh Alkabez, Moses Cordovero, Moses Chaim Luzzatto; early Hasidic masters such as the Baal Shem Tov, Schneur Zalman of Liadi and Nachman of Breslov, as well as virtually all later Hasidic masters; contemporary Hasidic teachers such as DovBer Pinson, Moshe Weinberger and Joel Landau; and key Mitnagdic leaders, such as the Vilna Gaon and Chaim Volozhin and their school, as well as Rabbi Shalom Sharabi (known at the RaShaSH), the Ben Ish Chai of Baghdad, and the Baba Sali. Rabbis who have rejected the idea include Saadia Gaon, David Kimhi, Hasdai Crescas, Joseph Albo, Abraham ibn Daud, Leon de Modena, Solomon ben Aderet, Maimonides and Asher ben Jehiel. Among the Geonim, Hai Gaon argued in favour of gilgulim. In the Western Hemisphere, belief in reincarnation is most prevalent in the now heavily Christian Polar North (now mainly parts of Greenland and Nunavut). The concept of reincarnation is enshrined in the Inuit languages, and in many Inuit cultures it is traditional to name a newborn child after a recently deceased person under the belief that the child is the namesake reincarnated. Reincarnation is an intrinsic part of some Northeastern Native American traditions. The following is a story of human-to-human reincarnation as told by Thunder Cloud, a Winnebago (Ho-Chunk) shaman. Here Thunder Cloud talks about his two previous lives and how he died and came back again to this his third lifetime. He describes his time between lives, when he was "blessed" by Earth Maker and all the abiding spirits and given special powers, including the ability to heal the sick. Thunder Cloud's account of his two reincarnations: I (my ghost) was taken to the place where the sun sets (the west). ... While at that place, I thought I would come back to earth again, and the old man with whom I was staying said to me, "My son, did you not speak about wanting to go to the earth again?" I had, as a matter of fact, only thought of it, yet he knew what I wanted. Then he said to me, "You can go, but you must ask the chief first." Then I went and told the chief of the village of my desire, and he said to me, "You may go and obtain your revenge upon the people who killed your relatives and you." Then I was brought down to earth. ... There I lived until I died of old age. ... As I was lying [in my grave], someone said to me, "Come, let us go away." So then we went toward the setting of the sun. There we came to a village where we met all the dead. ... From that place I came to this earth again for the third time, and here I am. Founded in the 15th century, Sikhism's founder Guru Nanak had a choice between the cyclical reincarnation concept of ancient Indian religions and the linear concept of Islam, he chose the cyclical concept of time. Sikhism teaches reincarnation theory similar to those in Hinduism, but with some differences from its traditional doctrines. Sikh rebirth theories about the nature of existence are similar to ideas that developed during the devotional Bhakti movement particularly within some Vaishnava traditions, which define liberation as a state of union with God attained through the grace of God. The doctrines of Sikhism teach that the soul exists, and is passed from one body to another in endless cycles of Saṃsāra, until liberation from the death and rebirth cycle. Each birth begins with karma (karam), and these actions leave a karmic signature (karni) on one's soul which influences future rebirths, but it is God whose grace that liberates from the death and rebirth cycle. The way out of the reincarnation cycle, asserts Sikhism, is to live an ethical life, devote oneself to God and constantly remember God's name. The precepts of Sikhism encourage the bhakti of One Lord for mukti (liberation from the death and rebirth cycle). The Yoruba religion teaches that Olodumare, the Supreme Being and divine Creator who rules over His Creation, created eniyan, or humanity, to achieve balance between heaven and earth and bring about Ipo Rere, or the Good Condition. To cause achievement of the Good Condition, humanity reincarnates. Once achieved, Ipo Rere provides the ultimate state of supreme existence with Olodumare, a goal which elevates reincarnation to a key position in the Yoruba religion. Atunwaye (also called atunwa) is the Yoruba term for reincarnation. Predestination is a foundational component of atunwaye. Just prior to incarnation, a person first chooses their Ayanmo (destiny) before also choosing their Akunyelan (lot) in the presence of Olodumare and Orunmila with Olodumare's approval. By atunwaye, a person may incarnate only in a human being and may choose to reincarnate in either sex, regardless of choice in the prior incarnation. The most common, widespread Yoruba reincarnation belief is ipadawaye, meaning "the ancestor's rebirth". According to this belief, the reincarnating person will reincarnate along their familial lineage. When a person dies, they go to orun (heaven) and will live with the ancestors in either orunrere (good heaven) or orunapaadi (bad heaven). Reincarnation is believed to be a gift bestowed on ancestors who lived well and experienced a "good" death. Only ancestors living in orunrere may return as grandchildren, reincarnating out of their love for the family or the world. Children may be given names to indicate which ancestor is believed to have returned, such as Babatide ("father has come"), Babatunde ("father has come again"), and Yetunde ("mother has come again"). A "bad" death (which includes deaths of children, cruel, or childless people and deaths by punishments from the gods, accidents, suicides, and gruesome murders) is generally believed to prevent the deceased from joining the ancestors and reincarnating again, though some practitioners also believe a person experiencing a "bad" death will be reborn much later into conditions of poverty. Another Yoruba reincarnation belief is abiku, meaning "born to die" According to Yoruba custom, an abiku is a reincarnating child who repeatedly experiences death and rebirth with the same mother in a vicious cycle. Because childlessness is considered a curse in Yoruba culture, parents with an abiku child will always attempt to help the abiku child by preventing their death. However, abiku are believed to possess a power to ensure their eventual death, so rendering assistance is often a frustrating endeavor causing significant pain to the parents. This pain is believed to bring happiness to the abiku. Abiku are believed to be a "species of spirit" thought to live apart from people in, for example, secluded parts of villages, jungles, and footpaths. Modern belief in abiku has significantly waned among urban populations, with the decline attributed to improved hygiene and medical care reducing infant mortality rates. Akudaaya, meaning "born to die and reappear" (also called akuda), is a Yoruba reincarnation belief of "a person that is dead[] but has not gone to heaven". Akudaaya is based on the belief that, if a recently-deceased person's destiny in that life remained unfulfilled, the deceased cannot join the ancestors and therefore must roam the world. Following death, an akudaaya returns to their previous existence by reappearing in the same physical form. However, the new existence will be lived in a different physical location from the first, and the akudaaya will not be recognized by a still-living relative, should they happen to meet. The akudaaya lives their new existence working to fulfill their destiny from the previous life. The concept of akudaaya is the subject of Akudaaya (The Wraith), a 2023 Nigerian drama film in the Yoruba language. The film is said to center on a deceased son who "has begun living life as a spirit in another state and has fallen in love". Spiritism, a Christian philosophy codified in the 19th century by the French educator Allan Kardec, teaches reincarnation or rebirth into human life after death. According to this doctrine, free will and cause and effect are the corollaries of reincarnation, and reincarnation provides a mechanism for a person's spiritual evolution in successive lives. The Theosophical Society draws much of its inspiration from India. In the Theosophical world-view reincarnation is the vast rhythmic process by which the soul, the part of a person which belongs to the formless non-material and timeless worlds, unfolds its spiritual powers in the world and comes to know itself. It descends from sublime, free, spiritual realms and gathers experience through its effort to express itself in the world. Afterwards there is a withdrawal from the physical plane to successively higher levels of reality, in death, a purification and assimilation of the past life. Having cast off all instruments of personal experience it stands again in its spiritual and formless nature, ready to begin its next rhythmic manifestation, every lifetime bringing it closer to complete self-knowledge and self-expression. However, it may attract old mental, emotional, and energetic karma patterns to form the new personality. Anthroposophy describes reincarnation from the point of view of Western philosophy and culture. The ego is believed to transmute transient soul experiences into universals that form the basis for an individuality that can endure after death. These universals include ideas, which are intersubjective and thus transcend the purely personal (spiritual consciousness), intentionally formed human character (spiritual life), and becoming a fully conscious human being (spiritual humanity). Rudolf Steiner described both the general principles he believed to be operative in reincarnation, such as that one's will activity in one life forms the basis for the thinking of the next, and a number of successive lives of various individualities. Similarly, other famous people's life stories are not primarily the result of genes, upbringing or biographical vicissitudes. Steiner relates that a large estate in north-eastern France was held during the early Middle Ages by a martial feudal lord. During a military campaign, this estate was captured by a rival. The previous owner had no means of retaliating, and was forced to see his property lost to an enemy. He was filled with a smoldering resentment towards the propertied classes, not only for the remainder of his life in the Middle Ages, but also in a much later incarnation—as Karl Marx. His rival was reborn as Friedrich Engels. Inspired by Helena Blavatsky's major works, including Isis Unveiled and The Secret Doctrine, astrologers in the early twentieth-century integrated the concepts of karma and reincarnation into the practice of Western astrology. Notable astrologers who advanced this development included Alan Leo, Charles E. O. Carter, Marc Edmund Jones, and Dane Rudhyar. A new synthesis of East and West resulted as Hindu and Buddhist concepts of reincarnation were fused with Western astrology's deep roots in Hermeticism and Neoplatonism. In the case of Rudhyar, this synthesis was enhanced with the addition of Jungian depth psychology. This dynamic integration of astrology, reincarnation and depth psychology has continued into the modern era with the work of astrologers Steven Forrest and Jeffrey Wolf Green. Their respective schools of Evolutionary Astrology are based on "an acceptance of the fact that human beings incarnate in a succession of lifetimes". Past reincarnation, usually termed past lives, is a key part of the principles and practices of the Church of Scientology. Scientologists believe that the human individual is actually a thetan, an immortal spiritual entity, that has fallen into a degraded state as a result of past-life experiences. Scientology auditing is intended to free the person of these past-life traumas and recover past-life memory, leading to a higher state of spiritual awareness. This idea is echoed in their highest fraternal religious order, Sea Org, whose motto is "Revenimus" ('We Come Back'), and whose members sign a "billion-year contract" as a sign of commitment to that ideal. L. Ron Hubbard, the founder of Scientology, does not use the word "reincarnation" to describe its beliefs, noting that: "The common definition of reincarnation has been altered from its original meaning. The word has come to mean 'to be born again in different life forms' whereas its actual definition is 'to be born again into the flesh of another body.' Scientology ascribes to this latter, original definition of reincarnation." The first writings in Scientology regarding past lives date from around 1951 and slightly earlier. In 1960, Hubbard published a book on past lives entitled Have You Lived Before This Life. In 1968 he wrote Mission Into Time, a report on a five-week sailing expedition to Sardinia, Sicily and Carthage to see if specific evidence could be found to substantiate L. Ron Hubbard's recall of incidents in his own past, centuries ago. Wicca is a neo-pagan religion focused on nature, guided by the philosophy of Wiccan Rede that advocates the tenets "Harm None, Do As Ye Will". Wiccans believe in a form of karmic return where one's deeds are returned, either in the current life or in another life, threefold or multiple times in order to teach one lessons (the Threefold Law). Reincarnation is therefore an accepted part of the Wiccan faith. Wiccans also believe that death and afterlife are important experiences for the soul to transform and prepare for future lifetimes. While there has been no scientific confirmation of the physical reality of reincarnation, where the subject has been discussed, there are questions of whether and how such beliefs may be justified within the discourse of science and religion. Some champions of academic parapsychology have argued that they have scientific evidence even while their detractors have accused them of practicing a form of pseudoscience. Skeptic Carl Sagan asked the Dalai Lama what he would do if a fundamental tenet of his religion (reincarnation) were definitively disproved by science. The Dalai Lama answered, "If science can disprove reincarnation, Tibetan Buddhism would abandon reincarnation...but it's going to be mighty hard to disprove reincarnation." Sagan considered claims of memories of past lives to be worthy of research, although he considered reincarnation to be an unlikely explanation for these. Over a period of 40 years, psychiatrist Ian Stevenson, from the University of Virginia, recorded case studies of young children who claimed to remember past lives. He published twelve books, including Twenty Cases Suggestive of Reincarnation, Reincarnation and Biology: A Contribution to the Etiology of Birthmarks and Birth Defects (a two-part monograph), European Cases of the Reincarnation Type, and Where Reincarnation and Biology Intersect. In his cases he reported the child's statements and testimony from family members and others, often along with what he considered to be correlates to a deceased person who in some ways seemed to match the child's memory. Stevenson also investigated cases where he thought that birthmarks and birth defects seemed to match wounds and scars on the deceased. Sometimes included in his documentation were medical records like autopsy photographs. As any claim of past life memory is subject to charges of false memories and the ease with which such claims can be hoaxed, Stevenson expected the controversy and skepticism of his beliefs that followed. He said that he looked for disconfirming evidence and alternative explanations for reports, but, as the Washington Post reported, he typically concluded that no normal explanation sufficed. Other academic researchers who have undertaken similar pursuits include Jim B. Tucker, Antonia Mills, Satwant Pasricha, Godwin Samararatne, and Erlendur Haraldsson, but Stevenson's publications remain the most well known. Stevenson's work in this regard was impressive enough to Carl Sagan that he referred to what were apparently Stevenson's investigations in his book The Demon-Haunted World as an example of carefully collected empirical data, and though he rejected reincarnation as a parsimonious explanation for the stories, he wrote that the phenomenon of alleged past-life memories should be further researched. Sam Harris cited Stevenson's works in his book The End of Faith as part of a body of data that seems to attest to the reality of psychic phenomena, but that only relies on subjective personal experience. Stevenson's claims have been subject to criticism and debunking, for example by the philosopher Paul Edwards, who contended that Ian Stevenson's accounts of reincarnation were purely anecdotal and cherry-picked. Edwards attributed the stories to selective thinking, suggestion, and false memories that result from the family's or researcher's belief systems and thus did not rise to the standard of fairly sampled empirical evidence. The philosopher Keith Augustine wrote in critique that the fact that "the vast majority of Stevenson's cases come from countries where a religious belief in reincarnation is strong, and rarely elsewhere, seems to indicate that cultural conditioning (rather than reincarnation) generates claims of spontaneous past-life memories." Further, Ian Wilson pointed out that a large number of Stevenson's cases consisted of poor children remembering wealthy lives or belonging to a higher caste. In these societies, claims of reincarnation have been used as schemes to obtain money from the richer families of alleged former incarnations. Robert Baker asserted that all the past-life experiences investigated by Stevenson and other parapsychologists are understandable in terms of known psychological factors including a mixture of cryptomnesia and confabulation. Edwards also objected that reincarnation invokes assumptions that are inconsistent with modern science. As the vast majority of people do not remember previous lives and there is no empirically documented mechanism known that allows personality to survive death and travel to another body, positing the existence of reincarnation is subject to the principle that "extraordinary claims require extraordinary evidence". Stevenson also claimed there were a handful of cases that suggested evidence of xenoglossy, including two where a subject under hypnosis allegedly conversed with people speaking the foreign language, instead of merely being able to recite foreign words. Sarah Thomason, a linguist (and skeptical researcher) at the University of Michigan, reanalyzed these cases, concluding that "the linguistic evidence is too weak to provide support for the claims of xenoglossy". Some believers in reincarnation (Stevenson famously not among them) give much importance to supposed past-life memories retrieved under hypnosis during past life regressions. Popularized by psychiatrist Brian Weiss, who claims he has regressed more than 4,000 patients since 1980, the technique is often identified as a kind of pseudoscientific practice. Such supposed memories have been documented to contain historical inaccuracies originating from modern popular culture, common beliefs about history, or books that discuss historical events. Experiments with subjects undergoing past life regression indicate that a belief in reincarnation and suggestions by the hypnotist are the two most important factors regarding the contents of memories reported. The use of hypnosis and suggestive questions can tend to leave the subject particularly likely to hold distorted or false memories. Rather than recall of a previous existence, the source of the memories is more likely cryptomnesia and confabulations that combine experiences, knowledge, imagination and suggestion or guidance from the hypnotist. Once created, those memories are indistinguishable from memories based on events that occurred during the subject's life. Past-life regression has been critiqued for being unethical on the grounds that it lacks any evidence to support its claims and that it increases one's susceptibility to false memories. Luis Cordón states that this can be problematic as it creates delusions under the guise of therapy. The memories are experienced as being as vivid as those based on events experienced in one's life and impossible to differentiate from true memories of actual events, and accordingly any damage can be difficult to undo. APA accredited organizations have challenged the use of past-life regressions as a therapeutic method, calling it unethical. Additionally, the hypnotic methodology that underpins past-life regression has been criticized as placing the participant in a vulnerable position, susceptible to implantation of false memories. Because the implantation of false memories may be harmful, Gabriel Andrade argues that past-life regression violates the principle of first, do no harm (non-maleficence), part of the Hippocratic Oath.
#REDIRECT [[Philosophy]]
null
{{DISPLAYTITLE:eBay}} {{More citations needed|date=December 2024}} {{Infobox company | name = eBay Inc. | logo = EBay logo.svg | logo_caption = Logo since 2012 | image = EBay Silicon Valley Headquarters in Willow Glen, San Jose, California 1938.jpg | image_size = 250px | image_caption = Headquarters in the [[Willow Glen, San Jose|Willow Glen]] district<br />of San Jose, California, in 2021. | former_name = AuctionWeb (1995–1997) | type = [[Public company|Public]] | traded_as = {{ubl|{{NASDAQ|EBAY}}|[[Nasdaq-100]] component|[[S&P 500]] component}} | industry = [[E-commerce]] | founded = {{start date and age|1995|9|3}} | founder = [[Pierre Omidyar]] | location = [[San Jose, California]], U.S. | area_served = Worldwide | key_people = {{unbulleted list| [[Thomas J. Tierney]]<br />([[Chairman#Public corporations|chairman]])|[[Jamie Iannone]]<br />([[President (corporate title)|president]]{{wbr}}&nbsp;& [[Chief executive officer|CEO]])}} | services = [[Online shopping]] | revenue = {{decrease}} {{US$|9.79 billion|link=yes}} (2022) | operating_income = {{decrease}} {{US$|2.35 billion}} (2022) | net_income = {{decrease}} {{US$|-1.27 billion}} (2022) | assets = {{decrease}} {{US$|20.85 billion}} (2022) | equity = {{decrease}} {{US$|5.15 billion}} (2022) | num_employees = {{circa|11,600}} (December 2022) | subsid = [[Auction Co.]], [[iBazar]], [[GittiGidiyor]], [[G-Market]], [[Half.com]], Qoo10.jp | footnotes = <ref name="10-K">{{cite web|url=https://www.sec.gov/ix?doc=/Archives/edgar/data/1065088/000106508823000006/ebay-20221231.htm|title=eBay, Inc. 2022 Annual Report (Form 10-K)|date=23 February 2023|publisher=[[U.S. Securities and Exchange Commission]]|access-date=June 11, 2023|archive-date=May 29, 2023|archive-url=https://web.archive.org/web/20230529224205/https://www.sec.gov/ix?doc=/Archives/edgar/data/1065088/000106508823000006/ebay-20221231.htm|url-status=live}}</ref> | homepage = {{official URL}} }} '''eBay''' is a [[website]]. It is a place for [[online]] [[auction]]s. Anybody can [[sales|sell]] almost anything they want there. People can [[buy]] many things there as well. People have to send what they sell themselves, usually using [[mail]]. Some items may be collected by the buyer such as items to large or heavy to post. Buyers can [[:wikt:transfer|transfer]] the [[money]] in any way they want as long as the sellers accept payments made that way. A lot of people use [[PayPal]] to transfer the money. PayPal was [[:wikt:own|owned]] by the same people as eBay until 2015. [[Margaret Whitman]] was the CEO of eBay from 1995 to 2008. == History == eBay was started in [[1995]] by [[Pierre Omidyar]]. It was first called ''Auctionweb''. The name 'eBay' came later when he tried to register a [[domain name]] for his planned name ''EchoBay''. It was already taken so he shortened the name to 'eBay'. == Statistics == Millions of items are bought on eBay every day. It is often very cheap to list an item on eBay. Total costs are usually below ordinary auction house costs. Many people look at others' items because they are interested in them. == References == <references /> == Other websites == * [http://www.ebay.com eBay site] {{San Jose, California}} {{authority control}} [[Category:American websites]] [[Category:E-commerce]]
eBay Inc. (/ˈiːbeɪ/ EE-bay, often stylized as ebay) is an American multinational e-commerce company based in San Jose, California, that brokers customer to customer and retail sales through online marketplaces in 190 markets worldwide. Sales occur either via online auctions or "buy it now" instant sales, and the company charges commissions to sellers upon sales. eBay was founded by Pierre Omidyar in September 1995. It has 134 million yearly active buyers worldwide and handled $74 billion in transactions in 2022, 49% of which was in the United States. In 2022, the company had a take rate (revenue as a percentage of volume) of 13.25%. eBay is used by individuals, companies, as well as governments to purchase and sell almost any legal, non-controversial item. eBay's auctions use a Vickrey auction (sealed-bid) proxy bid system. Buyers and sellers may rate and review each other after each transaction, resulting in a reputation system. The eBay service is accessible via websites and mobile apps. Software developers can create applications that integrate with eBay through the eBay API. Merchants can also earn commissions from affiliate marketing programs by eBay. eBay was founded as AuctionWeb in California on September 3, 1995, by French-born Iranian-American computer programmer Pierre Omidyar as a hobby to make some extra money. One of the first items sold on AuctionWeb was a broken laser pointer for $14.83. Astonished, Omidyar contacted the winning bidder to ask if he understood that the laser pointer was broken; the buyer explained: "I'm a collector of broken laser pointers." In February 1996, Omidyar's internet service provider informed him that he would need to upgrade to a business account due to the high web traffic of his website. The monthly price increase from $30 to $250 prompted Omidyar to start charging eBay users. The website made $1,000 in its first month, which was more than it cost to run, and $2,500 in its second month. Chris Agarpao was eBay's first employee; he processed mailed check payments. Jeffrey Skoll was hired as the first president of the company in early 1996. In November 1996, the company launched online auctions for airline seats, hotel rooms, cruise berths and other travel-related products in partnership with Electronic Travel Auctions. By that time, the company had hosted more than 200,000 auctions since its founding 14 months earlier. The company changed the name of its service from AuctionWeb to eBay in September 1997, after Echo Bay Technology Group, Omidyar's consulting firm. The echobay.com domain name was already registered by Echo Bay Mines, a gold mining company, so Omidyar shortened it to eBay.com. In 1997, the company received $6.7 million in venture capital funding from Benchmark. The frequently repeated story that eBay was founded to help Omidyar's fiancée trade Pez candy dispensers was fabricated in 1997 by public relations manager Mary Lou Song to give the media a human-interest story and to generate publicity with toy collectors. The most purchased and sold items on the website were Beanie Babies, the most difficult toys to find in retail stores, accounting for 10% of all listings in 1997. Ty, the manufacturer, had set up a website whereby people could trade used Beanie Babies. However, it was overwhelmed with unsortable listings. With a user-friendly interface, eBay became popular with collectors. Meg Whitman was appointed president and CEO in March 1998. At the time, the company had 30 employees, 500,000 users, and revenues of $4.7 million in the United States. In July 1998, eBay acquired Jump, the developer and operator of Up4Sale, an advertising-supported auction website which at the time had 27,000 separate auctions and 50,000 registered members. In September 1998, during the dot-com bubble, eBay became a public company via an initial public offering led by CFO Gary F. Bengier. Upon the initial public offering, which was priced at $18 per share and closed for trading on its first day at $53 per share, both Omidyar and Skoll became billionaires. In the risk factors section of the annual report filed with the U.S. Securities and Exchange Commission in 1998, Omidyar noted eBay's dependence on the continued strength of the Beanie Babies market. In June 2000, eBay acquired Half.com for $312 million in stock. In 2000, eBay partnered with Escrow.com to handle escrow for purchases and sales of motor vehicles, later expanded to other transaction types. By year-end, it had 22.5 million registered users and 79.4 million auctions per quarter. In January 2001, eBay acquired a majority stake in Internet Auction Co. Ltd, operator of the largest internet auction website in South Korea. In February 2002, eBay acquired iBazar, a French online auction site founded in 1998, for approximately $112 million in stock. eBay entered the Chinese market in 2002 and shut down its Chinese site in 2007 due to competition from local rival Taobao. In February 2002, eBay exited Japan due to competition from Yahoo! Japan and began operations in Taiwan with the acquisition of NeoCom Technology for $9.5 million. In June 2006, eBay turned over its operations in Taiwan to a joint venture partner. eBay acquired PayPal on October 3, 2002 for $1.4 billion. On May 28, 2003, in the case of eBay Inc. v. MercExchange, L.L.C., which had implications for the treatment of business method patents, a United States district court jury found eBay guilty of willful patent infringement and ordered the company to pay $35 million in damages after MercExchange accused eBay of infringing on three patents, one of which is used in eBay's "Buy It Now" feature. The decision was appealed to the United States Court of Appeals for the Federal Circuit (CAFC). The CAFC affirmed the judgment of willful infringement, and reversed the lower court and granted a permanent injunction. eBay appealed the permanent injunction to the Supreme Court of the United States, which on May 15, 2006 found an injunction is not required nor automatic in this or any patent case where guilt has been established. The case was sent back to the Virginia district court for consideration of the injunction and a trial on another MercExchange patent. In August 2004, eBay acquired 25% of classified advertising website Craigslist for $32 million. Former disgruntled Craigslist executive Phillip Knowlton was the seller. In December 2004, eBay acquired Rent.com for $415 million. In March 2005, eBay launched Kijiji, a classified advertising website, in international markets. It launched in the United States in July 2007. In May 2005, eBay acquired Gumtree, a classified advertising website in the United Kingdom. In October 2005, eBay Inc. acquired Skype Technologies for $2.6 billion. In February 2006, Intuit launched a web-based version of ItsDeductible, a donation tracking service, using data from eBay to help users assign a market value to the items they donate. In April 2006, eBay launched eBay Express, a site that was designed to work like a standard Internet shopping site, with fixed prices and no bidding involved. The website had 10 million items listed upon its launch. The site was shut down in October 2008. In January 2007, eBay acquired StubHub, an online marketplace for ticket resale, for $310 million. In April 2008, eBay sued Craigslist, claiming that in January 2008, Craigslist took actions that "unfairly diluted eBay's economic interest by more than 10%", making eBay lose its seat on the board of directors of Craigslist. Craigslist countersued in May 2008 alleging that eBay used its board seat to gain insider information about Craigslist that was used to compete against the company. In September 2010, Delaware Judge William B. Chandler III ruled that the actions of Craigslist were unlawful and that the actions were taken by Craigslist founders Jim Buckmaster and Craig Newmark had "breached their fiduciary duty of loyalty", and restored eBay's stake in the company to 28.4% from a diluted level of 24.85%. However, the judge dismissed eBay's objection to a staggered board provision, citing that Craigslist has the right to protect its own trade secrets. In January 2008, Meg Whitman resigned as president and CEO of eBay to enter politics, and was replaced with John Donahoe. Whitman remained on the board of directors and continued to advise Donahoe through 2008. In May 2008, eBay announced the opening of a building on the company's North Campus in San Jose, California, the first ground-up structure in the city to be built to LEED Gold standards. The building, the first the company had built in its 13-year existence, uses an array of 3,248 solar panels, spanning 60,000 square feet (5,600 m), and providing 650 kilowatts of power, 15–18% of the company's total energy requirements, reducing carbon dioxide usage by 37 million pounds over 30 years. The building also has energy-efficient lighting and water system and most waste is recycled. In April 2009, eBay agreed to acquire a controlling stake in G-Market, a South Korean online retailer, for $413 million. In May 2009, eBay launched the Selling Manager Applications program (SM Apps). The program allows approved developers to integrate their applications directly into the eBay.com interface. In November 2009, eBay sold a 70% stake in Skype to a consortium led by Silver Lake Partners and Marc Andreessen at a $2.75 billion valuation, while retaining a 30% minority ownership interest in Skype, after failing to integrate Skype into the company's online marketplace. Microsoft acquired the entire company for $8.5 billion in May 2011. In June 2011, eBay acquired GSI Commerce for $2.4 billion. In June 2013, it was renamed eBay Enterprise. In May 2012, RentPath, then known as Primedia, acquired Rent.com from eBay for approximately $415 million. In September 2012, eBay introduced a new logo using a thinner variation of the Univers typeface. It replaced the thicker Univers logo that had been used since eBay's inception in 1995. In October 2012, eBay launched an international shipping partnership with Pitney Bowes whereby a seller of an item to be shipped internationally can send the item to a Pitney Bowes facility in their home country, which then forwards it to the international buyer, taking care of all international shipping requirements. The company also launched a partnership with FedEx to offer discounted shipping options to sellers. In November 2012, eBay was charged in the High-Tech Employee Antitrust Litigation, accused by the United States Department of Justice of entering into non-solicitation agreements with other technology companies involving highly skilled employees. The litigation was settled in May 2014, with eBay required to end anti-competitive practices. On September 30, 2014, eBay announced it would spin off PayPal into a separate publicly traded company, a demand made nine months prior by activist hedge fund magnate Carl Icahn. The spinoff was completed on July 18, 2015. eBay's then chief executive, John Donahoe, stepped down from that role. In January 2015, eBay acquired Vivanuncios, a classified advertising website in Mexico. In June 2015, eBay sold its stake in Craigslist back to the company, ending the litigation. In August 2015, eBay sold a portion of its stake in Snapdeal. In September 2015, Propay and Skrill were eliminated as payment methods on the eBay website, citing low usage. Flipkart and eBay entered into a strategic partnership in July 2017 under which eBay acquired a 5.44% stake in Flipkart in exchange for the contribution of its India business unit valued at $211 million and a $514 million cash investment in Flipkart. Flipkart launched a program to allow its sellers to sell to customers globally in partnership with eBay. eBay reported a gain of $167 million on the sale of its India operations. In May 2018, eBay sold its stake in Flipkart to Walmart and relaunched its operations in India. In August 2017, eBay shut Half.com. In October 2017, eBay released image retrieval capability allowing users to find listings on the site that match an item depicted in a photo, using artificial intelligence and machine learning technologies. On January 31, 2018, eBay announced that it would replace PayPal as its primary payments provider with Netherlands-based start-up Adyen, resulting in lower costs and more control of merchants. In May 2018, eBay acquired Japanese e-commerce platform Qoo10 for $573 million. In July 2018, eBay announced support for Apple Pay as well as a partnership with Square for seller financing loans of up to $100,000. In September 2018, in response to the YouTube headquarters shooting, eBay announced plans to install a security fence around the perimeter of its San Jose headquarters to protect employees. On July 31, 2019, the company acquired a 5.59% stake in Paytm Mall. In September 2019, facing pressure from activist shareholder Elliott Investment Management, Devin Wenig resigned as CEO. Scott Schenkel, senior vice president and chief financial officer since 2015, was appointed as the interim CEO. In November 2019, eBay agreed to sell StubHub to Viagogo for $4.05 billion in cash; the sale was completed in February 2020. In April 2020, Jamie Iannone became the CEO of the company. In June 2020, Fred D. Anderson and Thomas J. Tierney resigned from the board of directors of the company; both had been directors since 2003. In July 2020, eBay sold its classified business to Adevinta for $2.5 billion in cash and 540 million shares of Adevinta. To gain regulatory approval, Gumtree was further divested. eBay sold its shares in Adevinta in 2023, when that company was acquired by private equity firms. In September 2020, Pierre Omidyar resigned from the board of directors, after resigning as chairman in 2015. In November 2021, eBay sold its South Korean business to Emart for $3 billion. In May 2022, eBay acquired a stake in Funko and became the preferred secondary marketplace for Funko. In June 2022, the company acquired KnownOrigin, a marketplace for non-fungible tokens. In August 2022, the company acquired the myFitment group of companies, specializing in online sales of automotive and powersports parts and accessories. In October 2022, the company acquired TCGplayer, a marketplace for collectible card games, for up to $295 million. In July 2023, the company acquired Certiligo, a provider of artificial intelligence-powered apparel and fashion goods digital IDs and authentication. Using MissionFish as an arbiter, eBay allows sellers to donate a portion of their auction proceeds to a charity of the seller's choice and charges discounted fees for charity auctions. High-profile charity auctions facilitated via eBay include the "Power Lunch" with investor Warren Buffett for 8 people at the Smith & Wollensky restaurant in New York City, with all of the proceeds going to the Glide Foundation. Auctions were held annually in 21 years between 2000 and 2022, with no auctions in 2020 and 2021 due to the COVID-19 pandemic. In total, auctions on eBay for lunch with Buffett raised $53.2 million for the Glide Foundation, with winning bids ranging from $2 million to as high as $19 million for the final auction in 2022. Charity auctions via eBay are planned for lunch with Marc Benioff, CEO of Salesforce. Also benefitting charity, a letter sent to Mark P. Mays, CEO of Clear Channel Communications by Senator Harry Reid and forty other Democratic senators, complaining about comments made by conservative talk show host Rush Limbaugh, sold for $2,100,100, with all of the proceeds going to the Marine Corps-Law Enforcement Foundation, benefiting the education of children of men and women who have died serving in the armed forces. The winning bid was matched by Limbaugh. In 2022, more than $163 million was raised for charities via the platform. Fraud committed by sellers includes selling counterfeit merchandise / bootleg recordings, shill bidding (undisclosed vendor bidding that is used to artificially inflate the price of a certain item by either the seller under an alternate account or another person in collusion with the seller), receiving payment and not shipping merchandise, shipping items other than those described, giving a deliberately misleading description and/or photo, knowingly and deliberately shipping faulty merchandise, denying warranty exchange after pre-agreeing to return merchandise authorization of defective on arrival merchandise, knowingly fencing (selling stolen goods), misrepresenting the cost of shipping, using bulk shipping prices to knowingly mask much higher costing, individual return shipping, and using pseudo-accounts to make high nonpaying bids on similar items that competitors are selling. eBay has been criticized for not doing enough to combat shill bidding. There are techniques such as auction sniping, which let buyers avoid shill bidders. Fraud committed by buyers includes filing a false shipping damage claim with the shipping company, friendly fraud (receiving merchandise and claiming otherwise), returning items other than received, removing parts from an item and returning it for a refund, sending a forged payment-service e-mail that states that he or she has made a payment to the seller's account as proof of payment, making a low bid then using pseudo-accounts to make high nonpaying bids in an attempt at gaining a low second chance offer price, damaging a non-refundable item to get a refund by claiming that the seller sent the item already damaged (in cases of buyer's remorse), and a package redirection scam, in which the return package is filled with garbage and sent to the wrong address. In March 2008, Professional Coin Grading Service issued an alert noting counterfeit PCGS slabs and various United States and Chinese coins originating from the People's Republic of China being sold on eBay. In 2004, Tiffany & Co. filed a lawsuit against eBay claiming that over 70% of the Tiffany silver jewelry offered for sale on eBay was fake and that eBay profited from the sales of counterfeit Tiffany items that infringed on its trademark. On July 14, 2008, a Federal District Court judge ruled that eBay does not have a legal responsibility to monitor users selling counterfeit items. In 2010, the Second Circuit affirmed this decision in Tiffany (NJ) Inc. v. eBay Inc. In June 2008, a court in Paris awarded damages of €40 million to LVMH over eBay auctions of counterfeit bags, perfumes, and other items sold by non-authorized retailers and entered a permanent injunction against eBay auctions of LVMH perfumes, whether counterfeit or not. eBay banned such items from its site. Also that month, a court in Troyes, France awarded eBay to pay luxury goods maker Hermès €20,000 due to the sale of two counterfeit bags on eBay in 2006. The court also ordered eBay to post the ruling on the home page of eBay's French website for three months. eBay allows buyers to rate any seller with positive, neutral, and negative comments. However, the option for sellers to leave anything other than positive feedback to buyers was removed in 2008. Criticism of the feedback system includes: In 2007 and 2008, during the period of eBay's ownership of PayPal, eBay required sellers to accept and buyers to pay with PayPal in many instances. This resulted in scrutiny by several regulatory agencies worldwide. The company later changed its payment requirements. In 2008, eBay reached a deal with Buy.com to list millions of items for sale by the retailer, angering sellers who faced additional competition. In January 2010, Auctionbytes.com held an open survey in which sellers could rate eBay, as well as competing auction and marketplace sites. In the survey, users were asked to rank 15 sites based on five criteria: profitability, customer service, communication, ease of use, and recommendation. eBay was ranked 13th, after other large sites such as Amazon.com and Craigslist, as well as lesser-known selling sites such as Atomic Mall, eCRATER, and Ruby Lane. In individual category rankings, eBay was rated the worst of all the 15 sites on customer service and communication, and average on ease of use. Some respondents stated they would have given eBay a rating of 10, three to five years ago. eBay was rated twelfth out of fifteen in the Recommended Selling Venue category. In 2011, eBay agreed to pay $30 million to settle a class action lawsuit alleging that it overcharged seller fees for sales of auto parts and accessories between April 2005 and August 2009. Members of the class received a refund of 6.67% of the fees paid in this category. eBay has been criticized for arranging its affairs so as to pay a low level of taxes in the United Kingdom. The Sunday Times reported in October 2012 that eBay paid only £1.2 million in tax on sales of over £800 million in 2010. eBay responded that it "complies fully with all applicable tax laws". On May 21, 2014, the company revealed that the consumer database of usernames, passwords, phone numbers, and physical addresses was breached between late February and early March. Users were forced to change their passwords. The Syrian Electronic Army took responsibility for the attack and said that it would not misuse the data; however, in a move of website defacement, replaced the front pages of the websites with their own logo. In June 2020, five employees were terminated and were subject to charges of cyberstalking after they were accused of targeting Ina and David Steiner, the editors and publishers of EcommerceBytes, a newsletter that eBay executives viewed as critical of the company. In addition to sending harassing messages and doxing, the defendants "ordered anonymous and disturbing deliveries to the victims’ home, including a preserved fetal pig, a bloody pig Halloween mask, a funeral wreath, a book on surviving the loss of a spouse, and pornography". The defendants also vandalized the couple's home in Natick, Massachusetts. The conspirators pleaded guilty and most were sentenced to prison terms. Wenig, the company's CEO at the time of the harassment campaign, who was frequently targeted by the newsletter and was described as having paranoia over the criticism, was not charged, instead leaving the company in September 2019 with a $57 million severance package. Steve Wymer, chief communication officer, who had ties with local politicians, was fired "for cause" for alleged involvement but was not charged and was hired by the local chapter of the Boys & Girls Clubs of America. Items prohibited to be sold on the website include illegal items such as child pornography, counterfeit products; or items that require licenses to sell such as tobacco, alcoholic beverages, firearms and ammunition, certain knives, human body parts, drugs, tarot readings and spells, virtual in-game items; as well as offensive items such as Nazi memorabilia, flags of the Confederate States of America, and used sex toys. Regulations vary by jurisdiction. In late 1999, a man offered one of his kidneys for auction on eBay, attempting to profit from the potentially lucrative (and, in the United States, illegal) market for organ transplants. Two previously undiscovered species, including the Coelopleurus exquisitus sea urchin in 2006, have been listed for sale on eBay. In January 2010, eBay withdrew a listing of a Dad's Army board game since the box graphics contained images of swastikas, claiming that it was Nazi paraphernalia and, as such, breached the terms of service. eBay was accused of pandering to political correctness. Items stolen from the British Museum in 2013 were auctioned on eBay in 2016. The museum reported that several items of jewelry made of gold, semi-precious stones, and glass, dating from between 1,500 BC and the 19th century AD, were among those missing. One piece of ancient Roman jewelry made from onyx – valued between £25,000 and £50,000, or US$32,000 and US$63,000 – was listed on eBay with a minimum price of £40 (US$50) in 2016. There were no bids made for the treasure. The police are investigating this case. The company said that it is supports local police in investigations and removes listings containing stolen property. In 2021, the estate of Dr. Seuss requested from eBay, and the company complied, to ban the sale of six Dr. Seuss books due to concerns that some images contained therein were racially insensitive. This led to backlash from followers of right-wing politics and ignited a surge of interest in the discontinued books. In September 2023, the United States Department of Justice sued eBay, accusing it of violating the Clean Air Act and other environmental laws by allowing the sale of several illegal products, including devices that defeat automobile pollution controls, restricted-use pesticides, and paint and coating removal products containing methylene chloride. 37°17′43″N 121°55′34″W / 37.2952°N 121.9260°W / 37.2952; -121.9260
{{Infobox Australian place | type = city | name = Canberra | native_name = ''Ngambri''<ref>{{cite web |url=https://www.sbs.com.au/nitv/article/2022/07/04/do-you-know-what-aboriginal-land-youre-today |title=Do you know what Aboriginal land you're on today? |date=4 July 2022}}</ref> | state = ACT | image = Canberra montage 2.jpg | caption = Canberra, from top left to bottom right&ndash;the city viewed from [[Mount Ainslie]], the Land Axis featuring [[Old Parliament House, Canberra|Old Parliament House]] and [[Parliament House, Canberra|New Parliament House]], the [[Australian War Memorial]], the [[National Carillon]], the [[National Gallery of Australia]] and the [[National Library of Australia]] on [[Lake Burley Griffin]] | image2 = Free printable and editable vector map of Canberra Australia.svg | caption2 = City map plan of Canberra | image2_alt = City map plan of Canberra | coordinates = {{coord|35|17|35|S|149|07|37|E|display=inline,title}} | force_national_map = yes | pushpin_label_position = left | pop = 456,692 | pop_year = June 2022 | pop_footnotes = <ref name=ABSGCCSA>{{cite web |title=Regional population, 2021-22 |date=20 April 2023 |url=https://www.abs.gov.au/statistics/people/population/regional-population/2021-22 |publisher=Australian Bureau of Statistics |access-date=27 April 2023}}</ref> | poprank = 8th | density = 503.932 | density_footnotes = | est = 12 March 1913 | elevation = 578 | elevation_footnotes = <ref>{{cite web |title=GFS / BOM data for CANBERRA AIRPORT |url=http://www.bom.gov.au/climate/averages/tables/cw_070014_All.shtml |access-date=16 June 2018 |archive-date=9 July 2015 |archive-url=https://web.archive.org/web/20150709190756/http://www.bom.gov.au/climate/averages/tables/cw_070014_All.shtml |url-status=live}}</ref> | area = 814.2 | area_footnotes = <ref name=area>{{cite web |url=http://www.actpla.act.gov.au/tools_resources/planning_data |title=Planning Data Statistics |publisher=[[Australian Capital Territory Planning and Land Authority|ACT Planning & Land Authority]] |date=21 July 2009 |access-date=13 May 2010 |archive-url=https://web.archive.org/web/20080802163103/http://www.actpla.act.gov.au/tools_resources/planning_data |archive-date=2 August 2008}}</ref> | timezone = AEST | utc = +10:00 | timezone-dst = AEDT | utc-dst = +11:00 | dist1 = 248 | dir1 = SW | location1 = [[Sydney]]<ref>{{cite web |url=http://www.ga.gov.au/cocky/cgi/run/distancedraw2?rec1=87421&placename=sydney&placetype=0&state=0&place1=CANBERRA&place1long=149.133331&place1lat=-35.299999 |title=Great Circle Distance between CANBERRA and SYDNEY |publisher=Geoscience Australia |date=March 2004 |access-date=2 May 2016 |archive-date=7 January 2019 |archive-url=https://web.archive.org/web/20190107044625/http://www.ga.gov.au/cocky/cgi/run/distancedraw2?rec1=87421&placename=sydney&placetype=0&state=0&place1=CANBERRA&place1long=149.133331&place1lat=-35.299999 |url-status=live}}</ref> | dist2 = 654 | dir2 = NE | location2 = [[Melbourne]]<ref>{{cite web |url=http://www.ga.gov.au/cocky/cgi/run/distancedraw2?rec1=248650&placename=melbourne&placetype=0&state=0&place1=CANBERRA&place1long=149.133331&place1lat=-35.299999 |title=Great Circle Distance between CANBERRA and MELBOURNE |publisher=Geoscience Australia |date=March 2004 |access-date=2 May 2016 |archive-date=7 January 2019 |archive-url=https://web.archive.org/web/20190107044626/http://www.ga.gov.au/cocky/cgi/run/distancedraw2?rec1=248650&placename=melbourne&placetype=0&state=0&place1=CANBERRA&place1long=149.133331&place1lat=-35.299999 |url-status=live}}</ref> | dist3 = 958 | dir3 = E | location3 = [[Adelaide]]<ref>{{cite web |url=http://www.ga.gov.au/cocky/cgi/run/distancedraw2?rec1=163285&placename=ADELAIDE&placetype=0&state=0&place1=CANBERRA&place1long=149.133331&place1lat=-35.299999 |title=Great Circle Distance between CANBERRA and ADELAIDE |publisher=Geoscience Australia |date=March 2004 |access-date=2 May 2016 |archive-date=7 January 2019 |archive-url=https://web.archive.org/web/20190107044620/http://www.ga.gov.au/cocky/cgi/run/distancedraw2?rec1=163285&placename=ADELAIDE&placetype=0&state=0&place1=CANBERRA&place1long=149.133331&place1lat=-35.299999 |url-status=live}}</ref> | dist4 = 945 | dir4 = SSW | location4 = [[Brisbane]]<ref>{{cite web |url=http://www.ga.gov.au/cocky/cgi/run/distancedraw2?rec1=126867&placename=BRISBANE&placetype=0&state=0&place1=CANBERRA&place1long=149.133331&place1lat=-35.299999 |title=Great Circle Distance between CANBERRA and BRISBANE |publisher=Geoscience Australia |date=March 2004 |access-date=2 May 2016 |archive-date=7 January 2019 |archive-url=https://web.archive.org/web/20190107044623/http://www.ga.gov.au/cocky/cgi/run/distancedraw2?rec1=126867&placename=BRISBANE&placetype=0&state=0&place1=CANBERRA&place1long=149.133331&place1lat=-35.299999 |url-status=live}}</ref> | dist5 = 3087 | dir5 = ESE | location5 = [[Perth, Western Australia|Perth]]<ref>{{cite web |url=http://www.ga.gov.au/cocky/cgi/run/distancedraw2?rec1=304529&placename=perth&placetype=0&state=0&place1=CANBERRA&place1long=149.133331&place1lat=-35.299999 |title=Great Circle Distance between CANBERRA and PERTH |publisher=Geoscience Australia |date=March 2004 |access-date=2 May 2016 |archive-date=7 January 2019 |archive-url=https://web.archive.org/web/20190107044628/http://www.ga.gov.au/cocky/cgi/run/distancedraw2?rec1=304529&placename=perth&placetype=0&state=0&place1=CANBERRA&place1long=149.133331&place1lat=-35.299999 |url-status=live}}</ref> | local_map = yes | region = | county = | stategov = [[Brindabella electorate|Brindabella]] | stategov2 = [[Ginninderra electorate|Ginninderra]] | stategov3 = [[Kurrajong electorate|Kurrajong]] | stategov4 = [[Murrumbidgee electorate|Murrumbidgee]] | stategov5 = [[Yerrabi electorate|Yerrabi]] | fedgov = [[Division of Canberra|Canberra]] | fedgov2 = [[Division of Fenner|Fenner]] | fedgov3 = [[Division of Bean|Bean]]<ref>{{cite web |url=https://www.aec.gov.au/Electorates/Redistributions/2017/act/final-report/files/act-2018-final-report.pdf |title=Redistribution of the Australian Capital Territory into electoral divisions |author=Augmented Electoral Commission for the Australian Capital Territory |date=July 2018 |quote=The electoral divisions described in this report came into effect from Friday 13 July 2018 ... However, members of the House of Representatives will not represent or contest these electoral divisions until ... a general election. |access-date=16 September 2018 |archive-date=16 September 2018 |archive-url=https://web.archive.org/web/20180916130414/https://www.aec.gov.au/Electorates/Redistributions/2017/act/final-report/files/act-2018-final-report.pdf |url-status=live}}</ref> | logo = | url = | maxtemp = 20.4 | mintemp = 7.0 | rainfall = 579.5 }} [[File:ACT in Australia map.png|thumb|Canberra is in the south-east of Australia]] [[Image:Flag of the Australian Capital Territory.svg|thumb|150px|Flag]] [[Image:Coat of Arms of the Australian Capital Territory.svg|thumb|90px|Coat of arms]] '''Canberra''' is the [[Capital (city)|capital city]] of [[Australia]]. There are 403,468 people who live there.<ref name="ABSGCCSA"/> It does not belong to a state but it is in the [[Australian Capital Territory]] (ACT). Canberra was started in 1913 and in the middle of the city is Lake Burley-Griffin. This lake was created especially for the city. There was a competition to design the new capital city of Australia. The competition was won by an [[architect]] from [[Chicago]] in the [[United States of America]]. His name was Walter Burley Griffin. Griffin designed his [[triangle|triangular]] town as a garden town, so its size expanded over a big area. But as Canberra grew larger and larger, the lake was no longer in the middle of the town, but the people who are in charge of planning the city make sure Canberra is as close to Griffin's original plan as possible. Canberra has a size in area nearly as big as [[Berlin]], but the number of people who live there is much smaller. Before Canberra was founded, Australia’s two most important cities, [[Sydney]] and [[Melbourne]], were arguing which city should be the capital of the [[continent]]. First, Melbourne was elected the capital, but in 1908 there was a vote. The result: an area of 2400 square [[kilometer]]s was given to the country by [[New South Wales]] to build Canberra in. This was the place, 300 kilometers away from Sydney and with a distance of 650 kilometers to Melbourne, where a few years later Canberra was built. The name Canberra is pronounced in various ways and most commonly now as "CAN-brugh" making it sound almost an English name - Canborough ! Some say "CAN-buh-ruh" ; a small number of people may pronounce it "can-BEAR-ruh" ; at an earlier time "can-BERR-a" was very common and that form is still heard although often looked down upon as ignorant and incorrect. But while now an uncommon pronunciation it may be closest to the original name and pronunciation. In the 19th century a farming property on the Limestone Plains was named "Canberry" (or "Camberry") and some think this name is derived in turn from the name of the local tribe - the Ngambri or Kamberra or Camberri people. (Others have suggested the name is derived from an Aboriginal word meaning "meeting place"). Canberra has an [[oceanic climate]] (''Cfb'' in the [[Köppen climate classification]]). Due to Canberra's inland location and fairly high [[altitude]], summers are warm to hot with frequent [[heat wave]]s and [[drought]]s and winters are fairly cold with frequent [[frost]]s during the nights. [[Snow]] is only common on the mountains. [[Tourist]]s who visit Canberra come to see the many national [[monument]]s and other famous sights. Some places to visit are the Government House, Parliament House, the [[High Court of Australia]], the [[Australian War Memorial]], the [[National Gallery of Australia]], the Royal Australian Mint, Telstra Tower, the [[National Library of Australia]], and the [[National Museum of Australia]]. == References == {{reflist}} ==Other websites== {{Commonscat}} [[Category:Canberra| ]] [[Category:1913 establishments]] [[Category:1910s establishments in Australia]]
Canberra (/ˈkænbərə/ KAN-bər-ə) is the capital city of Australia. Founded following the federation of the colonies of Australia as the seat of government for the new nation, it is Australia's largest inland city and the eighth-largest Australian city overall. The city is located at the northern end of the Australian Capital Territory at the northern tip of the Australian Alps, the country's highest mountain range. As of June 2022, Canberra's estimated population was 456,692. The area chosen for the capital had been inhabited by Aboriginal Australians for up to 21,000 years, by groups including the Ngunnawal, Ngunawal and Ngambri. European settlement commenced in the first half of the 19th century, as evidenced by surviving landmarks such as St John's Anglican Church and Blundells Cottage. On 1 January 1901, federation of the colonies of Australia was achieved. Following a long dispute over whether Sydney or Melbourne should be the national capital, a compromise was reached: the new capital would be built in New South Wales, so long as it was at least 100 mi (160 km) from Sydney. The capital city was founded and formally named as Canberra in 1913. A plan by the American architects Walter Burley Griffin and Marion Mahony Griffin was selected after an international design contest, and construction commenced in 1913. Unusual among Australian cities, it is an entirely planned city. The Griffins' plan featured geometric motifs and was centred on axes aligned with significant topographical landmarks such as Black Mountain, Mount Ainslie, Capital Hill and City Hill. Canberra's mountainous location makes it the only mainland Australian city where snow-capped mountains can be seen in winter; although snow in the city itself is uncommon. As the seat of the Government of Australia, Canberra is home to many important institutions of the federal government, national monuments and museums. This includes Parliament House, Government House, the High Court building and the headquarters of numerous government agencies. It is the location of many social and cultural institutions of national significance such as the Australian War Memorial, the Australian National University, the Royal Australian Mint, the Australian Institute of Sport, the National Gallery, the National Museum and the National Library. The city is home to many important institutions of the Australian Defence Force including the Royal Military College Duntroon and the Australian Defence Force Academy. It hosts all foreign embassies in Australia as well as regional headquarters of many international organisations, not-for-profit groups, lobbying groups and professional associations. Canberra has been ranked among the world's best cities to live in and visit. Although the Commonwealth Government remains the largest single employer in Canberra, it is no longer the majority employer. Other major industries have developed in the city, including in health care, professional services, education and training, retail, accommodation and food, and construction. Compared to the national averages, the unemployment rate is lower and the average income higher; tertiary education levels are higher, while the population is younger. At the 2016 Census, 32% of Canberra's inhabitants were reported as having been born overseas. Canberra's design is influenced by the garden city movement and incorporates significant areas of natural vegetation. Its design can be viewed from its highest point at the Telstra Tower and the summit of Mount Ainslie. Other notable features include the National Arboretum, born out of the 2003 Canberra bushfires, and Lake Burley Griffin, named for Walter Burley Griffin. Highlights in the annual calendar of cultural events include Floriade, the largest flower festival in the Southern Hemisphere, the Enlighten Festival, Skyfire, the National Multicultural Festival and Summernats. Canberra's main sporting venues are Canberra Stadium and Manuka Oval. The city is served with domestic and international flights at Canberra Airport, while interstate train and coach services depart from Canberra railway station and the Jolimont Centre respectively. City Interchange is the main hub of Canberra's bus and light rail transport network. The word "Canberra" is derived from the Ngunnawal language of a local Ngunnawal or Ngambri clan who resided in the area and were referred to by the early British colonists as either the Canberry, Kanberri or Nganbra tribe. Joshua John Moore, the first European land-owner in the region, named his grant "Canberry" in 1823 after these people. "Canberry Creek" and "Canberry" first appeared on regional maps from 1830, while the derivative name "Canberra" started to appear from around 1857. Numerous local commentators, including the Ngunnawal elder Don Bell, have speculated upon possible meanings of "Canberra" over the years. These include "meeting place", "woman's breasts" and "the hollow between a woman's breasts". Alternative proposals for the name of the city during its planning included Austral, Australville, Aurora, Captain Cook, Caucus City, Cookaburra, Dampier, Eden, Eucalypta, Flinders, Gonebroke, Home, Hopetoun, Kangaremu, Myola, Meladneyperbane, New Era, Olympus, Paradise, Shakespeare, Sydmelperadbrisho, Swindleville, The National City, Union City, Unison, Wattleton, Wheatwoolgold, Yass-Canberra. The first peoples of the Canberra area include the Ngunnawal, Ngunawal and Ngambri peoples. Other groups claiming a connection to the land include the Ngarigo (who also lived directly to the south) and the Ngambri-Guumaal. Neighbouring groups include the Wandandian to the east, the Walgulu also to the south, Gandangara people to the north and Wiradjuri to the north-west. The first British settlers into the Canberra area described two clans of Ngunnawal people resident to the vicinity. The Canberry or Nganbra clan lived mostly around Sullivan's Creek and had ceremonial grounds at the base of Galambary (Black Mountain), while the Pialligo clan had land around what is now Canberra Airport. The people living here carefully managed and cultivated the land with fire and farmed yams and hunted for food. Archaeological evidence of settlement in the region includes inhabited rock shelters, rock paintings and engravings, burial places, camps and quarry sites as well as stone tools and arrangements. Artefacts suggests early human activity occurred at some point in the area 21,000 years previously. Still today, Ngunnawal men into the present conduct ceremony on the banks of the river, Murrumbidgee River. They travel upstream as they receive their Totems and corresponding responsibilities for land management. 'Murrum' means 'Pathway' and Bidgee means 'Boss'. The submerged limestone caves beneath Lake Burley Griffin contained Aboriginal rock art, some of the only sites in the region. Galambary (Black Mountain) is an important Aboriginal meeting and business site, predominantly for men’s business. According to the Ngunnawal and Ngambri people, Mt Ainslie is primarily for place of women’s business. Black Mountain and Mount Ainslie are referred to as women’s breasts. Galambary was also used by Ngunnawal people as an initiation site, with the mountain itself said to represent the growth of a boy into a man. In October 1820, Charles Throsby led the first British expedition to the area. Four other expeditions occurred between 1820 and 1823 with the first accurate map being produced by explorer Mark John Currie in June 1823. By this stage the area had become known as the Limestone Plains. British settlement of the area probably dates from late 1823, when a sheep station was formed on what is now the Acton Peninsula by James Cowan, the head stockman employed by Joshua John Moore. Moore had received a land grant in the region in 1823 and formally applied to purchase the site on 16 December 1826. He named the property "Canberry". On 30 April 1827, Moore was told by letter that he could retain possession of 1,000 acres (405 ha) at Canberry. Other colonists soon followed Moore's example to take up land in the region. Around 1825 James Ainslie, working on behalf of the wealthy merchant Robert Campbell, arrived to establish a sheep station. He was guided to the region by a local Aboriginal girl who showed him the fine lands of her Pialligo clan. The area then became the property of Campbell and it was initially named Pialligo before Campbell changed it to the Scottish title of Duntroon. Campbell's family later built the imposing stone house that is now the officers' mess of the Royal Military College, Duntroon. The Campbells sponsored settlement by other farmer families to work their land, such as the Southwells of "Weetangera". Other notable early colonists included Henry Donnison, who established the Yarralumla estate—now the site of the official residence of the Governor-General of Australia—in 1827, and John Palmer who employed Duncan Macfarlane to form the Jerrabomberra property in 1828. A year later, John MacPherson established the Springbank estate, becoming the first British owner-occupier in the region. The Anglican church of St John the Baptist, in the suburb of Reid, was consecrated in 1845 and is now the oldest surviving public building in the city. St John's churchyard contains the earliest graves in the district. It has been described as a "sanctuary in the city", remaining a small English village-style church even as the capital grew around it. Canberra's first school, St John's School (now a museum), was situated next to the church and opened in the same year of 1845. It was built to educate local settlers children, including the Blundell children who lived in nearby Blundell's Cottage. As the European presence increased, the Indigenous population dwindled largely due to the destruction of their society, dislocation from their lands and from introduced diseases such as influenza, smallpox, alcoholism and measles. The district's change from a rural area in New South Wales to the national capital started during debates over federation in the late 19th century. Following a long dispute over whether Sydney or Melbourne should be the national capital, a compromise was reached: the new capital would be built in New South Wales, so long as it was at least 100 mi (160 km) from Sydney, with Melbourne to be the temporary seat of government while the new capital was built. A survey was conducted across several sites in New South Wales with Bombala, southern Monaro, Orange, Yass, Albury, Tamworth, Armidale, Tumut and Dalgety all discussed. Dalgety was chosen by the federal parliament and it passed the Seat of Government Act 1904 confirming Dalgety as the site of the nation's capital. However, the New South Wales government refused to cede the required territory as they did not accept the site. In 1906, the New South Wales Government finally agreed to cede sufficient land provided that it was in the Yass-Canberra region as this site was closer to Sydney. Newspaper proprietor John Gale circulated a pamphlet titled 'Dalgety or Canberra: Which?' advocating Canberra to every member of the Commonwealth's seven state and federal parliaments. By many accounts, it was decisive in the selection of Canberra as the site in 1908 as was a result of survey work done by the government surveyor Charles Scrivener. The NSW government ceded the district to the federal government in 1911 and the Federal Capital Territory was established. An international design competition was launched by the Department of Home Affairs on 30 April 1911, closing on 31 January 1912. The competition was boycotted by the Royal Institute of British Architects, the Institution of Civil Engineers and their affiliated bodies throughout the British Empire because the Minister for Home Affairs King O'Malley insisted that the final decision was for him to make rather than an expert in city planning. A total of 137 valid entries were received. O'Malley appointed a three-member board to advise him but they could not reach unanimity. On 24 May 1911, O'Malley came down on the side of the majority of the board with the design by Walter Burley Griffin and Marion Mahony Griffin of Chicago, Illinois, United States, being declared the winner. Second was Eliel Saarinen of Finland and third was Alfred Agache of Brazil but resident in Paris, France. O'Malley then appointed a six-member board to advise him on the implementation of the winning design. On 25 November 1912, the board advised that it could not support the Griffins' plan in its entirety and suggested an alternative plan of its own devising. This plan ostensibly incorporated the best features of the three place-getting designs as well as of a fourth design by H. Caswell, R.C.G. Coulter and W. Scott-Griffiths of Sydney, the rights to which it had purchased. It was this composite plan that was endorsed by Parliament and given formal approval by O'Malley on 10 January 1913. However, it was the Griffin plan which was ultimately proceeded with. In 1913, Walter Burley Griffin was appointed Federal Capital Director of Design and Construction and construction began. On 23 February, King O'Malley drove the first peg in the construction of the future capital city. In 1912, the government invited suggestions from the public as to the name of the future city. Almost 750 names were suggested. At midday on 12 March 1913, Lady Denman, the wife of Governor-General Lord Denman, announced that the city would be named "Canberra" at a ceremony at Kurrajong Hill, which has since become Capital Hill and the site of the present Parliament House. Canberra Day is a public holiday observed in the ACT on the second Monday in March to celebrate the founding of Canberra. After the ceremony, bureaucratic disputes hindered Griffin's work; a Royal Commission in 1916 ruled his authority had been usurped by certain officials and his original plan was reinstated. Griffin's relationship with the Australian authorities was strained and a lack of funding meant that by the time he was fired in 1920, little work had been done. By this time, Griffin had revised his plan, overseen the earthworks of major avenues and established the Glenloch Cork Plantation. The Commonwealth government purchased the pastoral property of Yarralumla in 1913 to provide an official residence for the Governor-General of Australia in the new capital. Renovations began in 1925 to enlarge and modernise the property. In 1927, the property was official dubbed Government House. On 9 May that year, the Commonwealth parliament moved to Canberra with the opening of the Provisional Parliament House. The Prime Minister Stanley Bruce had officially taken up residence in The Lodge a few days earlier. Planned development of the city slowed significantly during the depression of the 1930s and during World War II. Some projects planned for that time, including Roman Catholic and Anglican cathedrals, were never completed. (Nevertheless, in 1973 the Roman Catholic parish church of St. Christopher was remodelled into St. Christopher's Cathedral, Manuka, serving the Archdiocese of Canberra and Goulburn. It is the only cathedral in Canberra.) From 1920 to 1957, three bodies — successively the Federal Capital Advisory Committee, the Federal Capital Commission, and the National Capital Planning and Development Committee — continued to plan the further expansion of Canberra in the absence of Griffin. However, they were only advisory and development decisions were made without consulting them, which increased inefficiency. The largest event in Canberra up to World War II was the 24th Meeting of ANZAAS in January 1939. The Canberra Times described it as "a signal event ... in the history of this, the world's youngest capital city". The city's accommodation was not nearly sufficient to house the 1,250 delegates and a tent city had to be set up on the banks of the Molonglo River. One of the prominent speakers was H. G. Wells, who was a guest of the Governor-General Lord Gowrie for a week. This event coincided with a heatwave across south-eastern Australia during which the temperature in Canberra reached 108.5 degrees Fahrenheit (42.5 Celsius) on 11 January. On Friday, 13 January, the Black Friday bushfires caused 71 deaths in Victoria and Wells accompanied the Governor-General on his tour of areas threatened by fires. Immediately after the end of the war, Canberra was criticised for resembling a village and its disorganised collection of buildings was deemed ugly. Canberra was often derisively described as "several suburbs in search of a city". Prime Minister Sir Robert Menzies regarded the state of the national capital as an embarrassment. Over time his attitude changed from one of contempt to that of championing its development. He fired two ministers charged with the development of the city for poor performance. Menzies remained in office for over a decade and in that time the development of the capital sped up rapidly. The population grew by more than 50 per cent in every five-year period from 1955 to 1975. Several Government departments, together with public servants, were moved to Canberra from Melbourne following the war. Government housing projects were undertaken to accommodate the city's growing population. The National Capital Development Commission (NCDC) formed in 1957 with executive powers and ended four decades of disputes over the shape and design of Lake Burley Griffin — the centrepiece of Griffin's design — and construction was completed in 1964 after four years of work. The completion of the lake finally laid the platform for the development of Griffin's Parliamentary Triangle. Since the initial construction of the lake, various buildings of national importance have been constructed on its shores. The newly built Australian National University was expanded and sculptures as well as monuments were built. A new National Library was constructed within the Parliamentary Triangle, followed by the High Court and the National Gallery. Suburbs in Canberra Central (often referred to as North Canberra and South Canberra) were further developed in the 1950s and urban development in the districts of Woden Valley and Belconnen commenced in the mid and late 1960s respectively. Many of the new suburbs were named after Australian politicians such as Barton, Deakin, Reid, Braddon, Curtin, Chifley and Parkes. On 9 May 1988, a larger and permanent Parliament House was opened on Capital Hill as part of Australia's bicentenary celebrations. The Commonwealth Parliament moved there from the Provisional Parliament House, now known as Old Parliament House. In December 1988, the Australian Capital Territory was granted full self-government by the Commonwealth Parliament, a step proposed as early as 1965. Following the first election on 4 March 1989, a 17-member Legislative Assembly sat at temporary offices at 1 Constitution Avenue, Civic, on 11 May 1989. Permanent premises were opened on London Circuit in 1994. The Australian Labor Party formed the ACT's first government, led by the Chief Minister Rosemary Follett, who made history as Australia's first female head of government. Parts of Canberra were engulfed by bushfires on 18 January 2003 that killed four people, injured 435 and destroyed more than 500 homes as well as the major research telescopes of Australian National University's Mount Stromlo Observatory. Throughout 2013, several events celebrated the 100th anniversary of the naming of Canberra. On 11 March 2014, the last day of the centennial year, the Canberra Centenary Column was unveiled in City Hill. Other works included The Skywhale, a hot air balloon designed by the sculptor Patricia Piccinini, and StellrScope by visual media artist Eleanor Gates-Stuart. On 7 February 2021, The Skywhale was joined by Skywhalepapa to create a Skywhale family, an event marked by Skywhale-themed pastries and beer produced by local companies as well as an art pop song entitled "We are the Skywhales". In 2014, Canberra was named the best city to live in the world by the Organisation for Economic Co-operation and Development, and was named the third best city to visit in the world by Lonely Planet in 2017. Canberra covers an area of 814.2 km (314.4 sq mi) and is located near the Brindabella Ranges (part of the Australian Alps), approximately 150 km (93 mi) inland from Australia's east coast. It has an elevation of approximately 580 m (1,900 ft) AHD; the highest point is Mount Majura at 888 m (2,913 ft). Other low mountains include Mount Taylor 855 m (2,805 ft), Mount Ainslie 843 m (2,766 ft), Mount Mugga Mugga 812 m (2,664 ft) and Black Mountain 812 m (2,664 ft). The native forest in the Canberra region was almost wholly eucalypt species and provided a resource for fuel and domestic purposes. By the early 1960s, logging had depleted the eucalypt, and concern about water quality led to the forests being closed. Interest in forestry began in 1915 with trials of a number of species including Pinus radiata on the slopes of Mount Stromlo. Since then, plantations have been expanded, with the benefit of reducing erosion in the Cotter catchment, and the forests are also popular recreation areas. The urban environs of the city of Canberra straddle the Ginninderra plain, Molonglo plain, the Limestone plain, and the Tuggeranong plain (Isabella's Plain). The Molonglo River which flows across the Molonglo plain has been dammed to form the national capital's iconic feature Lake Burley Griffin. The Molonglo then flows into the Murrumbidgee north-west of Canberra, which in turn flows north-west toward the New South Wales town of Yass. The Queanbeyan River joins the Molonglo River at Oaks Estate just within the ACT. A number of creeks, including Jerrabomberra and Yarralumla Creeks, flow into the Molonglo and Murrumbidgee. Two of these creeks, the Ginninderra and Tuggeranong, have similarly been dammed to form Lakes Ginninderra and Tuggeranong. Until recently the Molonglo River had a history of sometimes calamitous floods; the area was a flood plain prior to the filling of Lake Burley Griffin. Under the Köppen-Geiger classification, Canberra has an oceanic climate (Cfb). In January, the warmest month, the average high is approximately 29 °C (84 °F); in July, the coldest month, the average high drops to approximately 12 °C (54 °F). Frost is common in the winter months. Snow is rare in the CBD (central business district) due to being on the leeward (eastern) side of the dividing range, but the surrounding areas get annual snowfall through winter and often the snow-capped Brindabella Range can be seen from the CBD. The last significant snowfall in the city centre was in 1968. Canberra is often affected by foehn winds, especially in winter and spring, evident by its anomalously warm maxima relative to altitude. The highest recorded maximum temperature was 44.0 °C (111.2 °F) on 4 January 2020. Winter 2011 was Canberra's warmest winter on record, approximately 2 °C (4 °F) above the average temperature. The lowest recorded minimum temperature was −10.0 °C (14.0 °F) on the morning of 11 July 1971. Light snow falls only once in every few years, and is usually not widespread and quickly dissipates. Canberra is protected from the west by the Brindabellas which create a strong rain shadow in Canberra's valleys. Canberra gets 100.4 clear days annually. Annual rainfall is the third lowest of the capital cities (after Adelaide and Hobart) and is spread fairly evenly over the seasons, with late spring bringing the highest rainfall. Thunderstorms occur mostly between October and April, owing to the effect of summer and the mountains. The area is generally sheltered from a westerly wind, though strong northwesterlies can develop. A cool, vigorous afternoon easterly change, colloquially referred to as a 'sea-breeze' or the 'Braidwood Butcher', is common during the summer months and often exceeds 40 km/h in the city. Canberra is also less humid than the nearby coastal areas. Canberra was severely affected by smoke haze during the 2019/2020 bushfires. On 1 January 2020, Canberra had the worst air quality of any major city in the world, with an AQI of 7700 (USAQI 949). Canberra is a planned city and the inner-city area was originally designed by Walter Burley Griffin, a major 20th-century American architect. Within the central area of the city near Lake Burley Griffin, major roads follow a wheel-and-spoke pattern rather than a grid. Griffin's proposal had an abundance of geometric patterns, including concentric hexagonal and octagonal streets emanating from several radii. However, the outer areas of the city, built later, are not laid out geometrically. Lake Burley Griffin was deliberately designed so that the orientation of the components was related to various topographical landmarks in Canberra. The lakes stretch from east to west and divided the city in two; a land axis perpendicular to the central basin stretches from Capital Hill—the eventual location of the new Parliament House on a mound on the southern side—north northeast across the central basin to the northern banks along Anzac Parade to the Australian War Memorial. This was designed so that looking from Capital Hill, the War Memorial stood directly at the foot of Mount Ainslie. At the southwestern end of the land axis was Bimberi Peak, the highest mountain in the ACT, approximately 52 km (32 mi) south west of Canberra. The straight edge of the circular segment that formed the central basin of Lake Burley Griffin was perpendicular to the land axis and designated the water axis, and it extended northwest towards Black Mountain. A line parallel to the water axis, on the northern side of the city, was designated the municipal axis. The municipal axis became the location of Constitution Avenue, which links City Hill in Civic Centre and both Market Centre and the Defence precinct on Russell Hill. Commonwealth Avenue and Kings Avenue were to run from the southern side from Capital Hill to City Hill and Market Centre on the north respectively, and they formed the western and eastern edges of the central basin. The area enclosed by the three avenues was known as the Parliamentary Triangle, and formed the centrepiece of Griffin's work. The Griffins assigned spiritual values to Mount Ainslie, Black Mountain, and Red Hill and originally planned to cover each of these in flowers. That way each hill would be covered with a single, primary colour which represented its spiritual value. This part of their plan never came to fruition, as World War I slowed construction and planning disputes led to Griffin's dismissal by Prime Minister Billy Hughes after the war ended. The urban areas of Canberra are organised into a hierarchy of districts, town centres, group centres, local suburbs as well as other industrial areas and villages. There are seven residential districts, each of which is divided into smaller suburbs, and most of which have a town centre which is the focus of commercial and social activities. The districts were settled in the following chronological order: The Canberra Central district is substantially based on Walter Burley Griffin's designs. In 1967 the then National Capital Development Commission adopted the "Y Plan" which laid out future urban development in Canberra around a series of central shopping and commercial area known as the 'town centres' linked by freeways, the layout of which roughly resembled the shape of the letter Y, with Tuggeranong at the base of the Y and Belconnen and Gungahlin located at the ends of the arms of the Y. Development in Canberra has been closely regulated by government, both through planning processes and the use of crown lease terms that have tightly limited the use of parcels of land. Land in the ACT is held on 99-year crown leases from the national government, although most leases are now administered by the Territory government. There have been persistent calls for constraints on development to be liberalised, but also voices in support of planning consistent with the original 'bush capital' and 'urban forest' ideals that underpin Canberra's design. Many of Canberra's suburbs are named after former Prime Ministers, famous Australians, early settlers, or use Aboriginal words for their title. Street names typically follow a particular theme; for example, the streets of Duffy are named after Australian dams and reservoirs, the streets of Dunlop are named after Australian inventions, inventors and artists and the streets of Page are named after biologists and naturalists. Most diplomatic missions are located in the suburbs of Yarralumla, Deakin and O'Malley. There are three light industrial areas: the suburbs of Fyshwick, Mitchell and Hume. The average Canberran was responsible for 13.7 tonnes of greenhouse gases in 2005. In 2012, the ACT Government legislated greenhouse gas targets to reduce its emissions by 40 per cent from 1990 levels by 2020, 80 per cent by 2050, with no net emissions by 2060. The government announced in 2013 a target for 90% of electricity consumed in the ACT to be supplied from renewable sources by 2020, and in 2016 set an ambitious target of 100% by 2020. In 1996 Canberra became the first city in the world to set a vision of no waste, proposing an ambitious target of 2010 for completion. The strategy aimed to achieve a waste-free society by 2010, through the combined efforts of industry, government and community. By early 2010, it was apparent that though it had reduced waste going to landfill, the ACT initiative's original 2010 target for absolutely zero landfill waste would be delayed or revised to meet the reality. Plastic bags made of polyethylene polymer with a thickness of less than 35 µm were banned from retail distribution in the ACT from November 2011. The ban was introduced by the ACT Government in an effort to make Canberra more sustainable. Of all waste produced in the ACT, 75 per cent is recycled. Average household food waste in the ACT remains above the Australian average, costing an average $641 per household per annum. Canberra's annual Floriade festival features a large display of flowers every Spring in Commonwealth Park. The organisers of the event have a strong environmental standpoint, promoting and using green energy, "green catering", sustainable paper, the conservation and saving of water. The event is also smoke-free. There is no local council or city government for the city of Canberra. The Australian Capital Territory Legislative Assembly performs the roles of both a city council for the city and a territory government for the rest of the Australian Capital Territory. However, the vast majority of the population of the Territory reside in Canberra and the city is therefore the primary focus of the ACT Government. The assembly consists of 25 members elected from five districts using proportional representation. The five districts are Brindabella, Ginninderra, Kurrajong, Murrumbidgee and Yerrabi, which each elect five members. The Chief Minister is elected by the Members of the Legislative Assembly (MLA) and selects colleagues to serve as ministers alongside him or her in the Executive, known informally as the cabinet. Whereas the ACT has federally been dominated by Labor, the Liberals have been able to gain some footing in the ACT Legislative Assembly and were in government during a period of six and half years from 1995 and 2001. Labor took back control of the Assembly in 2001. At the 2004 election, Chief Minister Jon Stanhope and the Labor Party won 9 of the 17 seats allowing them to form the ACT's first majority government. Since 2008, the ACT has been governed by a coalition of Labor and the Greens. As of 2022, the Chief Minister was Andrew Barr from the Australian Labor Party. The Australian federal government retains some influence over the ACT government. In the administrative sphere, most frequently this is through the actions of the National Capital Authority which is responsible for planning and development in areas of Canberra which are considered to be of national importance or which are central to Griffin's plan for the city, such as the Parliamentary Triangle, Lake Burley Griffin, major approach and processional roads, areas where the Commonwealth retains ownership of the land or undeveloped hills and ridge-lines (which form part of the Canberra Nature Park). The national government also retains a level of control over the Territory Assembly through the provisions of the Australian Capital Territory (Self-Government) Act 1988. This federal act defines the legislative power of the ACT assembly. The ACT was given its first federal parliamentary representation in 1949 when it gained a seat in the House of Representatives, the Division of Australian Capital Territory. However, the ACT member could only vote on matters directly affecting the territory. In 1974, the ACT was allocated two Senate seats and the House of Representatives seat was divided into two. A third was created in 1996, but was abolished in 1998 because of changes to the regional demographic distribution. At the 2019 election, the third seat has been reintroduced as the Division of Bean. The House of Representatives seats have mostly been held by Labor and usually by comfortable margins. The Labor Party has polled at least seven percentage points more than the Liberal Party at every federal election since 1990 and their average lead since then has been 15 percentage points. The ALP and the Liberal Party held one Senate seat each until the 2022 election when Independent candidate David Pocock unseated the Liberal candidate Zed Seselja. The Australian Federal Police (AFP) provides all of the constabulary services in the territory in a manner similar to state police forces, under a contractual agreement with the ACT Government. The AFP does so through its community policing arm ACT Policing. People who have been charged with offences are tried either in the ACT Magistrates Court or, for more severe offences, the ACT Supreme Court. Prior to its closure in 2009, prisoners were held in remand at the Belconnen Remand Centre in the ACT but usually imprisoned in New South Wales. The Alexander Maconochie Centre was officially opened on 11 September 2008 by then Chief Minister Jon Stanhope. The total cost for construction was $130 million. The ACT Civil and Administrative Tribunal deal with minor civil law actions and other various legal matters. Canberra has the lowest rate of crime of any capital city in Australia as of 2019. As of 2016 the most common crimes in the ACT were property related crimes, unlawful entry with intent and motor vehicle theft. They affected 2,304 and 966 people (580 and 243 per 100,000 persons respectively). Homicide and related offences—murder, attempted murder and manslaughter, but excluding driving causing death and conspiracy to murder—affect 1.0 per 100,000 persons, which is below the national average of 1.9 per 100,000. Rates of sexual assault (64.4 per 100,000 persons) are also below the national average (98.5 per 100,000). However the 2017 crime statistics showed a rise in some types of personal crime, notably burglaries, thefts and assaults. In February 2020, the unemployment rate in Canberra was 2.9% which was lower than the national unemployment rate of 5.1%. As a result of low unemployment and substantial levels of public sector and commercial employment, Canberra has the highest average level of disposable income of any Australian capital city. The gross average weekly wage in Canberra is $1827 compared with the national average of $1658 (November 2019). The median house price in Canberra as of February 2020 was $745,000, lower than only Sydney among capital cities of more than 100,000 people, having surpassed Melbourne and Perth since 2005. The median weekly rent paid by Canberra residents is higher than rents in all other states and territories. As of January 2014 the median unit rent in Canberra was $410 per week and median housing rent was $460, making the city the third most expensive in the country. Factors contributing to this higher weekly rental market include; higher average weekly incomes, restricted land supply, and inflationary clauses in the ACT Residential Tenancies Act. The city's main industry is public administration and safety, which accounted for 27.1% of Gross Territory Product in 2018-19 and employed 32.49% of Canberra's workforce. The headquarters of many Australian Public Service agencies are located in Canberra, and Canberra is also host to several Australian Defence Force establishments, most notably the Australian Defence Force headquarters and HMAS Harman, which is a naval communications centre that is being converted into a tri-service, multi-user depot. Other major sectors by employment include Health Care (10.54%), Professional Services (9.77%), Education and Training (9.64%), Retail (7.27%), Accommodation & Food (6.39%) and Construction (5.80%). The former RAAF Fairbairn, adjacent to the Canberra Airport was sold to the operators of the airport, but the base continues to be used for RAAF VIP flights. A growing number of software vendors have based themselves in Canberra, to capitalise on the concentration of government customers; these include Tower Software and RuleBurst. A consortium of private and government investors is making plans for a billion-dollar data hub, with the aim of making Canberra a leading centre of such activity in the Asia-Pacific region. A Canberra Cyber Security Innovation Node was established in 2019 to grow the ACT's cyber security sector and related space, defence and education industries. At the 2021 census, the population of Canberra was 453,558, up from 395,790 at the 2016 census, and 355,596 at the 2011 census. Canberra has been the fastest-growing city in Australia in recent years, having grown 23.3% from 2011-2021. Canberrans are relatively young, highly mobile and well educated. The median age is 35 years and only 12.7% of the population is aged over 65 years. Between 1996 and 2001, 61.9% of the population either moved to or from Canberra, which was the second highest mobility rate of any Australian capital city. As at May 2017, 43% of ACT residents (25–64) had a level of educational attainment equal to at least a bachelor's degree, significantly higher that the national average of 31%. According to statistics collected by the National Australia Bank and reported in The Canberra Times, Canberrans on average give significantly more money to charity than Australians in other states and territories, for both dollar giving and as a proportion of income. At the 2016 census, the most commonly nominated ancestries were: The 2016 census showed that 32% of Canberra's inhabitants were born overseas. Of inhabitants born outside Australia, the most prevalent countries of birth were England, China, India, New Zealand and the Philippines. 1.6% of the population, or 6,476 people, identified as Indigenous Australians (Aboriginal Australians and Torres Strait Islanders) in 2016. At the 2016 census, 72.7% of people spoke only English at home. The other languages most commonly spoken at home were Mandarin (3.1%), Vietnamese (1.1%), Cantonese (1%), Hindi (0.9%) and Spanish (0.8%). On census night in 2016, approximately 50.0% of ACT residents described themselves as Christian (excluding not stated responses), the most common denominations being Catholic and Anglican; 36.2% described themselves as having no religion. The two main tertiary institutions are the Australian National University (ANU) in Acton and the University of Canberra (UC) in Bruce, with over 10,500 and 8,000 full-time-equivalent students respectively. Established in 1946, the ANU has always had a strong research focus and is ranked among the leading universities in the world and the best in Australia by The Times Higher Education Supplement and the Shanghai Jiao Tong World University Rankings. There are two religious university campuses in Canberra: Signadou in the northern suburb of Watson is a campus of the Australian Catholic University; St Mark's Theological College in Barton is part of the secular Charles Sturt University. The ACT Government announced on 5 March 2020 that the CIT campus and an adjoining carpark in Reid would be leased to the University of New South Wales (UNSW) for a peppercorn lease, for it to develop as a campus for a new UNSW Canberra. UNSW released a master plan in 2021 for a 6,000 student campus to be realised over 15 years at a cost of $1 billion. The Australian Defence College has two campuses: the Australian Command and Staff College (ACSC) plus the Centre for Defence and Strategic Studies (CDSS) at Weston, and the Australian Defence Force Academy (ADFA) beside the Royal Military College, Duntroon located in the inner-northern suburb of Campbell. ADFA teaches military undergraduates and postgraduates and includes UNSW@ADFA, a campus of the University of New South Wales; Duntroon provides Australian Army officer training. Tertiary level vocational education is also available through the Canberra Institute of Technology (CIT), with campuses in Bruce, Reid, Gungahlin, Tuggeranong and Fyshwick. The combined enrolment of the CIT campuses was over 28,000 students in 2019. Following the transfer of land in Reid for the new UNSW Canberra, a new CIT Woden is scheduled to be completed by 2025. In 2016 there were 132 schools in Canberra; 87 were operated by the government and 45 were private. During 2006, the ACT Government announced closures of up to 39 schools, to take effect from the end of the school year, and after a series of consultations unveiled its Towards 2020: Renewing Our Schools policy. As a result, some schools closed during the 2006–08 period, while others were merged; the creation of combined primary and secondary government schools was to proceed over a decade. The closure of schools provoked significant opposition. Most suburbs were planned to include a primary and a nearby preschool; these were usually located near open areas where recreational and sporting activities were easily available. Canberra also has the highest percentage of non-government (private) school students in Australia, accounting for 40.6 per cent of ACT enrollments. Canberra is home to many national monuments and institutions such as the Australian War Memorial, the Australian Institute of Aboriginal and Torres Strait Islander Studies, the National Gallery of Australia, the National Portrait Gallery, the National Library, the National Archives, the Australian Academy of Science, the National Film & Sound Archive and the National Museum. Many Commonwealth government buildings in Canberra are open to the public, including Parliament House, the High Court and the Royal Australian Mint. Lake Burley Griffin is the site of the Captain James Cook Memorial and the National Carillon. Other sites of interest include the Australian–American Memorial, Commonwealth Park, Commonwealth Place, the Telstra Tower, the Australian National Botanic Gardens, the National Zoo and Aquarium, the National Dinosaur Museum, and Questacon – the National Science and Technology Centre. The Canberra Museum and Gallery in the city is a repository of local history and art, housing a permanent collection and visiting exhibitions. Several historic homes are open to the public: Lanyon and Tuggeranong Homesteads in the Tuggeranong Valley, Mugga-Mugga in Symonston, and Blundells' Cottage in Parkes all display the lifestyle of the early European settlers. Calthorpes' House in Red Hill is a well-preserved example of a 1920s house from Canberra's very early days. Strathnairn Homestead is an historic building which also dates from the 1920s. Canberra has many venues for live music and theatre: the Canberra Theatre and Playhouse which hosts many major concerts and productions; and Llewellyn Hall (within the ANU School of Music), a world-class concert hall are two of the most notable. The Street Theatre is a venue with less mainstream offerings. The Albert Hall was the city's first performing arts venue, opened in 1928. It was the original performance venue for theatre groups such as the Canberra Repertory Society. Stonefest was a large annual festival, for some years one of the biggest festivals in Canberra. It was downsized and rebranded as Stone Day in 2012. There are numerous bars and nightclubs which also offer live entertainment, particularly concentrated in the areas of Dickson, Kingston and the city. Most town centres have facilities for a community theatre and a cinema, and they all have a library. Popular cultural events include the National Folk Festival, the Royal Canberra Show, the Summernats car festival, Enlighten festival, the National Multicultural Festival in February and the Celebrate Canberra festival held over 10 days in March in conjunction with Canberra Day. Canberra maintains sister-city relationships with both Nara, Japan and Beijing, China. Canberra has friendship-city relationships with both Dili, East Timor and Hangzhou, China. City-to-city relationships encourage communities and special interest groups both locally and abroad to engage in a wide range of exchange activities. The Canberra Nara Candle Festival held annually in spring, is a community celebration of the Canberra Nara Sister City relationship. The festival is held in Canberra Nara Park on the shores of Lake Burley Griffin. The history of Canberra was told in the 1938 radio feature Canberra the Great. As Australia's capital, Canberra is the most important centre for much of Australia's political reportage and thus all the major media, including the Australian Broadcasting Corporation, the commercial television networks, and the metropolitan newspapers maintain local bureaus. News organisations are represented in the press gallery, a group of journalists who report on the national parliament. The National Press Club of Australia in Barton has regular television broadcasts of its lunches at which a prominent guest, typically a politician or other public figure, delivers a speech followed by a question-and-answer session. Canberra has a daily newspaper, The Canberra Times, which was established in 1926. There are also several free weekly publications, including news magazines CityNews and Canberra Weekly as well as entertainment guide BMA Magazine. BMA Magazine first went to print in 1992; the inaugural edition featured coverage of the Nirvana Nevermind tour. There are a number of AM and FM stations broadcasting in Canberra (AM/FM Listing). The main commercial operators are the Capital Radio Network (2CA and 2CC), and Austereo/ARN (104.7 and Mix 106.3). There are also several community operated stations. A DAB+ digital radio trial is also in operation, it simulcasts some of the AM/FM stations, and also provides several digital only stations (DAB+ Trial Listing). Five free-to-air television stations service Canberra: Each station broadcasts a primary channel and several multichannels. Of the three main commercial networks: Prior to 1989, Canberra was serviced by just the ABC, SBS and Capital Television (CTC), which later became Ten Capital in 1994 then Southern Cross Ten in 2002 then Channel 9/Southern Cross Nine in 2016 and finally Channel 10 in 2021, with Prime Television (now Prime7) and WIN Television arriving as part of the Government's regional aggregation program in that year. Pay television services are available from Foxtel (via satellite) and telecommunications company TransACT (via cable). In addition to local sporting leagues, Canberra has a number of sporting teams that compete in national and international competitions. The best known teams are the Canberra Raiders and the ACT Brumbies who play rugby league and rugby union respectively; both have been champions of their leagues. Both teams play their home games at Canberra Stadium, which is the city's largest stadium and was used to hold group matches in football for the 2000 Summer Olympics and in rugby union for the 2003 Rugby World Cup. Canberra United represents the city in the A-League Women (formerly the W-League), the national women's soccer league and were champions in the 2011–12 season. A men's team is set to join the A-League Men in the 2024–25 season. The city also has a successful basketball team, the Canberra Capitals, which has won seven out of the last eleven national women's basketball titles. The Canberra Vikings represent the city in the National Rugby Championship and finished second in the 2015 season. There are also teams that participate in national competitions in netball, field hockey, ice hockey, cricket and baseball. The historic Prime Minister's XI cricket match is played at Manuka Oval annually. Other significant annual sporting events include the Canberra Marathon and the City of Canberra Half Ironman Triathlon. Canberra has been bidding for an Australian Football League club since 1981 when Australian rules in the Australian Capital Territory was more popular. While the league has knocked back numerous proposals, according to the AFL Canberra belongs to the Greater Western Sydney Giants who play three home games at Manuka Oval each season. Other significant annual sporting events include the Canberra Marathon and the City of Canberra Half Ironman Triathlon. The Australian Institute of Sport (AIS) is located in the Canberra suburb of Bruce. The AIS is a specialised educational and training institution providing coaching for elite junior and senior athletes in a number of sports. The AIS has been operating since 1981 and has achieved significant success in producing elite athletes, both local and international. The majority of Australia's team members and medallists at the 2000 Summer Olympics in Sydney were AIS graduates. Canberra has numerous sporting ovals, golf courses, skate parks, and swimming pools that are open to the public. Tennis courts include those at the National Sports Club, Lyneham, former home of the Canberra Women's Tennis Classic. A Canberra-wide series of bicycle paths are available to cyclists for recreational and sporting purposes. Canberra Nature Parks have a large range of walking paths, horse and mountain bike trails. Water sports like sailing, rowing, dragon boating and water skiing are held on Canberra's lakes. The Rally of Canberra is an annual motor sport event, and from 2000 to 2002, Canberra hosted the Canberra 400 event for V8 Supercars on the temporary Canberra Street Circuit, which was located inside the Parliamentary Triangle. A popular form of exercise for people working near or in the Parliamentary Triangle is to do the "bridge to bridge walk/run" of about 5 km around Lake Burley Griffin, crossing the Commonwealth Avenue Bridge and Kings Avenue Bridge, using the paths beside the lake. The walk takes about 1 hour, making it ideal for a lunchtime excursion. This is also popular on weekends. Such was the popularity during the COVID-19 isolation in 2020 that the ACT Government initiated a 'Clockwise is COVID-wise' rule for walkers and runners. Canberra has two large public hospitals, the approximately 600-bed Canberra Hospital—formerly the Woden Valley Hospital—in Garran and the 174-bed Calvary Public Hospital in Bruce. Both are teaching institutions. The largest private hospital is the Calvary John James Hospital in Deakin. Calvary Private Hospital in Bruce and Healthscope's National Capital Private Hospital in Garran are also major healthcare providers. The Royal Canberra Hospital was located on Acton Peninsula on Lake Burley Griffin; it was closed in 1991 and was demolished in 1997 in a controversial and fatal implosion to facilitate construction of the National Museum of Australia. The city has 10 aged care facilities. Canberra's hospitals receive emergency cases from throughout southern New South Wales, and ACT Ambulance Service is one of four operational agencies of the ACT Emergency Services Authority. NETS provides a dedicated ambulance service for inter-hospital transport of sick newborns within the ACT and into surrounding New South Wales. The automobile is by far the dominant form of transport in Canberra. The city is laid out so that arterial roads connecting inhabited clusters run through undeveloped areas of open land or forest, which results in a low population density; this also means that idle land is available for the development of future transport corridors if necessary without the need to build tunnels or acquire developed residential land. In contrast, other capital cities in Australia have substantially less green space. Canberra's districts are generally connected by parkways—limited access dual carriageway roads with speed limits generally set at a maximum of 100 km/h (62 mph). An example is the Tuggeranong Parkway which links Canberra's CBD and Tuggeranong, and bypasses Weston Creek. In most districts, discrete residential suburbs are bounded by main arterial roads with only a few residential linking in, to deter non-local traffic from cutting through areas of housing. In an effort to improve road safety, traffic cameras were first introduced to Canberra by the Kate Carnell Government in 1999. The traffic cameras installed in Canberra include fixed red-light and speed cameras and point-to-point speed cameras; together they bring in revenue of approximately $11 million per year in fines. ACTION, the government-operated bus service, provides public transport throughout the city. CDC Canberra provides bus services between Canberra and nearby areas of New South Wales of (Murrumbateman and Yass) and as Qcity Transit (Queanbeyan). A light rail line commenced service on 20 April 2019 linking the CBD with the northern district of Gungahlin. A planned Stage 2A of Canberra's light rail network will run from Alinga Street station to Commonwealth Park, adding three new stops at City West, City South and Commonwealth Park. In February 2021 ACT Minister for Transport and City Services Chris Steel said he expects construction on Stage 2A to commence in the 2021-22 financial year, and for "tracks to be laid" by the next Territory election in 2024. At the 2016 census, 7.1% of the journeys to work involved public transport, while 4.5% walked to work. There are two local taxi companies. Aerial Capital Group enjoyed monopoly status until the arrival of Cabxpress in 2007. In October 2015 the ACT Government passed legislation to regulate ride sharing, allowing ride share services including Uber to operate legally in Canberra. The ACT Government was the first jurisdiction in Australia to enact legislation to regulate the service. Since then many other ride sharing and taxi services have started in ACT namely Ola, Glide Taxi and GoCatch An interstate NSW TrainLink railway service connects Canberra to Sydney. Canberra railway station is in the inner south suburb of Kingston. Between 1920 and 1922 the train line crossed the Molonglo River and ran as far north as the city centre, although the line was closed following major flooding and was never rebuilt, while plans for a line to Yass were abandoned. A 3 ft 6 in (1,067 mm) gauge construction railway was built in 1923 between the Yarralumla brickworks and the provisional Parliament House; it was later extended to Civic, but the whole line was closed in May 1927. Train services to Melbourne are provided by way of a NSW TrainLink bus service which connects with a rail service between Sydney and Melbourne in Yass, about a one-hour drive from Canberra. Plans to establish a high-speed rail service between Melbourne, Canberra and Sydney, have not been implemented, as the various proposals have been deemed economically unviable. The original plans for Canberra included proposals for railed transport within the city, however none eventuated. The phase 2 report of the most recent proposal, the High Speed Rail Study, was published by the Department of Infrastructure and Transport on 11 April 2013. A railway connecting Canberra to Jervis Bay was also planned but never constructed. Canberra is about three hours by road from Sydney on the Federal Highway (National Highway 23), which connects with the Hume Highway (National Highway 31) near Goulburn, and seven hours by road from Melbourne on the Barton Highway (National Highway 25), which joins the Hume Highway at Yass. It is a two-hour drive on the Monaro Highway (National Highway 23) to the ski fields of the Snowy Mountains and the Kosciuszko National Park. Batemans Bay, a popular holiday spot on the New South Wales coast, is also two hours away via the Kings Highway. Canberra Airport provides direct domestic services to Adelaide, Brisbane, Cairns, Darwin, Gold Coast, Hobart, Melbourne, Perth, Sunshine Coast and Sydney with connections to other domestic centres. There are also direct flights to small regional towns: Ballina, Dubbo, Newcastle and Port Macquarie in New South Wales. Canberra Airport is, as of September 2013, designated by the Australian Government Department of Infrastructure and Regional Development as a restricted use designated international airport. International flights have previously been operated by both Singapore Airlines and Qatar Airways. Fiji Airways has announced direct flights to Nadi commencing in July 2023. Until 2003 the civilian airport shared runways with RAAF Base Fairbairn. In June of that year, the Air Force base was decommissioned and from that time the airport was fully under civilian control. Canberra has one of the highest rates of active travel of all Australian major cities, with 7.1 per cent of commuters walking or cycling to work in 2011. An ACT Government survey conducted in late 2010 found that Canberrans walk an average of 26 minutes each day. According to The Canberra Times in March 2014, Canberra's cyclists are involved in an average of four reported collisions every week. The newspaper also reported that Canberra is home to 87,000 cyclists, translating to the highest cycling participation rate in Australia; and, with higher popularity, bike injury rates in 2012 were twice the national average. Since late 2020, two scooter-sharing systems have been operational in Canberra: orange scooters from Neuron Mobility and purple scooters from Beam Mobility, both Singapore-based companies that operate in many Australian cities. These services cover much of Canberra Central and Central Belconnen, with plans to expand coverage to more areas of the city in 2022. The government-owned Icon Water manages Canberra's water and sewerage infrastructure. ActewAGL is a joint venture between ACTEW and AGL, and is the retail provider of Canberra's utility services including water, natural gas, electricity, and also some telecommunications services via a subsidiary TransACT. Canberra's water is stored in four reservoirs, the Corin, Bendora and Cotter dams on the Cotter River and the Googong Dam on the Queanbeyan River. Although the Googong Dam is located in New South Wales, it is managed by the ACT government. Icon Water owns Canberra's two wastewater treatment plants, located at Fyshwick and on the lower reaches of the Molonglo River. Electricity for Canberra mainly comes from the national power grid through substations at Holt and Fyshwick (via Queanbeyan). Power was first supplied from the Kingston Powerhouse near the Molonglo River, a thermal plant built in 1913, but this was finally closed in 1957. The ACT has four solar farms, which were opened between 2014 and 2017: Royalla (rated output of 20 megawatts, 2014), Mount Majura (2.3 MW, 2016), Mugga Lane (13 MW, 2017) and Williamsdale (11 MW, 2017). In addition, numerous houses in Canberra have photovoltaic panels or solar hot water systems. In 2015 and 2016, rooftop solar systems supported by the ACT government's feed-in tariff had a capacity of 26.3 megawatts, producing 34,910 MWh. In the same year, retailer-supported schemes had a capacity of 25.2 megawatts and exported 28,815 MWh to the grid (power consumed locally was not recorded). There are no wind-power generators in Canberra, but several have been built or are being built or planned in nearby New South Wales, such as the 140.7 megawatt Capital Wind Farm. The ACT government announced in 2013 that it was raising the target for electricity consumed in the ACT to be supplied from renewable sources to 90% by 2020, raising the target from 210 to 550 megawatts. It announced in February 2015 that three wind farms in Victoria and South Australia would supply 200 megawatts of capacity; these are expected to be operational by 2017. Contracts for the purchase of an additional 200 megawatts of power from two wind farms in South Australia and New South Wales were announced in December 2015 and March 2016. The ACT government announced in 2014 that up to 23 megawatts of feed-in-tariff entitlements would be made available for the establishment of a facility in the ACT or surrounding region for burning household and business waste to produce electricity by 2020. The ACT has the highest rate with internet access at home (94 per cent of households in 2014–15). Canberra has three sister cities: In addition, Canberra has the following friendship cities:
{{otheruses}} {{Use dmy dates|date=April 2020}} {{Infobox Australian place | type = city | name = Sydney | state = nsw | native_name = ''Gadi''<br>''Warrang'' | image = Sydney skyline from the north August 2016 (29009142591).jpg | caption = [[Sydney Harbour]] (Landmarks include [[Sydney Opera House]] and [[Sydney Harbour Bridge]].) | image_alt = | image2 = Free vector map of Sydney city Australia Level 12.svg | image2_alt = Map of the Sydney metropolitan area | caption2 = Map of the Sydney metropolitan area | coordinates = {{coord|33|51|54|S|151|12|34|E|display=inline,title}} | force_national_map = yes | pop = 5029768 | pop_year = 2016 | pop_footnotes =<ref name="ABSGCCSA">{{cite web|title=3218.0 – Regional Population Growth, Australia, 2015–16:ESTIMATED RESIDENT POPULATION – Australia's capital city populations, June 2016|url=http://www.abs.gov.au/ausstats/abs@.nsf/Latestproducts/3218.0Media%20Release12016?opendocument&tabname=Summary&prodno=3218.0&issue=2016&num=&view=|website=Australian Bureau of Statistics|publisher=[[Australian Bureau of Statistics]]|accessdate=July 29, 2017|archive-date=September 30, 2017|archive-url=https://web.archive.org/web/20170930073403/http://www.abs.gov.au/ausstats/abs@.nsf/Latestproducts/3218.0Media%20Release12016?opendocument&tabname=Summary&prodno=3218.0&issue=2016&num=&view=|url-status=dead}} Estimated resident population, June 30, 2016.</ref> | poprank = 1st | density = 400 | density_footnotes = (2015)<ref>{{cite web|url=http://abs.gov.au/ausstats/abs@.nsf/Latestproducts/3218.0Main%20Features202014-15?opendocument&tabname=Summary&prodno=3218.0&issue=2014-15&num=&view=|title=3218.0 – Regional Population Growth, Australia, 2014–15: New South Wales: Population Density|publisher=[[Australian Bureau of Statistics]]|date=March 30, 2016|accessdate=March 31, 2016|url-status=dead|archiveurl=https://web.archive.org/web/20160405011204/http://www.abs.gov.au/AUSSTATS/abs@.nsf/Latestproducts/3218.0Main%20Features202014-15?opendocument&tabname=Summary&prodno=3218.0&issue=2014-15&num=&view=|archivedate=April 5, 2016|df=dmy-all}}</ref> | established = [[Australia Day|January 26]] 1788 | area = 12367.7 | area_footnotes = (GCCSA)<ref name=ABSGCCSAXLS>{{cite web|url=http://www.censusdata.abs.gov.au/CensusOutput/copsub.NSF/All%20docs%20by%20catNo/2011~Community%20Profile~1GSYD/$File/BCP_1GSYD.zip?OpenElement|title=Greater Sydney: Basic Community Profile|publisher=[[Australian Bureau of Statistics]]|work=2011 Census Community Profiles|date=March 28, 2013|format=xls|accessdate=April 9, 2014|archive-date=November 7, 2022|archive-url=https://web.archive.org/web/20221107230439/https://www.censusdata.abs.gov.au/CensusOutput/copsub.NSF/All%20docs%20by%20catNo/2011~Community%20Profile~1GSYD/%24File/BCP_1GSYD.zip?OpenElement|url-status=dead}}</ref> | timezone = [[Australian Eastern Standard Time|AEST +10]] [[Australian Eastern Daylight Time|AEDT +11]] | dist1 = 877 | dir1 = NE | location1 = Melbourne | dist2 = 923 | dir2 = S | location2 = Brisbane | dist3 = 287 | dir3 = NE | location3 = Canberra | dist4 = 3936 | dir4 = E | location4 = [[Perth, Western Australia|Perth]] | dist5 = 1404 | dir5 = E | location5 = Adelaide | lga = [[Sydney#Government|Various]] | county = [[Cumberland County, New South Wales|Cumberland]]<ref name=gnbcounty>{{NSW GNR|id=JPYbwptLTR|title=Cumberland County|accessdate=September 20, 2017}}</ref> | division = | stategov = [[Electoral districts of New South Wales|Various]] | fedgov = [[:File:Sydney divisions overview 2010.png|various]] (24) | maxtemp = 22.8 | maxtemp_footnotes =<ref name=metdata>{{BoM Aust stats|site_ref=cw_066062_All|site_name=Sydney (Observatory Hill)|accessdate=December 15, 2016}}</ref> | mintemp = 14.7 | mintemp_footnotes =<ref name=metdata/> | rainfall = 1149.7 | rainfall_footnotes =<ref name=metdata/> | footnotes = '''Coordinates''':<ref>{{NSW GNR|id=TRlpoerXGH|title=Sydney|accessdate=February 28, 2015}}</ref> | local_map = yes | zoom = 7 }} [[File:City of Sydney Flag.svg|thumb|250px|Flag of Sydney]] [[File:Sydney - COA.svg|thumb|200px|Coat of arms of Sydney]] '''Sydney''' ({{lang-xdk|Gadi}}, also {{lang|xdk|Warrang}}) is a city on the east coast of [[Australia]] which is the [[capital city]] of [[New South Wales]]. About five and a half [[million]] people live in Sydney which makes it the biggest city in both [[Australia]] and [[Oceania]]. Sydney started in 1788, when the Captain [[Arthur Phillip]] brought the [[First Fleet]] to settle in Australia. The settlers were mostly [[convict]]s from crowded prisons in [[England]] and [[Ireland]], with a group of soldiers to guard them. The country is home to more than 200 spoken languages with a large population of overseas-born residents.<ref>{{Cite web|title=Sydney Information|url=https://www.besydney.com.au/incentive-events/why-sydney/|url-status=live|archive-url=https://web.archive.org/web/20160911092158/https://businesseventssydney.com.au/sydney-shines/global-city/|archive-date=September 11, 2016|website=Business Events Sydney|accessdate=August 30, 2016}}</ref> The first people to occupy the area were [[Indigenous Australians|Australian Aboriginals]].<ref>{{cite web | url=https://www.cityofsydney.nsw.gov.au/learn/archives-history/sydneys-history/aboriginal-history | title=Aboriginal history and the Gadigal people }}</ref> ==Attractions== In Sydney, there are many famous buildings: the [[Sydney Opera House]], the [[Queen Victoria Building]] and the [[Sydney Harbour Bridge]]. Sydney has a large [[harbour]] and many [[beaches]]. The most famous beach is [[Bondi Beach]], some other famous beaches are [[Coogee Beach]] and [[Manly Beach]]. A popular coastal walk to do is the Bondi to Coogee Coastal Walk.<ref>{{Cite web|url=http://www.visitnsw.com/things-to-do/nature-and-parks/walks/bondi-to-coogee-coastal-walk|title=Bondi to Coogee Coastal Walk|website=Destination NSW|accessdate=March 17, 2016}}</ref> Famous parts of the harbour are [[Darling Harbour]] and Circular Quay. The [[Royal National Park]] is a popular and famous park in southern Sydney. Sydney has many things to see and do. These include visiting [[Taronga Zoo]] (a park for animals) on the northern side of the harbour, eating food, looking at art or watching sports (like [[cricket]] or [[Rugby football|rugby]]) or walking down George Street which has many nice shops.<ref>{{Cite web|url=http://www.sydney.com/things-to-do/fashion-and-shopping/shopping-precincts|title=Sydney shopping precincts|website=Destination NSW|accessdate=March 17, 2016}}</ref> Sydney has some of Australia's best shopping, and the [[Blue Mountains (New South Wales)|Blue Mountains]] are about two hours' drive away to the west. Many people go to Sydney City to watch the great [[fireworks]] display over the harbour on [[New Year's Eve]]. Sydney holds events throughout the year that attract tourists from around the world, some namely events are Vivid Sydney, Royal Easter Show, and more.<ref>{{Cite web|title=Sydney Festivals and Events|url=https://www.besydney.com.au/|url-status=live|archive-url=https://web.archive.org/web/20160804182258/https://businesseventssydney.com.au/sydney-shines/explore-the-city/festivals-events/|archive-date=August 4, 2016|website=Business Events Sydney Festivals and Events|accessdate=August 30, 2016}}</ref> == Events == Sydney is home to a range of events that happen throughout the year, drawing many tourists and visitors to the city. One of Sydney's biggest event holding convention centre is the newly rebuilt [[International Convention Centre]] located by Darling Harbour, Sydney.<ref>{{Cite web|title=International Convention Centre|url=https://www.besydney.com.au/find-a-supplier/international-convention-centre-sydney/|url-status=live|archive-url=https://web.archive.org/web/20160304170446/https://businesseventssydney.com.au/supplier-search/results/international-convention-centre-sydney/|archive-date=March 4, 2016|website=Business Events Sydney|accessdate=August 30, 2016}}</ref> High-profile global events held in Sydney were: 2007 APEC Leaders Conference, 2008 World Youth Day and the 2014 G20 Finance Ministers and Central Bank Governors meeting.<ref>{{Cite web|title=Sydney Shines|url=https://www.besydney.com.au/|url-status=live|archive-url=https://web.archive.org/web/20170215054826/http://businesseventssydney.com.au/sydney-shines/meeting-city/|archive-date=February 15, 2017|website=Business Events Sydney|accessdate=February 9, 2017}}</ref> === Vivid Sydney === Vivid Sydney is held every year in May and June in the winter and is a display of light shows, art installations and music from local and international artists. In 2016, Vivid Sydney ran for 23 nights from May 27 to June 18.<ref>{{Cite web|url=http://www.sydney.com/destinations/sydney/sydney-city/vivid-sydney|title=Vivid Sydney Event|website=Sydney Tourism NSW|accessdate=March 31, 2016}}</ref> === Royal Easter Show === The Royal Easter Show is held in Sydney Olympic Park throughout the Easter public holiday period in March or April. It is a family attraction with rides, carnival games, showbags and food tents. There are also agriculture shows where children can learn more about farm animals as well as animal feeding areas. Shows such as woodcutting and animal competitions are popular every year.<ref>{{Cite web|url=http://www.sydney.com/events/sydney-royal-easter-show|title=Royal Easter Show Event|website=Sydney Tourism NSW|accessdate=March 31, 2016}}</ref> === Mercedes Benz Fashion Weekend === Mercedes Benz Fashion weekend runs for two days every year and features runway shows by leading Australian designers. The event is attended by fashion journalists, magazine editors, bloggers and fashion buyers. There are also Styling Sessions and VIP parties at the home of fashion week in Carriageworks in the inner city suburb of Redfern.<ref>{{Cite web|url=http://www.sydney.com/events/mercedesbenz-fashion-weekend|title=Mercedes Benz Fashion Weekend|website=Sydney Tourism NSW|accessdate=March 31, 2016|archive-date=March 26, 2016|archive-url=https://web.archive.org/web/20160326125325/http://www.sydney.com/events/mercedesbenz-fashion-weekend|url-status=dead}}</ref> == Geography == Sydney is a city located on the East Coast of Australia. It is currently the largest city in Australia, and the continent of Oceania. Sydney is known for having been the original landing spot for the British. It is situated next to the Blue Mountains. Sydney has more than 650 suburbs, and 38 local government areas. Sydney lies on a drowned coastline on the east coast of New South Wales, where the ocean level has risen up to flood deep river valleys carved in the Sydney [[Triassic]] [[sandstone]], which was laid down about about 200 million years ago. Sydney has over two major regions: the Cumberland Plain, a mostly flying to the west of Sydney Harbour, and the [[Hornsby]] Plateau, a tableland north you of the Harbour. Sydney's native plant species are mostly [[eucalyptus]] trees,<ref>{{cite web|url=http://www.agriculture.gov.au/abares/forestsaustralia/profiles/eucalypt-forest|title=Eucalypt forest|access-date=29 January 2017|publisher=Commonwealth of Australia|year=2017}}</ref> and its soils are red and yellow in colour. Sydney has nine rivers and just over a hundred creeks. ==Climate== Sydney has a [[humid subtropical climate]] with warm to hot summers and cool winters, and rainfall spread throughout the year. Hotter temperatures are recorded inland in the western suburbs, because the coast is moderated by the ocean. Rainfall is higher in the first half of the year when easterly winds are common, and lower in late winter/early spring when winds are more westerly, but rain has been very changeable in recent years.<ref>{{cite news|title=Sydney future: high temps, erratic rain| url =https://www.smh.com.au/national/sydney-future-high-temps-erratic-rain-20081028-5a7x.html| publisher = The Sydney Morning Herald}}</ref> East coast lows bring heavy rainfall to the Sydney area, typically in autumn to early winter.<ref>{{cite web|title=About East Coast Lows|url=http://www.bom.gov.au/nsw/sevwx/facts/ecl.shtml|publisher=Bureau of Meteorology|access-date=12 December 2022}}</ref> In the warm months, rain comes in short heavy falls in the afternoons, usually with a [[thunderstorm]]. Snow is unheard of, but major snowfall was last reported in Sydney on 28 June 1836. The [[El Niño–Southern Oscillation]] are an important feature in Sydney's weather patterns: [[drought]] and [[bushfire]] on the one hand, and storms and flooding on the other.<ref>[https://www.smh.com.au/environment/weather/the-beasts-to-our-east-what-are-el-ninos-and-la-ninas-20201209-p56lya.html#:~:text=Events%20usually%20last%20nine%20to,with%20flooding%20in%20eastern%20states. What is La Nina and what does it mean for your summer?] By Peter Hannam and Laura Chung. The Sydney Morning Herald. November 25, 2021.</ref> The highest recorded temperature in the city was {{convert|45.3|C}}, which was experienced on 18 January 2013. The highest recorded temperature in the west was {{convert|48.9|C}}, observed on 4 January 2020 in the suburb of Penrith.<ref>{{cite web |url=http://www.bom.gov.au/climate/averages/tables/cw_067113_All.shtml |title=Climate statistics for Penrith Lakes |website=Bureau of Meteorology}}</ref> Hot days are usually ended by a southerly buster, a windy sea breeze that comes from the southeast. While Sydney CBD (Observatory Hill) has never recorded [[frost]], the far western suburbs usually get frost in winter. The [[ultraviolet]] index rating in the summer averages at 12, but can reach as high as 13.<ref>{{cite news|url=https://www.9news.com.au/national/weather-uv-ratings-extreme-sydney-new-south-wales-forecast-skin-cancer-danger/d8a7c657-ce79-4907-bace-13c8ebf34a35 |access-date=25 March 2022|title=Australia's UV index peaks as thousands flock to beach unprotected }}</ref> Windspeed recorded in [[Sydney Airport]] averages at 24.3 km/h (15.0 mph), making Sydney the windiest capital city in Australia.<ref>[http://www.bom.gov.au/jsp/ncc/cdio/cvg/av?p_stn_num=066037&p_prim_element_index=0&p_comp_element_index=0&redraw=null&p_display_type=full_statistics_table&normals_years=1991-2020&tablesizebutt=normal SYDNEY AIRPORT AMO] Bureau of Meterology, Sydney Airport (1991-2020 normals). Retrieved 17 December 2020.</ref> {{Sydney weatherbox}} == Related pages == * [[First Fleet]] * [[Australia]] * [[Koala]] == References == {{wikivoyage}} {{Commons category|Sydney|Sydney Australia}} {{reflist}} {{New South Wales cities}} {{authority control}} [[Category:Sydney| ]] [[Category:Olympic cities]] [[Category:1788 establishments in Australia]]
Sydney is the capital city of the state of New South Wales, and the most populous city in Australia. Located on Australia's east coast, the metropolis surrounds Sydney Harbour and extends about 80 km (50 mi) from the Pacific Ocean in the east to the Blue Mountains in the west, and about 80 km (50 mi) from the Ku-ring-gai Chase National Park and the Hawkesbury River in the north and north-west, to the Royal National Park and Macarthur in the south and south-west. Greater Sydney consists of 658 suburbs, spread across 33 local government areas. Residents of the city are colloquially known as "Sydneysiders". The estimated population in June 2022 was 5,297,089; the city is home to approximately 66% of the state's population. The city's nicknames include the "Emerald City" and the "Harbour City". Aboriginal Australians have inhabited the Greater Sydney region for at least 30,000 years, and Aboriginal engravings and cultural sites are common. The traditional custodians of the land on which modern Sydney stands are the clans of the Darug, Dharawal and Eora peoples. During his first Pacific voyage in 1770, James Cook charted the eastern coast of Australia, making landfall at Botany Bay. In 1788, the First Fleet of convicts, led by Arthur Phillip, founded Sydney as a British penal colony, the first European settlement in Australia. After World War II, Sydney experienced mass migration and by 2021 over 40 per cent of the population was born overseas. Foreign countries of birth with the greatest representation are mainland China, India, the United Kingdom, Vietnam and the Philippines. Despite being one of the most expensive cities in the world, Sydney frequently ranks in the top ten most liveable cities. It is classified as an Alpha city by the Globalization and World Cities Research Network, indicating its influence in the region and throughout the world. Ranked eleventh in the world for economic opportunity, Sydney has an advanced market economy with strengths in finance, manufacturing and tourism. Established in 1850, the University of Sydney was Australia's first university and is regarded as one of the world's leading universities. Sydney has hosted major international sporting events such as the 2000 Summer Olympics. The city is among the top fifteen most-visited, with millions of tourists coming each year to see the city's landmarks. The city has over 1,000,000 ha (2,500,000 acres) of nature reserves and parks, and its notable natural features include Sydney Harbour and Royal National Park. The Sydney Harbour Bridge and the World Heritage-listed Sydney Opera House are major tourist attractions. Central Station is the hub of Sydney's suburban rail and light rail networks, with metro platforms under construction. The main passenger airport serving the city is Kingsford Smith Airport, one of the world's oldest continually operating airports. In 1788, Captain Arthur Phillip, the first governor of New South Wales, named the cove where the first British settlement was established Sydney Cove after Home Secretary Thomas Townshend, 1st Viscount Sydney. The cove was called Warrane by the Aboriginal inhabitants. Phillip considered naming the settlement Albion, but this name was never officially used. By 1790 Phillip and other officials were regularly calling the township Sydney. Sydney was declared a city in 1842. The Gadigal (Cadigal) clan, whose territory stretches along the southern shore of Port Jackson from South Head to Darling Harbour, are the traditional owners of the land on which the British settlement was initially established, and call their territory Gadi (Cadi). Aboriginal clan names within the Sydney region were often formed by adding the suffix "-gal" to a word denoting the name for their territory, a specific place in their territory, a food source, or totem. Greater Sydney covers the traditional lands of 28 known Aboriginal clans. The first people to inhabit the area now known as Sydney were Aboriginal Australians who had migrated from southeast Asia via northern Australia. Flaked pebbles found in Western Sydney's gravel sediments might indicate human occupation from 45,000 to 50,000 years ago, while radiocarbon dating has shown evidence of human activity in the region from around 30,000 years ago. Prior to the arrival of the British, there were 4,000 to 8,000 Aboriginal people in the greater Sydney region. The inhabitants subsisted on fishing, hunting, and gathering plants and shellfish. The diet of the coastal clans was more reliant on seafood whereas hinterland clans ate more forest animals and plants. The clans had distinctive equipment and weapons mostly made of stone, wood, plant materials, bone and shell. They also differed in their body decorations, hairstyles, songs and dances. Aboriginal clans had a rich ceremonial life, part of a belief system centring on ancestral, totemic and supernatural beings. People from different clans and language groups came together to participate in initiation and other ceremonies. These occasions fostered trade, marriages and clan alliances. The earliest British settlers recorded the word 'Eora' as an Aboriginal term meaning either 'people' or 'from this place'. The clans of the Sydney area occupied land with traditional boundaries. There is debate, however, about which group or nation these clans belonged to, and the extent of differences in language and rites. The major groups were the coastal Eora people, the Dharug (Darug) occupying the inland area from Parramatta to the Blue Mountains, and the Dharawal people south of Botany Bay. Darginung and Gundungurra languages were spoken on the fringes of the Sydney area. The first meeting between Aboriginals and British explorers occurred on 29 April 1770 when Lieutenant James Cook landed at Botany Bay (Kamay) and encountered the Gweagal clan. Two Gweagal men opposed the landing party and one was shot and wounded. Cook and his crew stayed at Botany Bay for a week, collecting water, timber, fodder and botanical specimens and exploring the surrounding area. Cook sought to establish relations with the Aboriginal population without success. Britain had been sending convicts to its American colonies for most of the eighteenth century, and the loss of these colonies in 1783 was the impetus to establish a penal colony at Botany Bay. Proponents of colonisation also pointed to the strategic importance of a new base in the Asia-Pacific region and its potential to provide much-needed timber and flax for the navy. The First Fleet of 11 ships under the command of Captain Arthur Phillip arrived in Botany Bay in January 1788. It comprised more than a thousand settlers, including 736 convicts. The fleet soon moved to the more suitable Port Jackson where a settlement was established at Sydney Cove on 26 January 1788. The colony of New South Wales was formally proclaimed by Governor Phillip on 7 February 1788. Sydney Cove offered a fresh water supply and a safe harbour, which Philip described as "the finest Harbour in the World ... Here a Thousand Sail of the Line may ride in the most perfect Security". The settlement was planned to be a self-sufficient penal colony based on subsistence agriculture. Trade and shipbuilding were banned in order to keep the convicts isolated. However, the soil around the settlement proved poor and the first crops failed, leading to several years of hunger and strict rationing. The food crisis was relieved with the arrival of the Second Fleet in mid-1790 and the Third Fleet in 1791. Former convicts received small grants of land, and government and private farms spread to the more fertile lands around Parramatta, Windsor and Camden on the Cumberland Plain. By 1804, the colony was self-sufficient in food. A smallpox epidemic in April 1789 killed about half the region's Indigenous population. In November 1790 Bennelong led a group of survivors of the Sydney clans into the settlement, establishing a continuous presence of Aboriginal Australians in settled Sydney. Phillip had been given no instructions for urban development, but in July 1788 submitted a plan for the new town at Sydney Cove. It included a wide central avenue, a permanent Government House, law courts, hospital and other public buildings, but no provision for warehouses, shops, or other commercial buildings. Phillip promptly ignored his own plan, and unplanned development became a feature of Sydney's topography. After Phillip's departure in December 1792, the colony's military officers began acquiring land and importing consumer goods from visiting ships. Former convicts engaged in trade and opened small businesses. Soldiers and former convicts built houses on Crown land, with or without official permission, in what was now commonly called Sydney town. Governor William Bligh (1806–08) imposed restrictions on commerce and ordered the demolition of buildings erected on Crown land, including some owned by past and serving military officers. The resulting conflict culminated in the Rum Rebellion of 1808, in which Bligh was deposed by the New South Wales Corps. Governor Lachlan Macquarie (1810–1821) played a leading role in the development of Sydney and New South Wales, establishing a bank, a currency and a hospital. He employed a planner to design the street layout of Sydney and commissioned the construction of roads, wharves, churches, and public buildings. Parramatta Road, linking Sydney and Parramatta, was opened in 1811, and a road across the Blue Mountains was completed in 1815, opening the way for large-scale farming and grazing west of the Great Dividing Range. Following the departure of Macquarie, official policy encouraged the emigration of free British settlers to New South Wales. Immigration to the colony increased from 900 free settlers in 1826–30 to 29,000 in 1836–40, many of whom settled in Sydney. By the 1840s Sydney exhibited a geographic divide between poor and working-class residents living west of the Tank Stream in areas such as The Rocks, and the more affluent residents living to its east. Free settlers, free-born residents and former convicts now represented the vast majority of the population of Sydney, leading to increasing public agitation for responsible government and an end to transportation. Transportation to New South Wales ceased in 1840. In 1804, Irish convicts led around 300 rebels in the Castle Hill Rebellion, an attempt to march on Sydney, commandeer a ship, and sail to freedom. Poorly armed, and with their leader Philip Cunningham captured, the main body of insurgents were routed by about 100 troops and volunteers at Rouse Hill. At least 39 convicts were killed in the uprising and subsequent executions. As the colony spread to the more fertile lands around the Hawkesbury River, north-west of Sydney, conflict between the settlers and the Darug people intensified, reaching a peak from 1794 to 1810. Bands of Darug people, led by Pemulwuy and later by his son Tedbury, burned crops, killed livestock and raided settler stores in a pattern of resistance that was to be repeated as the colonial frontier expanded. A military garrison was established on the Hawkesbury in 1795. The death toll from 1794 to 1800 was 26 settlers and up to 200 Darug. Conflict again erupted from 1814 to 1816 with the expansion of the colony into Dharawal country in the Nepean region south-west of Sydney. Following the deaths of several settlers, Governor Macquarie despatched three military detachments into Dharawal lands, culminating in the Appin massacre (April 1816) in which at least 14 Aboriginal people were killed. The New South Wales Legislative Council became a semi-elected body in 1842. Sydney was declared a city the same year, and a governing council established, elected on a restrictive property franchise. The discovery of gold in New South Wales and Victoria in 1851 initially caused economic disruption as men moved to the goldfields. Melbourne soon overtook Sydney as Australia's largest city, leading to an enduring rivalry between the two. However, increased immigration from overseas and wealth from gold exports increased demand for housing, consumer goods, services and urban amenities. The New South Wales government also stimulated growth by investing heavily in railways, trams, roads, ports, telegraph, schools and urban services. The population of Sydney and its suburbs grew from 95,600 in 1861 to 386,900 in 1891. The city developed many of its characteristic features. The growing population packed into rows of terrace houses in narrow streets. New public buildings of sandstone abounded, including at the University of Sydney (1854–61), the Australian Museum (1858–66), the Town Hall (1868–88), and the General Post Office (1866–92). Elaborate coffee palaces and hotels were erected. Daylight bathing at Sydney's beaches was banned, but segregated bathing at designated ocean baths was popular. Drought, the winding down of public works and a financial crisis led to economic depression in Sydney throughout most of the 1890s. Meanwhile, the Sydney-based premier of New South Wales, George Reid, became a key figure in the process of federation. When the six colonies federated on 1 January 1901, Sydney became the capital of the State of New South Wales. The spread of bubonic plague in 1900 prompted the state government to modernise the wharves and demolish inner-city slums. The outbreak of the First World War in 1914 saw more Sydney males volunteer for the armed forces than the Commonwealth authorities could process, and helped reduce unemployment. Those returning from the war in 1918 were promised "homes fit for heroes" in new suburbs such as Daceyville and Matraville. "Garden suburbs" and mixed industrial and residential developments also grew along the rail and tram corridors. The population reached one million in 1926, after Sydney had regained its position as the most populous city in Australia. The government created jobs with massive public projects such as the electrification of the Sydney rail network and building the Sydney Harbour Bridge. Sydney was more severely affected by the Great Depression of the 1930s than regional New South Wales or Melbourne. New building almost came to a standstill, and by 1933 the unemployment rate for male workers was 28 per cent, but over 40 per cent in working class areas such as Alexandria and Redfern. Many families were evicted from their homes and shanty towns grew along coastal Sydney and Botany Bay, the largest being "Happy Valley" at La Perouse. The Depression also exacerbated political divisions. In March 1932, when populist Labor premier Jack Lang attempted to open the Sydney Harbour Bridge he was upstaged by Francis de Groot of the far-right New Guard, who slashed the ribbon with a sabre. In January 1938, Sydney celebrated the Empire Games and the sesquicentenary of European settlement in Australia. One journalist wrote, "Golden beaches. Sun tanned men and maidens...Red-roofed villas terraced above the blue waters of the harbour...Even Melbourne seems like some grey and stately city of Northern Europe compared with Sydney's sub-tropical splendours." A congress of the "Aborigines of Australia" declared 26 January "A Day of Mourning" for "the whiteman's seizure of our country." With the outbreak of Second World War in 1939, Sydney experienced a surge in industrial development. Unemployment virtually disappeared and women moved into jobs previously typically reserved for males. Sydney was attacked by Japanese submarines in May and June 1942 with 21 killed. Households built air raid shelters and performed drills. Military establishments in response to World War II in Australia included the Garden Island Tunnel System, the only tunnel warfare complex in Sydney, and the heritage-listed military fortification systems Bradleys Head Fortification Complex and Middle Head Fortifications, which were part of a total defence system for Sydney Harbour. A post-war immigration and baby boom saw a rapid increase in Sydney's population and the spread of low-density housing in suburbs throughout the Cumberland Plain. Immigrants—mostly from Britain and continental Europe—and their children accounted for over three-quarters of Sydney's population growth between 1947 and 1971. The newly created Cumberland County Council oversaw low-density residential developments, the largest at Green Valley and Mount Druitt. Older residential centres such as Parramatta, Bankstown and Liverpool became suburbs of the metropolis. Manufacturing, protected by high tariffs, employed over a third of the workforce from 1945 to the 1960s. However, as the long post-war economic boom progressed, retail and other service industries became the main source of new jobs. An estimated one million onlookers, most of the city's population, watched Queen Elizabeth II land in 1954 at Farm Cove where Captain Phillip had raised the Union Jack 165 years earlier, commencing her Australian Royal Tour. It was the first time a reigning monarch stepped onto Australian soil. Increasing high-rise development in Sydney and the expansion of suburbs beyond the "green belt" envisaged by the planners of the 1950s resulted in community protests. In the early 1970s, trade unions and resident action groups imposed green bans on development projects in historic areas such as The Rocks. Federal, State and local governments introduced heritage and environmental legislation. The Sydney Opera House was also controversial for its cost and disputes between architect Jørn Utzon and government officials. However, soon after it opened in 1973 it became a major tourist attraction and symbol of the city. The progressive reduction in tariff protection from 1974 began the transformation of Sydney from a manufacturing centre to a "world city". From the 1980s, overseas immigration grew rapidly, with Asia, the Middle East and Africa becoming major sources. By 2021, the population of Sydney was over 5.2 million, with 40% of the population born overseas. China and India overtook England as the largest source countries for overseas-born residents. Sydney is a coastal basin with the Tasman Sea to the east, the Blue Mountains to the west, the Hawkesbury River to the north, and the Woronora Plateau to the south. Sydney spans two geographic regions. The Cumberland Plain lies to the south and west of the Harbour and is relatively flat. The Hornsby Plateau is located to the north and is dissected by steep valleys. The flat areas of the south were the first to be developed; it was not until the construction of the Sydney Harbour Bridge that the northern reaches became more heavily populated. Seventy surf beaches can be found along its coastline, with Bondi Beach being the most famous. The Nepean River wraps around the western edge of the city and becomes the Hawkesbury River before reaching Broken Bay. Most of Sydney's water storages can be found on tributaries of the Nepean River. The Parramatta River is mostly industrial and drains a large area of Sydney's western suburbs into Port Jackson. The southern parts of the city are drained by the Georges River and the Cooks River into Botany Bay. There is no single definition of the boundaries of Sydney. The Australian Statistical Geography Standard definition of Greater Sydney covers 12,369 km (4,776 sq mi) and includes the local government areas of Central Coast in the north, Hawkesbury in the north-west, Blue Mountains in the west, Sutherland Shire in the south, and Wollondilly in the south-west. The local government area of the City of Sydney covers about 26 square kilometres from Garden island in the east to Bicentennial Park in the west, and south to the suburbs of Alexandria and Rosebery. Sydney is made up of mostly Triassic rock with some recent igneous dykes and volcanic necks (typically found in the Prospect dolerite intrusion, west of Sydney). The Sydney Basin was formed in the early Triassic period. The sand that was to become the sandstone of today was laid down between 360 and 200 million years ago. The sandstone has shale lenses and fossil riverbeds. The Sydney Basin bioregion includes coastal features of cliffs, beaches, and estuaries. Deep river valleys known as rias were carved during the Triassic period in the Hawkesbury sandstone of the coastal region. The rising sea level between 18,000 and 6,000 years ago flooded the rias to form estuaries and deep harbours. Port Jackson, better known as Sydney Harbour, is one such ria. Sydney features two major soil types: sandy soils (which originate from the Hawkesbury sandstone) and clay (which are from shales and volcanic rocks), though some soils may be a mixture of the two. Directly overlying the older Hawkesbury sandstone is the Wianamatta shale, a geological feature found in western Sydney that was deposited in connection with a large river delta during the Middle Triassic. The Wianamatta shale generally comprises fine grained sedimentary rocks such as shales, mudstones, ironstones, siltstones and laminites, with less common sandstone units. The Wianamatta Group is made up of Bringelly Shale, Minchinbury Sandstone and Ashfield Shale. The most prevalent plant communities in the Sydney region are grassy woodlands (i.e. savannas) and some pockets of dry sclerophyll forests, which consist of eucalyptus trees, casuarinas, melaleucas, corymbias and angophoras, with shrubs (typically wattles, callistemons, grevilleas and banksias), and a semi-continuous grass in the understory. The plants in this community tend to have rough, spiky leaves due to low soil fertility. Sydney also features a few areas of wet sclerophyll forests in the wetter, elevated areas in the north and northeast. These forests are defined by straight, tall tree canopies with a moist understory of soft-leaved shrubs, tree ferns and herbs. The predominant vegetation community in Sydney is the Cumberland Plain Woodland in Western Sydney (Cumberland Plain), followed by the Sydney Turpentine-Ironbark Forest in the Inner West and Northern Sydney, the Eastern Suburbs Banksia Scrub in the coastline and the Blue Gum High Forest scantily present in the North Shore – all of which are critically endangered. The city also includes the Sydney Sandstone Ridgetop Woodland found in Ku-ring-gai Chase National Park on the Hornsby Plateau to the north. Sydney is home to dozens of bird species, which commonly include the Australian raven, Australian magpie, crested pigeon, noisy miner and the pied currawong. Introduced bird species ubiquitously found in Sydney are the common myna, common starling, house sparrow and the spotted dove. Reptile species are also numerous and predominantly include skinks. Sydney has a few mammal and spider species, such as the grey-headed flying fox and the Sydney funnel-web, respectively, and has a huge diversity of marine species inhabiting its harbour and beaches. Under the Köppen–Geiger classification, Sydney has a humid subtropical climate (Cfa) with "warm, sometimes hot" summers and "generally mild", to "cool" winters. The El Niño–Southern Oscillation, the Indian Ocean Dipole and the Southern Annular Mode play an important role in determining Sydney's weather patterns: drought and bushfire on the one hand, and storms and flooding on the other, associated with the opposite phases of the oscillation in Australia. The weather is moderated by proximity to the ocean, and more extreme temperatures are recorded in the inland western suburbs. At Sydney's primary weather station at Observatory Hill, extreme temperatures have ranged from 45.8 °C (114.4 °F) on 18 January 2013 to 2.1 °C (35.8 °F) on 22 June 1932. An average of 14.9 days a year have temperatures at or above 30 °C (86 °F) in the central business district (CBD). In contrast, the metropolitan area averages between 35 and 65 days, depending on the suburb. The hottest day in the metropolitan area occurred in Penrith on 4 January 2020, where a high of 48.9 °C (120.0 °F) was recorded. The average annual temperature of the sea ranges from 18.5 °C (65.3 °F) in September to 23.7 °C (74.7 °F) in February. Sydney has an average of 7.2 hours of sunshine per day and 109.5 clear days annually. Due to the inland location, frost is recorded early in the morning in Western Sydney a few times in winter. Autumn and spring are the transitional seasons, with spring showing a larger temperature variation than autumn. Sydney experiences an urban heat island effect. This makes certain parts of the city more vulnerable to extreme heat, including coastal suburbs. In late spring and summer, temperatures over 35 °C (95 °F) are not uncommon, though hot, dry conditions are usually ended by a southerly buster, a powerful southerly that brings gale winds and a rapid fall in temperature. Since Sydney is downwind of the Great Dividing Range, it occasionally experiences dry, westerly foehn winds typically in winter and early spring (which are the reason for its warm maximum temperatures). Westerly winds are intense when the Roaring Forties (or the Southern Annular Mode) shift towards southeastern Australia, where they may damage homes and affect flights, in addition to making the temperature seem colder than it actually is. Rainfall has a moderate to low variability and has historically been fairly uniform throughout the year, although in recent years it has been more summer-dominant and erratic. Precipitation is usually higher in late summer through to early winter and lower in late winter to early spring. In late autumn and winter, east coast lows may bring large amounts of rainfall, especially in the CBD. In the warm season black nor'easters are usually the cause of heavy rain events, though other forms of low-pressure areas, including remnants of ex-cyclones, may also bring heavy deluge and afternoon thunderstorms. Snowfall was last reported in 1836, though a fall of graupel, or soft hail, in Lindfield, Roseville and Killara was mistaken by many for snow, in July 2008. In 2009, dry conditions brought a severe dust storm towards the city. The Greater Sydney Commission divides Sydney into three "cities" and five "districts" based on the 33 LGAs in the metropolitan area. The "metropolis of three cities" comprises Eastern Harbour City, Central River City and Western Parkland City. The Australian Bureau of Statistics also includes City of Central Coast (the former Gosford City and Wyong Shire) as part of Greater Sydney for population counts, adding 330,000 people. The CBD extends about 3 km (1.9 mi) south from Sydney Cove. It is bordered by Farm Cove within the Royal Botanic Garden to the east and Darling Harbour to the west. Suburbs surrounding the CBD include Woolloomooloo and Potts Point to the east, Surry Hills and Darlinghurst to the south, Pyrmont and Ultimo to the west, and Millers Point and The Rocks to the north. Most of these suburbs measure less than 1 km (0.4 sq mi) in area. The Sydney CBD is characterised by narrow streets and thoroughfares, created in its convict beginnings. Several localities, distinct from suburbs, exist throughout Sydney's inner reaches. Central and Circular Quay are transport hubs with ferry, rail, and bus interchanges. Chinatown, Darling Harbour, and Kings Cross are important locations for culture, tourism, and recreation. the Strand Arcade, located between Pitt Street Mall and George Street, is a historical Victorian-style shopping arcade. Opened on 1 April 1892, its shop fronts are an exact replica of the original internal shopping facades. Westfield Sydney, located beneath the Sydney Tower, is the largest shopping centre by area in Sydney. Since the late 20th century, there has been a trend of gentrification amongst Sydney's inner suburbs. Pyrmont, located on the harbour, was redeveloped from a centre of shipping and international trade to an area of high density housing, tourist accommodation, and gambling. Originally located well outside of the city, Darlinghurst is the location of the historic Darlinghurst Gaol, manufacturing, and mixed housing. For a period it was known as an area of prostitution. The terrace-style housing has largely been retained and Darlinghurst has undergone significant gentrification since the 1980s. Green Square is a former industrial area of Waterloo which is undergoing urban renewal worth $8 billion. On the city harbour edge, the historic suburb and wharves of Millers Point are being built up as the new area of Barangaroo. The suburb of Paddington is known for its restored terrace houses, Victoria Barracks, and shopping including the weekly Oxford Street markets. The Inner West generally includes the Inner West Council, Municipality of Burwood, Municipality of Strathfield, and City of Canada Bay. These span up to about 11 km west of the CBD. Historically, especially prior to the building of the Harbour Bridge, the outer suburbs of the Inner West such as Strathfield were the location of "country" estates for the colony's elites. By contrast, the inner suburbs in the Inner West, being close to transport and industry, have historically housed working-class industrial workers. These areas have undergone gentrification in the late 20th century, and many parts are now highly valued residential suburbs. As of 2021, an Inner West suburb (Strathfield) remained one of the 20 most expensive postcodes in Australia by median house price (the others were all in metropolitan Sydney, all in Northern Sydney or the Eastern Suburbs). The University of Sydney is located in this area, as well as the University of Technology, Sydney and a campus of the Australian Catholic University. The Anzac Bridge spans Johnstons Bay and connects Rozelle to Pyrmont and the city, forming part of the Western Distributor. The Inner West is today well known as the location of village commercial centres with cosmopolitan flavours, such as the "Little Italy" commercial centres of Leichardt, Five Dock and Haberfield, "Little Portugal" in Petersham, "Little Korea" in Strathfield or "Little Shanghai" in Ashfield. Large-scale shopping centres in the area include Westfield Burwood, DFO Homebush and Birkenhead Point Outlet Centre. There is a large cosmopolitan community and nightlife hub on King Street in Newtown. The area is serviced by the T1, T2, and T3 railway lines, including the Main Suburban Line, which was the first to be constructed in New South Wales. Strathfield railway station is a secondary railway hub within Sydney, and major station on the Suburban and Northern lines. It was constructed in 1876. The future Sydney Metro West will also connect this area with the City and Parramatta. The area is also serviced by the Parramatta River services of Sydney Ferries, numerous bus routes and cycleways. The Eastern Suburbs encompass the Municipality of Woollahra, the City of Randwick, the Waverley Municipal Council, and parts of the Bayside Council. They include some of the most affluent and advantaged areas in the country, with some streets being amongst the most expensive in the world. As at 2014, Wolseley Road, Point Piper, had a top price of $20,900 per square metre, making it the ninth-most expensive street in the world. More than 75% of neighbourhoods in the Electoral District of Wentworth fall under the top decile of SEIFA advantage, making it the least disadvantaged area in the country. As of 2021, of the 20 most expensive postcodes in Australia by median house price, nine were in the Eastern Suburbs. Major landmarks include Bondi Beach, which was added to the Australian National Heritage List in 2008; and Bondi Junction, featuring a Westfield shopping centre and an estimated office workforce of 6,400 by 2035, as well as a railway station on the T4 Eastern Suburbs Line. The suburb of Randwick contains Randwick Racecourse, the Royal Hospital for Women, the Prince of Wales Hospital, Sydney Children's Hospital, and University of New South Wales Kensington Campus. Construction of the CBD and South East Light Rail was completed in April 2020. The project aims to provide reliable and high-capacity tram services to residents in the City and South-East. Major shopping centres in the area include Westfield Bondi Junction and Westfield Eastgardens. Southern Sydney includes the suburbs in the local government areas of former Rockdale, Georges River Council (collectively known as St George), and broadly also includes the suburbs in the local government area of Sutherland, south of the Georges River (colloquially known as 'The Shire'). The Kurnell peninsula, near Botany Bay, is the site of the first landfall on the eastern coastline made by James Cook in 1770. La Perouse, a historic suburb named after the French navigator Jean-François de Galaup, comte de Lapérouse, is notable for its old military outpost at Bare Island and the Botany Bay National Park. The suburb of Cronulla in southern Sydney is close to Royal National Park, Australia's oldest national park. Hurstville, a large suburb with commercial and high-rise residential buildings dominating the skyline, has become a CBD for the southern suburbs. 'Northern Sydney' may also include the suburbs in the Upper North Shore, Lower North Shore and the Northern Beaches. The Northern Suburbs include several landmarks – Macquarie University, Gladesville Bridge, Ryde Bridge, Macquarie Centre and Curzon Hall in Marsfield. This area includes suburbs in the local government areas of Hornsby Shire, City of Ryde, the Municipality of Hunter's Hill and parts of the City of Parramatta. The North Shore includes the commercial centres of North Sydney and Chatswood. North Sydney itself consists of a large commercial centre, which contains the second largest concentration of high-rise buildings in Sydney after the CBD. North Sydney is dominated by advertising, marketing and associated trades, with many large corporations holding offices. The Northern Beaches area includes Manly, one of Sydney's most popular holiday destinations for much of the nineteenth and twentieth centuries. The region also features Sydney Heads, a series of headlands which form the entrance to Sydney Harbour. The Northern Beaches area extends south to the entrance of Port Jackson (Sydney Harbour), west to Middle Harbour and north to the entrance of Broken Bay. The 2011 Australian census found the Northern Beaches to be the most white and mono-ethnic district in Australia, contrasting with its more-diverse neighbours, the North Shore and the Central Coast. As of the end of 2021, half of the 20 most expensive postcodes in Australia (by median house price) were in Northern Sydney, including four on the Northern Beaches, two on the Lower North Shore, three on the Upper North Shore, and one straddling Hunters Hill and Woolwich. The Hills district generally refers to the suburbs in north-western Sydney including the local government areas of The Hills Shire, parts of the City of Parramatta Council and Hornsby Shire. Actual suburbs and localities that are considered to be in the Hills District can be somewhat amorphous. For example, the Hills District Historical Society restricts its definition to the Hills Shire local government area, yet its study area extends from Parramatta to the Hawkesbury. The region is so named for its characteristically comparatively hilly topography as the Cumberland Plain lifts up, joining the Hornsby Plateau. Windsor and Old Windsor Roads are the second and third roads, respectively, laid in Australia. The greater western suburbs encompasses the areas of Parramatta, the sixth largest business district in Australia, settled the same year as the harbour-side colony, Bankstown, Liverpool, Penrith, and Fairfield. Covering 5,800 km (2,200 sq mi) and having an estimated population as at 2017 of 2,288,554, western Sydney has the most multicultural suburbs in the country. The population is predominantly of a working class background, with major employment in the heavy industries and vocational trade. Toongabbie is noted for being the third mainland settlement (after Sydney and Parramatta) set up after British colonisation began in 1788, although the site of the settlement is actually in the separate suburb of Old Toongabbie. The western suburb of Prospect, in the City of Blacktown, is home to Raging Waters, a water park operated by Parques Reunidos. Auburn Botanic Gardens, a botanical garden in Auburn, attracts thousands of visitors each year, including many from outside Australia. The greater west also includes Sydney Olympic Park, a suburb created to host the 2000 Summer Olympics, and Sydney Motorsport Park, a circuit in Eastern Creek. Prospect Hill, a historically significant ridge in the west and the only area in Sydney with ancient volcanic activity, is also listed on the State Heritage Register. To the northwest, Featherdale Wildlife Park, a zoo in Doonside, near Blacktown, is a major tourist attraction. Sydney Zoo, opened in 2019, is another prominent zoo situated in Bungaribee. Established in 1799, the Old Government House, a historic house museum and tourist spot in Parramatta, was included in the Australian National Heritage List on 1 August 2007 and World Heritage List in 2010 (as part of the 11 penal sites constituting the Australian Convict Sites), making it the only site in greater western Sydney to be featured in such lists. The house is Australia's oldest surviving public building. Further to the southwest is the region of Macarthur and the city of Campbelltown, a significant population centre until the 1990s considered a region separate to Sydney proper. Macarthur Square, a shopping complex in Campbelltown, has become one of the largest shopping complexes in Sydney. The southwest also features Bankstown Reservoir, the oldest elevated reservoir constructed in reinforced concrete that is still in use and is listed on the State Heritage Register. The southwest is home to one of Sydney's oldest trees, the Bland Oak, which was planted in the 1840s by William Bland in Carramar. The earliest structures in the colony were built to the bare minimum of standards. Governor Macquarie set ambitious targets for the design of new construction projects. The city now has a world heritage listed building, several national heritage listed buildings, and dozens of Commonwealth heritage listed buildings as evidence of the survival of Macquarie's ideals. In 1814 the Governor called on a convict named Francis Greenway to design Macquarie Lighthouse. The lighthouse's Classical design earned Greenway a pardon from Macquarie in 1818 and introduced a culture of refined architecture that remains to this day. Greenway went on to design the Hyde Park Barracks in 1819 and the Georgian style St James's Church in 1824. Gothic-inspired architecture became more popular from the 1830s. John Verge's Elizabeth Bay House and St Philip's Church of 1856 were built in Gothic Revival style along with Edward Blore's Government House of 1845. Kirribilli House, completed in 1858, and St Andrew's Cathedral, Australia's oldest cathedral, are rare examples of Victorian Gothic construction. From the late 1850s there was a shift towards Classical architecture. Mortimer Lewis designed the Australian Museum in 1857. The General Post Office, completed in 1891 in Victorian Free Classical style, was designed by James Barnet. Barnet also oversaw the 1883 reconstruction of Greenway's Macquarie Lighthouse. Customs House was built in 1844. The neo-Classical and French Second Empire style Town Hall was completed in 1889. Romanesque designs gained favour from the early 1890s. Sydney Technical College was completed in 1893 using both Romanesque Revival and Queen Anne approaches. The Queen Victoria Building was designed in Romanesque Revival fashion by George McRae; completed in 1898, it accommodates 200 shops across its three storeys. As the wealth of the settlement increased and Sydney developed into a metropolis after Federation in 1901, its buildings became taller. Sydney's first tower was Culwulla Chambers which topped out at 50 m (160 ft) making 12 floors. The Commercial Traveller's Club, built in 1908, was of similar height at 10 floors. It was built in a brick stone veneer and demolished in 1972. This heralded a change in Sydney's cityscape and with the lifting of height restrictions in the 1960s there came a surge of high-rise construction. The Great Depression had a tangible influence on Sydney's architecture. New structures became more restrained with far less ornamentation. The most notable architectural feat of this period is the Harbour Bridge. Its steel arch was designed by John Bradfield and completed in 1932. A total of 39,000 tonnes of structural steel span the 503 m (1,650 ft) between Milsons Point and Dawes Point. Modern and International architecture came to Sydney from the 1940s. Since its completion in 1973 the city's Opera House has become a World Heritage Site and one of the world's most renowned pieces of Modern design. Jørn Utzon was awarded the Pritzker Prize in 2003 for his work on the Opera House. Sydney is home to Australia's first building by renowned Canadian-American architect Frank Gehry, the Dr Chau Chak Wing Building (2015). An entrance from The Goods Line–a pedestrian pathway and former railway line–is located on the eastern border of the site. Contemporary buildings in the CBD include Citigroup Centre, Aurora Place, Chifley Tower, the Reserve Bank building, Deutsche Bank Place, MLC Centre, and Capita Centre. The tallest structure is Sydney Tower, designed by Donald Crone and completed in 1981. Due to the proximity of Sydney Airport, a maximum height restriction was imposed, now sitting at 330 metres (1083 feet). Green bans and heritage overlays have been in place since at least 1977 to protect Sydney's heritage after controversial demolitions in the 1970s. Sydney surpasses both New York City and Paris real estate prices, having some of the most expensive in the world. The city remains Australia's most expensive housing market, with the mean house price at $1,142,212 as of December 2019 (over 25% higher the national mean house price). It is only second to Hong Kong with the average property costing 14 times the annual Sydney salary as of December 2016. There were 1.76 million dwellings in Sydney in 2016 including 925,000 (57%) detached houses, 227,000 (14%) semi-detached terrace houses and 456,000 (28%) units and apartments. Whilst terrace houses are common in the inner city areas, detached houses dominate the landscape in the outer suburbs. Due to environmental and economic pressures, there has been a noted trend towards denser housing, with a 30% increase in the number of apartments between 1996 and 2006. Public housing in Sydney is managed by the Government of New South Wales. Suburbs with large concentrations of public housing include Claymore, Macquarie Fields, Waterloo, and Mount Druitt. A range of heritage housing styles can be found throughout Sydney. Terrace houses are found in the inner suburbs such as Paddington, The Rocks, Potts Point and Balmain–many of which have been the subject of gentrification. These terraces, particularly those in suburbs such as The Rocks, were historically home to Sydney's miners and labourers. In the present day, terrace houses now make up some of the most valuable real estate in the city. Federation homes, constructed around the time of Federation in 1901, are located in suburbs such as Penshurst, Turramurra, and in Haberfield. Workers cottages are found in Surry Hills, Redfern, and Balmain. California bungalows are common in Ashfield, Concord, and Beecroft. Larger modern homes are predominantly found in the outer suburbs, such as Stanhope Gardens, Kellyville Ridge, Bella Vista to the northwest, Bossley Park, Abbotsbury, and Cecil Hills to the west, and Hoxton Park, Harrington Park, and Oran Park to the southwest. The Anzac War Memorial in Hyde Park is a public memorial dedicated to the Australian Imperial Force of World War I. The Royal Botanic Garden is the most iconic green space in the region, hosting both scientific and leisure activities. There are 15 separate parks under the City administration. Parks within the city centre include Hyde Park, The Domain and Prince Alfred Park. The Centennial Parklands is the largest park in the City of Sydney, comprising 189 ha (470 acres). The inner suburbs include Centennial Park and Moore Park in the east (both within the City of Sydney local government area), while the outer suburbs contain Sydney Park and Royal National Park in the south, Ku-ring-gai Chase National Park in the north, and Western Sydney Parklands in the west, which is one of the largest urban parks in the world. The Royal National Park was proclaimed in 1879 and with 13,200 ha (51 sq mi) is the second oldest national park in the world. Hyde Park is the oldest parkland in the country. The largest park in the Sydney metropolitan area is Ku-ring-gai Chase National Park, established in 1894 with an area of 15,400 ha (59 sq mi). It is regarded for its well-preserved records of indigenous habitation – more than 800 rock engravings, cave drawings and middens. The area now known as The Domain was set aside by Governor Arthur Phillip in 1788 as his private reserve. Under the orders of Macquarie the land to the immediate north of The Domain became the Royal Botanic Garden in 1816. This makes them the oldest botanic garden in Australia. The Gardens host scientific research with herbarium collections, a library and laboratories. The two parks have a total area of 64 ha (0.2 sq mi) with 8,900 individual plant species and receive over 3.5 million annual visits. To the south of The Domain is Hyde Park, the oldest public parkland in Australia which measures 16.2 ha (0.1 sq mi). Its location was used for both relaxation and grazing of animals from the earliest days of the colony. Macquarie dedicated it in 1810 for the "recreation and amusement of the inhabitants of the town" and named it in honour of Hyde Park in London. Researchers from Loughborough University have ranked Sydney amongst the top ten world cities that are highly integrated into the global economy. The Global Economic Power Index ranks Sydney eleventh in the world. The Global Cities Index recognises it as fourteenth in the world based on global engagement. There is a significant concentration of foreign banks and multinational corporations in Sydney and the city is promoted as Australia's financial capital and one of Asia Pacific's leading financial hubs. The prevailing economic theory during early colonial days was mercantilism, as it was throughout most of Western Europe. The economy struggled at first due to difficulties in cultivating the land and the lack of a stable monetary system. Governor Macquarie created two coins from every Spanish silver dollar in circulation. The economy was capitalist in nature by the 1840s as the proportion of free settlers increased, the maritime and wool industries flourished, and the powers of the East India Company were curtailed. Wheat, gold, and other minerals became export industries towards the end of the 1800s. Significant capital began to flow into the city from the 1870s to finance roads, railways, bridges, docks, courthouses, schools and hospitals. Protectionist policies after federation allowed for the creation of a manufacturing industry which became the city's largest employer by the 1920s. These same policies helped to relieve the effects of the Great Depression during which the unemployment rate in New South Wales reached as high as 32%. From the 1960s onwards Parramatta gained recognition as the city's second CBD and finance and tourism became major industries and sources of employment. Sydney's nominal gross domestic product was AU$400.9 billion and AU$80,000 per capita in 2015. Its gross domestic product was AU$337 billion in 2013, the largest in Australia. The Financial and Insurance Services industry accounts for 18.1% of gross product, ahead of Professional Services with 9% and Manufacturing with 7.2%. The Creative and Technology sectors are also focus industries for the City of Sydney and represented 9% and 11% of its economic output in 2012. There were 451,000 businesses based in Sydney in 2011, including 48% of the top 500 companies in Australia and two-thirds of the regional headquarters of multinational corporations. Global companies are attracted to the city in part because its time zone spans the closing of business in North America and the opening of business in Europe. Most foreign companies in Sydney maintain significant sales and service functions but comparably less production, research, and development capabilities. There are 283 multinational companies with regional offices in Sydney. Sydney has been ranked between the fifteenth and the fifth most expensive city in the world and is the most expensive city in Australia. Of the 15 categories only measured by UBS in 2012, workers receive the seventh highest wage levels of 77 cities in the world. Working residents of Sydney work an average of 1,846 hours per annum with 15 days of leave. The labour force of Greater Sydney Region in 2016 was 2,272,722 with a participation rate of 61.6%. It comprised 61.2% full-time workers, 30.9% part-time workers, and 6.0% unemployed individuals. The largest reported occupations are professionals, clerical and administrative workers, managers, technicians and trades workers, and community and personal service workers. The largest industries by employment across Greater Sydney are Health Care and Social Assistance (11.6%), Professional Services (9.8%), Retail Trade (9.3%), Construction (8.2%), Education and Training (8.0%), Accommodation and Food Services (6.7%), and Financial and Insurance Services (6.6%). The Professional Services and Financial and Insurance Services industries account for 25.4% of employment within the City of Sydney. In 2016, 57.6% of working-age residents had a weekly income of less than $1,000 and 14.4% had a weekly income of $1,750 or more. The median weekly income for the same period was $719 for individuals, $1,988 for families, and $1,750 for households. Unemployment in the City of Sydney averaged 4.6% for the decade to 2013, much lower than the current rate of unemployment in Western Sydney of 7.3%. Western Sydney continues to struggle to create jobs to meet its population growth despite the development of commercial centres like Parramatta. Each day about 200,000 commuters travel from Western Sydney to the CBD and suburbs in the east and north of the city. Home ownership in Sydney was less common than renting prior to the Second World War but this trend has since reversed. Median house prices have increased by an average of 8.6% per annum since 1970. The median house price in March 2014 was $630,000. The primary cause of rising prices is the increasing cost of land and scarcity. 31.6% of dwellings in Sydney are rented, 30.4% are owned outright and 34.8% are owned with a mortgage. 11.8% of mortgagees in 2011 had monthly loan repayments of less than $1,000 and 82.9% had monthly repayments of $1,000 or more. 44.9% of renters for the same period had weekly rent of less than $350 whilst 51.7% had weekly rent of $350 or more. The median weekly rent in Sydney in 2011 was $450. Macquarie gave a charter in 1817 to form the first bank in Australia, the Bank of New South Wales. New private banks opened throughout the 1800s but the financial system was unstable. Bank collapses were frequent and a crisis point was reached in 1893 when 12 banks failed. The Bank of New South Wales exists to this day as Westpac. The Commonwealth Bank of Australia was formed in Sydney in 1911 and began to issue notes backed by the resources of the nation. It was replaced in this role in 1959 by the Reserve Bank of Australia, also based in Sydney. The Australian Securities Exchange began operating in 1987 and with a market capitalisation of $1.6 trillion is now one of the ten largest exchanges in the world. The Financial and Insurance Services industry now constitutes 43% of the economic product of the City of Sydney. Sydney makes up half of Australia's finance sector and has been promoted by consecutive Commonwealth Governments as Asia Pacific's leading financial centre. In the 2017 Global Financial Centres Index, Sydney was ranked as having the eighth most competitive financial centre in the world. In 1985 the Federal Government granted 16 banking licences to foreign banks and now 40 of the 43 foreign banks operating in Australia are based in Sydney, including the People's Bank of China, Bank of America, Citigroup, UBS, Mizuho Bank, Bank of China, Banco Santander, Credit Suisse, Standard Chartered, State Street, HSBC, Deutsche Bank, Barclays, Royal Bank of Canada, Société Générale, Royal Bank of Scotland, Sumitomo Mitsui, ING Group, BNP Paribas, and Investec. Sydney has been a manufacturing city since the 1920s. By 1961 the industry accounted for 39% of all employment and by 1970 over 30% of all Australian manufacturing jobs were in Sydney. Its status has declined in recent decades, making up 12.6% of employment in 2001 and 8.5% in 2011. Between 1970 and 1985 there was a loss of 180,000 manufacturing jobs. Despite this, Sydney still overtook Melbourne as the largest manufacturing centre in Australia in the 2010s, with a manufacturing output of $21.7 billion in 2013. Observers have credited Sydney's focus on the domestic market and high-tech manufacturing for its resilience against the high Australian dollar of the early 2010s. The Smithfield-Wetherill Park Industrial Estate in Western Sydney is the largest industrial estate in the Southern Hemisphere and is the centre of manufacturing and distribution in the region. Sydney is a gateway to Australia for many international visitors and ranks among the top sixty most visited cities in the world. It has hosted over 2.8 million international visitors in 2013, or nearly half of all international visits to Australia. These visitors spent 59 million nights in the city and a total of $5.9 billion. The countries of origin in descending order were China, New Zealand, the United Kingdom, the United States, South Korea, Japan, Singapore, Germany, Hong Kong, and India. The city also received 8.3 million domestic overnight visitors in 2013 who spent a total of $6 billion. 26,700 workers in the City of Sydney were directly employed by tourism in 2011. There were 480,000 visitors and 27,500 people staying overnight each day in 2012. On average, the tourism industry contributes $36 million to the city's economy per day. Popular destinations include the Sydney Opera House, the Sydney Harbour Bridge, Watsons Bay, The Rocks, Sydney Tower, Darling Harbour, the Royal Botanic Garden, the Australian Museum, the Museum of Contemporary Art, the Art Gallery of New South Wales, the Queen Victoria Building, Sea Life Sydney Aquarium, Taronga Zoo, Bondi Beach, Luna Park and Sydney Olympic Park. Major developmental projects designed to increase Sydney's tourism sector include a casino and hotel at Barangaroo and the redevelopment of East Darling Harbour, which involves a new exhibition and convention centre, now Australia's largest. Sydney is the highest-ranking city in the world for international students. More than 50,000 international students study at the city's universities and a further 50,000 study at its vocational and English language schools. International education contributes $1.6 billion to the local economy and creates demand for 4,000 local jobs each year. The population of Sydney in 1788 was less than 1,000. With convict transportation it almost tripled in ten years to 2,953. For each decade since 1961 the population has increased by more than 250,000. The 2021 census recorded the population of Greater Sydney as 5,231,150. The Australian Bureau of Statistics (ABS) projects the population will grow to between 8 and 8.9 million by 2061, but that Melbourne will replace Sydney as Australia's most populous city by 2030. The four most densely populated suburbs in Australia are located in Sydney with each having more than 13,000 residents per square kilometre (33,700 residents per square mile). Between 1971 and 2018, Sydney experienced a net loss of 716,832 people to the rest of Australia, but its population grew due to overseas arrivals and a healthy birth rate. The median age of Sydney residents is 37 and 14.8% of people are 65 or older. 48.6% of Sydney's population is married whilst 36.7% have never been married. 49.0% of families are couples with children, 34.4% are couples without children, and 14.8% are single-parent families. Most immigrants to Sydney between 1840 and 1930 were British, Irish or Chinese. At the 2021 census, the most common ancestries were: At the 2021 census, 40.5% of Sydney's population was born overseas. Foreign countries of birth with the greatest representation are Mainland China, India, England, Vietnam, Philippines and New Zealand. At the 2021 census, 1.7% of Sydney's population identified as being Indigenous — Aboriginal Australians and Torres Strait Islanders. 42% of households in Sydney use a language other than English, with the most common being Mandarin (5%), Arabic (4.2%), Cantonese (2.8%), Vietnamese (2.2%) and Hindi (1.5%). In 2021, Christianity was the largest religious affiliation at 46%, the largest denominations of which were Catholicism at 23.1% and Anglicanism at 9.2%. 30.3% of Sydney residents identified as having no religion. The most common non-Christian religious affiliations were Islam (6.3%), Hinduism (4.8%), Buddhism (3.8%), Sikhism (0.7%), and Judaism (0.7%). About 500 people identified with traditional Aboriginal religions. The Church of England was the only recognised church before Governor Macquarie appointed official Catholic chaplains in 1820. Macquarie also ordered the construction of churches such as St Matthew's, St Luke's, St James's, and St Andrew's. Religious groups, alongside secular institutions, have played a significant role in education, health and charitable services throughout Sydney's history. Crime in Sydney is low, with The Independent ranking Sydney as the fifth safest city in the world in 2019. However, drug use is a significant problem. Methamphetamine is heavily consumed compared to other countries, while heroin is less common. One of the biggest crime-related issues in recent times was the introduction of lockout laws in February 2014, in an attempt to curb alcohol-fuelled violence. Patrons could not enter clubs or bars in the inner-city after 1:30am, and last drinks were called at 3am. The lockout laws were removed in January 2020. Ku-ring-gai Chase National Park is rich in Indigenous Australian heritage, containing around 1,500 pieces of Aboriginal rock art – the largest cluster of Indigenous sites in Australia. The park's indigenous sites include petroglyphs, art sites, burial sites, caves, marriage areas, birthing areas, midden sites, and tool manufacturing locations, which are dated to be around 5,000 years old. The inhabitants of the area were the Garigal people. Other rock art sites exist in the Sydney region, such as in Terrey Hills and Bondi, although the locations of most are not publicised to prevent damage by vandalism, and to retain their quality, as they are still regarded as sacred sites by Indigenous Australians. The Australian Museum opened in Sydney in 1827 with the purpose of collecting and displaying the natural wealth of the colony. It remains Australia's oldest natural history museum. In 1995 the Museum of Sydney opened on the site of the first Government House. It recounts the story of the city's development. Other museums include the Powerhouse Museum and the Australian National Maritime Museum. The State Library of New South Wales holds the oldest library collections in Australia, being established as the Australian Subscription Library in 1826. The Royal Society of New South Wales, formed in 1866, encourages "studies and investigations in science, art, literature, and philosophy". It is based in a terrace house in Darlington owned by the University of Sydney. The Sydney Observatory building was constructed in 1859 and used for astronomy and meteorology research until 1982 before being converted into a museum. The Museum of Contemporary Art was opened in 1991 and occupies an Art Deco building in Circular Quay. Its collection was founded in the 1940s by artist and art collector John Power and has been maintained by the University of Sydney. Sydney's other significant art institution is the Art Gallery of New South Wales which coordinates the Archibald Prize for portraiture. Sydney is also home to contemporary art gallery Artspace, housed in the historic Gunnery Building in Woolloomooloo, fronting Sydney Harbour. Sydney's first commercial theatre opened in 1832 and nine more had commenced performances by the late 1920s. The live medium lost much of its popularity to the cinema during the Great Depression before experiencing a revival after World War II. Prominent theatres in the city today include State Theatre, Theatre Royal, Sydney Theatre, The Wharf Theatre, and Capitol Theatre. Sydney Theatre Company maintains a roster of local, classical, and international plays. It occasionally features Australian theatre icons such as David Williamson, Hugo Weaving, and Geoffrey Rush. The city's other prominent theatre companies are New Theatre, Belvoir, and Griffin Theatre Company. Sydney is also home to Event Cinemas' first theatre, which opened on George St in 1913, under its former Greater Union brand; the theatre currently operates, and is regarded as one of Australia's busiest cinema locations. The Sydney Opera House is the home of Opera Australia and Sydney Symphony. It has staged over 100,000 performances and received 100 million visitors since opening in 1973. Two other important performance venues in Sydney are Town Hall and the City Recital Hall. The Sydney Conservatorium of Music is located adjacent to the Royal Botanic Garden and serves the Australian music community through education and its biannual Australian Music Examinations Board exams. Many writers have originated in and set their work in Sydney. Others have visited the city and commented on it. Some of them are commemorated in the Sydney Writers Walk at Circular Quay. The city was the headquarters for Australia's first published newspaper, the Sydney Gazette. Watkin Tench's A Narrative of the Expedition to Botany Bay (1789) and A Complete Account of the Settlement at Port Jackson in New South Wales (1793) have remained the best-known accounts of life in early Sydney. Since the infancy of the establishment, much of the literature set in Sydney were concerned with life in the city's slums and working-class communities, notably William Lane's The Working Man's Paradise (1892), Christina Stead's Seven Poor Men of Sydney (1934) and Ruth Park's The Harp in the South (1948). The first Australian-born female novelist, Louisa Atkinson, set several novels in Sydney. Contemporary writers, such as Elizabeth Harrower, were born in the city and set most of their work there–Harrower's debut novel Down in the City (1957) was mostly set in a King's Cross apartment. Well known contemporary novels set in the city include Melina Marchetta's Looking for Alibrandi (1992), Peter Carey's 30 Days in Sydney: A Wildly Distorted Account (1999), J. M. Coetzee's Diary of a Bad Year (2007) and Kate Grenville's The Secret River (2010). The Sydney Writers' Festival is held annually between April and May. Filmmaking in Sydney was prolific until the 1920s when spoken films were introduced and American productions gained dominance. The Australian New Wave saw a resurgence in film production, with many notable features shot in the city between the 1970s and 80s, helmed by directors such as Bruce Beresford, Peter Weir and Gillian Armstrong. Fox Studios Australia commenced production in Sydney in 1998. Successful films shot in Sydney since then include The Matrix, Lantana, Mission: Impossible 2, Moulin Rouge!, Star Wars: Episode II – Attack of the Clones, Australia, Superman Returns, and The Great Gatsby. The National Institute of Dramatic Art is based in Sydney and has several famous alumni such as Mel Gibson, Judy Davis, Baz Luhrmann, Cate Blanchett, Hugo Weaving and Jacqueline Mckenzie. Sydney hosts several festivals throughout the year. The city's New Year's Eve celebrations are the largest in Australia. The Royal Easter Show is held every year at Sydney Olympic Park. Sydney Festival is Australia's largest arts festival. The travelling rock music festival Big Day Out originated in Sydney. The city's two largest film festivals are Sydney Film Festival and Tropfest. Vivid Sydney is an annual outdoor exhibition of art installations, light projections, and music. In 2015, Sydney was ranked the 13th top fashion capital in the world. It hosts the Australian Fashion Week in autumn. Sydney Mardi Gras has commenced each February since 1979. Sydney's Chinatown has had numerous locations since the 1850s. It moved from George Street to Campbell Street to its current setting in Dixon Street in 1980. Little Italy is located in Stanley Street. Restaurants, bars and nightclubs can be found in the entertainment hubs in the Sydney CBD (Darling Harbour, Barangaroo, The Rocks and George Street), Oxford Street, Surry Hills, Newtown and Parramatta. Kings Cross was previously considered the red-light district. The Star is the city's casino and is situated next to Darling Harbour while the new Crown Sydney resort is in nearby Barangaroo. The Sydney Morning Herald is Australia's oldest newspaper still in print; it has been published continuously since 1831. Its competitor is The Daily Telegraph, in print since 1879. Both papers have Sunday tabloid editions called The Sun-Herald and The Sunday Telegraph respectively. The Bulletin was founded in Sydney in 1880 and became Australia's longest running magazine. It closed after 128 years of continuous publication. Sydney heralded Australia's first newspaper, the Sydney Gazette, published until 1842. Each of Australia's three commercial television networks and two public broadcasters is headquartered in Sydney. Nine's offices and news studios are in North Sydney, Ten is based in Pyrmont, and Seven is based in South Eveleigh in Redfern. The Australian Broadcasting Corporation is located in Ultimo, and the Special Broadcasting Service is based in Artarmon. Multiple digital channels have been provided by all five networks since 2000. Foxtel is based in North Ryde and sells subscription cable television to most of the urban area. Sydney's first radio stations commenced broadcasting in the 1920s. Radio has managed to survive despite the introduction of television and the Internet. 2UE was founded in 1925 and under the ownership of Nine Entertainment is the oldest station still broadcasting. Competing stations include the more popular 2GB, ABC Radio Sydney, KIIS 106.5, Triple M, Nova 96.9 and 2Day FM. Sydney's earliest migrants brought with them a passion for sport but were restricted by the lack of facilities and equipment. The first organised sports were boxing, wrestling, and horse racing from 1810 in Hyde Park. Horse racing remains popular and events such as the Golden Slipper Stakes attract widespread attention. The first cricket club was formed in 1826 and matches were played within Hyde Park throughout the 1830s and 1840s. Cricket is a favoured sport in summer and big matches have been held at the Sydney Cricket Ground since 1878. The New South Wales Blues compete in the Sheffield Shield league and the Sydney Sixers and Sydney Thunder contest the national Big Bash Twenty20 competition. First played in Sydney in 1865, rugby grew to be the city's most popular football code by the 1880s. One-tenth of the state's population attended a New South Wales versus New Zealand rugby match in 1907. Rugby league separated from rugby union in 1908. The New South Wales Waratahs contest the Super Rugby competition, while the Sydney Rays represent the city in the National Rugby Championship. The national Wallabies rugby union team competes in Sydney in international matches such as the Bledisloe Cup, Rugby Championship, and World Cup. Sydney is home to nine of the sixteen teams in the National Rugby League competition: Canterbury-Bankstown Bulldogs, Cronulla-Sutherland Sharks, Manly-Warringah Sea Eagles, Penrith Panthers, Parramatta Eels, South Sydney Rabbitohs, St George Illawarra Dragons, Sydney Roosters, and Wests Tigers. New South Wales contests the annual State of Origin series against Queensland. Sydney FC and the Western Sydney Wanderers compete in the A-League (men's) and W-League (women's) soccer competitions and Sydney frequently hosts matches for the Australian national men's team, the Socceroos. The Sydney Swans and Greater Western Sydney Giants are local Australian rules football clubs that play in the Australian Football League and the AFL Women's. The Sydney Kings compete in the National Basketball League. The Sydney Uni Flames play in the Women's National Basketball League. The Sydney Blue Sox contest the Australian Baseball League. The NSW Pride are a member of the Hockey One League. The Sydney Bears and Sydney Ice Dogs play in the Australian Ice Hockey League. The Swifts are competitors in the national women's netball league. Women were first allowed to participate in recreational swimming when separate baths were opened at Woolloomooloo Bay in the 1830s. From being illegal at the beginning of the century, sea bathing gained immense popularity during the early 1900s and the first surf lifesaving club was established at Bondi Beach. Disputes about appropriate clothing for surf bathing surfaced occasionally and concerned men as well as women. The City2Surf is an annual 14 km (8.7 mi) running race from the CBD to Bondi Beach and has been held since 1971. In 2010, 80,000 runners participated which made it the largest run of its kind in the world. Sailing races have been held on Sydney Harbour since 1827. Yachting has been popular amongst wealthier residents since the 1840s and the Royal Sydney Yacht Squadron was founded in 1862. The Sydney to Hobart Yacht Race is a 1,170 km (727 mi) event that starts from Sydney Harbour on Boxing Day. Since its inception in 1945 it has been recognised as one of the most difficult yacht races in the world. Six sailors died and 71 vessels of 115 failed to finish in the 1998 edition. The Royal Sydney Golf Club is based in Rose Bay and since its opening in 1893 has hosted the Australian Open on 13 occasions. Royal Randwick Racecourse opened in 1833 and holds several major cups throughout the year. Sydney benefitted from the construction of significant sporting infrastructure in preparation for its hosting of the 2000 Summer Olympics. The Sydney Olympic Park accommodates athletics, aquatics, tennis, hockey, archery, baseball, cycling, equestrian, and rowing facilities. It also includes the high capacity Stadium Australia used for rugby, soccer, and Australian rules football. The Sydney Football Stadium was completed in 1988 and was used for rugby and soccer matches. Sydney Cricket Ground was opened in 1878 and is used for both cricket and Australian rules football fixtures. The Sydney International tennis tournament is held here at the beginning of each year as the warm-up for the Grand Slam in Melbourne. Two of the most successful tennis players in history (Ken Rosewall and Todd Woodbridge) were born in and live in the city. Sydney co-hosted the FIBA Oceania Championship in 1979, 1985, 1989, 1995, 2007, 2009 and 2011. The first five governors had near autocratic power in the colony of New South Wales, subject only to the laws of England and the supervision of the Colonial Office in London. Sydney was the seat of government for the colony which encompassed over half the Australian continent. The first Legislative Council met in 1826, and in 1842, the imperial parliament expanded and reformed the council, making it partly elected. In the same year, the town of Sydney officially became a city and an elected municipal council was established. The council had limited powers, mostly relating to services such as street lighting and drainage. Its boundaries were restricted to an area of 11.6 square kilometres, taking in the city centre and the modern suburbs of Woolloomooloo, Surry Hills, Chippendale, and Pyrmont. As Sydney grew, other municipal councils were formed to provide local administration. In 1856, New South Wales achieved responsible government with the introduction of a bicameral parliament, based in Sydney, comprising a directly elected Legislative Assembly and a nominated Legislative Council. With the federation of the Australian colonies in 1901, Sydney became the capital of the state of New South Wales and its administration was divided between the Commonwealth, State and constituent local governments. In common with other Australian capital cities, Sydney has no single local government covering its whole area. Local government areas have responsibilities such as local roads, libraries, child care, community services and waste collection, whereas the state government retains responsibility for main roads, traffic control, public transport, policing, education, and major infrastructure project. There are 33 local government areas which are wholly or mostly within Greater Sydney as defined by the Australian Statistical Geography Standard. Sydney is the location of the secondary official residences of the Governor-General and Prime Minister – Admiralty House and Kirribilli House respectively. The Parliament of New South Wales sits in Parliament House on Macquarie Street. This building was completed in 1816 and first served as a hospital. The Legislative Council moved into its northern wing in 1829 and by 1852 had entirely supplanted the surgeons from their quarters. Several additions have been made as the Parliament has expanded, but it retains its original Georgian façade. Government House was completed in 1845 and has served as the home of 25 Governors and 5 Governors-General. The Cabinet of Australia also meets in Sydney when needed. The highest court in the state is the Supreme Court of New South Wales, located in Queen's Square. The city is also the home of numerous branches of the intermediate District Court of New South Wales and the lower Local Court of New South Wales. In the past, the state has tended to resist amalgamating Sydney's more populated local government areas as merged councils could pose a threat to its governmental power. Established in 1842, the City of Sydney is one such local government area and includes the CBD and some adjoining inner suburbs. It is responsible for fostering development in the local area, providing local services (waste collection and recycling, libraries, parks, sporting facilities), promoting the interests of residents, supporting organisations that target the local community, and attracting and providing infrastructure for commerce, tourism, and industry. The City of Sydney is led by an elected Council and Lord Mayor. In federal politics, Sydney was initially considered as a possibility for Australia's capital city; the newly created city of Canberra ultimately filled this role. Seven Australian Prime Ministers have been born in Sydney, more than any other city, including first Prime Minister Edmund Barton and current Prime Minister Anthony Albanese. Essential public emergency services are provided and managed by the State Government. Greater Sydney is served by: Education became a focus for the colony from the 1870s when public schools began to form and schooling became compulsory. By 2011, 90% of working age residents had completed some schooling and 57% had completed the highest level of school. 1,390,703 people were enrolled in an educational institution in 2011 with 45.1% of these attending school and 16.5% studying at a university. Undergraduate or postgraduate qualifications are held by 22.5% of working age Sydney residents and 40.2% of working age residents of the City of Sydney. The most common fields of tertiary qualification are commerce (22.8%), engineering (13.4%), society and culture (10.8%), health (7.8%), and education (6.6%). There are six public universities based in Sydney: The University of Sydney, University of New South Wales, University of Technology Sydney, Macquarie University, Western Sydney University, and Australian Catholic University. Five public universities maintain secondary campuses in the city: the University of Notre Dame Australia, Central Queensland University, Victoria University, University of Wollongong, and University of Newcastle. Charles Sturt University and Southern Cross University operate secondary campuses only designated for international students. In addition, four public universities offer programmes in Sydney through third-party providers: University of the Sunshine Coast, La Trobe University, Federation University Australia and Charles Darwin University. 5.2% of residents of Sydney are attending a university. The University of New South Wales and the University of Sydney are ranked top 50 in the world, the University of Technology Sydney is ranked 133, while Macquarie University is ranked 237, and Western Sydney University is ranked 474. Sydney has public, denominational, and independent schools. 7.8% of Sydney residents are attending primary school and 6.4% are enrolled in secondary school. There are 935 public preschool, primary, and secondary schools in Sydney that are administered by the New South Wales Department of Education. 14 of the 17 selective secondary schools in New South Wales are based in Sydney. Public vocational education and training in Sydney are run by TAFE New South Wales and began with the opening of the Sydney Technical College in 1878. The college became the Sydney Institute in 1992 and now operates alongside its sister TAFE facilities across the Sydney metropolitan area, namely the Northern Sydney Institute, the Western Sydney Institute, and the South Western Sydney Institute. At the 2011 census, 2.4% of Sydney residents are enrolled in a TAFE course. The first hospital in the new colony was a collection of tents at The Rocks. Many of the convicts that survived the trip suffered from dysentery, smallpox, scurvy, and typhoid. Healthcare facilities remained inadequate despite the arrival of a prefabricated hospital with the Second Fleet and the construction of new hospitals at Parramatta, Windsor, and Liverpool in the 1790s. Governor Macquarie arranged for the construction of Sydney Hospital, completed in 1816. Parts of the facility have been repurposed for use as Parliament House but the hospital itself still operates. The city's first emergency department was established at Sydney Hospital in 1870. Demand for emergency medical care increased from 1895 with the introduction of an ambulance service. The Sydney Hospital also housed Australia's first teaching facility for nurses, the Nightingale Wing, established with the input of Florence Nightingale in 1868. Healthcare was recognised as a right in the early 1900s and Sydney's public hospitals came under the oversight of the Government of New South Wales. The administration of healthcare across Sydney is handled by eight local health districts: Central Coast, Illawarra Shoalhaven, Sydney, Nepean Blue Mountains, Northern Sydney, South Eastern Sydney, South Western Sydney, and Western Sydney. The Prince of Wales Hospital was established in 1852 and became the first of several major hospitals to be opened. St Vincent's Hospital was founded in 1857, followed by Royal Alexandra Hospital for Children in 1880, the Prince Henry Hospital in 1881, the Royal Prince Alfred Hospital in 1882, the Royal North Shore Hospital in 1885, the St George Hospital in 1894, and the Nepean Hospital in 1895. Westmead Hospital in 1978 was the last major facility to open. The motor vehicle, more than any other factor, has determined the pattern of Sydney's urban development since World War II. The growth of low-density housing in the city's outer suburbs has made car ownership necessary for hundreds of thousands of households. The percentage of trips taken by car has increased from 13% in 1947 to 50% in 1960 and 70% in 1971. The most important roads in Sydney were the nine Metroads, including the 110 km (68 mi) Sydney Orbital Network. Sydney's reliance on motor vehicles and its sprawling road network has been criticised by proponents of mass public transport and high-density housing. The Light Horse Interchange in western Sydney is the largest in the southern hemisphere. There can be up to 350,000 cars using Sydney's roads simultaneously during peak hour, leading to significant traffic congestion. 84.9% of Sydney households own a motor vehicle and 46.5% own two or more. Car dependency is an ongoing issue in Sydney–of people who travel to work, 58.4% use a car, 9.1% catch a train, 5.2% take a bus, and 4.1% walk. In contrast, only 25.2% of working residents in the City of Sydney use a car, whilst 15.8% take a train, 13.3% use a bus, and 25.3% walk. With a rate of 26.3%, Sydney has the highest utilisation of public transport for travel to work of any Australian capital. The CBD features a series of alleyways and lanes that provide off-street vehicular access to city buildings and as well as pedestrian routes through city buildings. Bus services are conducted by private operators under contract to Transport for NSW. Integrated tickets called Opal cards operate on bus routes. In total, nearly 225 million boardings were recorded across the bus network. NightRide is a nightly bus service that operate between midnight and 5am. Sydney once had one of the largest tram networks in the British Empire after London. It served routes covering 291 km (181 mi). The internal combustion engine made buses more flexible than trams and consequently more popular, leading to the progressive closure of the network with the final tram operating in 1961. From 1930 there were 612 buses across Sydney carrying 90 million passengers per annum. In 1997, the Inner West Light Rail opened between Central station and Wentworth Park. It was extended to Lilyfield in 2000 and then Dulwich Hill in 2014. It links the Inner West and Darling Harbour with Central station and facilitated 9.1 million journeys in the 2016–17 financial year. A second, the CBD and South East Light Rail 12 km (7.5 mi) line serving the CBD and south-eastern suburbs opened in 2019–2020. A light rail line serving Western Sydney has also been announced, due to open in 2024. Established in 1906, Central station is the largest and busiest railway station in the state and is the main hub of the city's rail network. Sydney Trains is the suburban rail service. Its tracks form part of the New South Wales railway network. It serves 175 stations across the city and had an annual ridership of 359 million passenger journeys in 2017–18. Sydney's railway was first constructed in 1854 with progressive extension to the network to serve both freight and passengers. The main station is the Central railway station in the southern part of the CBD. In the 1850s and 1860s, the railway reached areas that are now outer suburbs of Sydney. Main article: Sydney Metro Sydney Metro, a driverless rapid transit system separate from the suburban commuter network, commenced operation in May 2019 and will be extended into the city and down southwest by 2024 and through the inner west to Parramatta by 2030. It currently serves 13 stations. A line to serve the greater west is planned for 2026 and will include a station for the second international airport. At the time the Sydney Harbour Bridge opened in 1932, the city's ferry service was the largest in the world. Patronage declined from 37 million passengers in 1945 to 11 million in 1963 but has recovered somewhat in recent years. From its hub at Circular Quay, the ferry network extends from Manly to Parramatta. Sydney Airport, officially "Sydney Kingsford-Smith Airport", is located in Mascot. It services 46 international and 23 domestic destinations. As the busiest airport in Australia, it handled 37.9 million passengers in 2013 and 530,000 tonnes of freight in 2011. It has been announced that a new facility named Western Sydney Airport will be constructed at Badgerys Creek from 2016 at a cost of $2.5 billion. Bankstown Airport is Sydney's second busiest airport, and serves general aviation, charter and some scheduled cargo flights. Bankstown is also the fourth busiest airport in Australia by number of aircraft movements. Port Botany has surpassed Port Jackson as the city's major shipping port. Cruise ship terminals are located at Sydney Cove and White Bay. Obtaining sufficient fresh water was difficult during early colonial times. A catchment called the Tank Stream sourced water from what is now the CBD but was little more than an open sewer by the end of the 1700s. The Botany Swamps Scheme was one of several ventures during the mid-1800s that saw the construction of wells, tunnels, steam pumping stations, and small dams to service Sydney's growing population. The Upper Nepean Scheme came into operation in 1886. It transports water 100 km (62 mi) from the Nepean, Cataract, and Cordeaux rivers and continues to service about 15% of Sydney's water needs. Dams were built on these three rivers between 1907 and 1935. In 1977 the Shoalhaven Scheme brought several more dams into service. The state-owned corporation WaterNSW now manages eleven major dams: Warragamba, one of the largest domestic water supply dams in the world, Woronora, Cataract, Cordeaux, Nepean, Avon, Wingecarribee Reservoir, Fitzroy Falls Reservoir, Tallowa, the Blue Mountains Dams, and Prospect Reservoir. Water is collected from five catchment areas covering 16,000 km (6,178 sq mi) and total storage amounts to 2.6 TL (0.6 cu mi). The Sydney Desalination Plant came into operation in 2010. WaterNSW supplies bulk water to Sydney Water, a state-owned corporation that operates water distribution, sewerage and storm water management services. Sydney's electricity infrastructure is maintained by Ausgrid and Endeavour Energy. Their combined networks include over 815,000 poles and 83,000 km (52,000 mi) of cables. Submarine communications cable systems in Sydney include the Australia–Japan Cable, Telstra Endeavour and the Southern Cross Cable, which link Australia and countries in the Pacific. As climate change, greenhouse gas emissions and pollution have become a major issue for Australia, Sydney has in the past been criticised for its lack of focus on reducing pollution and emissions and maintaining water quality. The release of the Metropolitan Air Quality Scheme (MAQS) led to a broader understanding of the causation of pollution in Sydney, allowing the government to form appropriate responses. The 2019–20 Australian bushfire season significantly impacted outer Sydney and dramatically reduced air quality, leading to a smoky haze that lingered for days. The air quality was 11 times the hazardous level in some days, worse than New Delhi's; it was compared to "smoking 32 cigarettes" by Brian Oliver, a respiratory diseases scientist at the University of Technology Sydney. Since Sydney is surrounded by bushland and forest, bushfires can ring the region in a natural phenomena that is labelled "ring of fire". The City of Sydney became the first council in Australia to achieve formal certification as carbon-neutral in 2008. The city has reduced its 2007 carbon emissions by 6% and since 2006 has reduced carbon emissions from city buildings by up to 20%. The Sustainable Sydney 2030 program presented a guide to reducing energy in homes and offices by 30%. Reductions in energy consumption have slashed energy bills by $30 million a year. Solar panels have been established on many CBD buildings to minimise carbon pollution by around 3,000 tonnes a year. The city also has an "urban forest growth strategy", in which it aims to regular increase the tree coverage in the city by frequently planting trees with strong leaf density and vegetation to provide cleaner air and create moisture during hot weather, thus lowering city temperatures. Sydney has also become a leader in the development of green office buildings and enforcing the requirement of all building proposals to be energy-efficient. The One Central Park development, completed in 2013, is an example of this implementation. Australian cities are some of the most car-dependent cities in the world, especially by world city standards, although Sydney's is the lowest of Australia's major cities at 66%. Sydney also has the highest usage of public transport in an Australian city, at 27%–comparable with New York City, Shanghai and Berlin. Despite its high ranking for an Australian city, Sydney has a low level of mass-transit services, with a historically low-density layout and significant urban sprawl, thus increasing the likelihood of car dependency. Strategies have been implemented to reduce private vehicle pollution by encouraging mass and public transit, initiating the development of high density housing and introducing a fleet of 10 new electric cars, the largest order of the pollution-free vehicle in Australia. Electric cars do not produce carbon monoxide and nitrous oxide, which contribute to climate change. Cycling trips have increased by 113% across Sydney's inner-city since March 2010, with about 2,000 bikes passing through top peak-hour intersections on an average weekday. Transport developments in the north-west and east of the city have been designed to encourage use of the expanding public transportation system.
[[File:Greenhouse Effect (2017 NASA data).svg|thumb|222x222px]] [[File:The green house effect.svg|thumb]] The '''greenhouse effect''' occurs when certain gases in the [[Earth's atmosphere]] (the [[air]] around the [[Earth]]) trap [[infrared radiation]]. This makes the planet become warmer, similar to the way a [[greenhouse]] becomes warmer.<ref>IPCC, 2013: [https://www.ipcc.ch/site/assets/uploads/2018/02/WG1AR5_AnnexIII_FINAL.pdf Annex III: Glossary] [Planton S. (ed)]. In: ''Climate Change 2013: The Physical Science Basis. Contribution of Working Group I to the Fifth Assessment Report of the Intergovernmental Panel on Climate Change'' Stocker T.F. D. Qin G.-K. Plattner M. Tignor S.K. Allen J. Boschung A. Nauels Y. Xia V. Bex and P.M. Midgley (eds). Cambridge University Press, Cambridge, United Kingdom and New York.</ref> The most important [[greenhouse gas]]es in [[Earth's atmosphere]] are: [[water vapor]], [[carbon dioxide]](CO<sub>2</sub>), and [[methane]]. When there is more greenhouse gas in the air, the air holds more heat. This is why more greenhouse gases cause [[climate change]] and [[global warming]].<ref> {{cite book | title = The Earth's biosphere: evolution, dynamics, and change| author = Vaclav Smil | publisher = MIT Press | date = 2003 | isbn = 978-0-262-69298-4 | page = 107 | url = https://books.google.com/books?id=8ntHWPMUgpMC&pg=PA107}}</ref> The greenhouse effect is natural. It is important for life on Earth. Without the greenhouse effect, the Earth's average temperature would be around {{ndash}}18 or –19 degrees [[Celsius]] (0 or 1 degree Fahrenheit). Earth would be locked in an [[ice age]]. Because of the greenhouse effect, the Earth's actual average temperature is 14 degrees Celsius (57 degrees Fahrenheit). The problem is that recently, the greenhouse effect has become stronger. This is because humans have been burning large amounts of [[fossil fuel]]s, which releases carbon dioxide. Since carbon dioxide is a greenhouse gas, it has [[global warming|caused the planet to warm]] over the past 150 years. [[File:191203 Furnaces of the world - Popular Mechanics - Global warming.jpg|thumb| upright=2.0| The greenhouse effect and its impact on climate were succinctly described in this 1912 ''[[Popular Mechanics]]'' article.]] About 10,000 years ago, before people started burning large amounts of [[fossil fuels]], there were 260 to 280 parts per million (ppm) of carbon dioxide (CO2) in the atmosphere, but now there is over 400&nbsp;ppm. Most scientists say that having 350&nbsp;ppm or less is safe for the [[environment]] and that [[species]] on the planet can adapt to this level. Higher levels can make severe problems for animal and marine life that are already being seen today, such as [[ocean acidification]]. The greenhouse effect was first proposed by [[Joseph Fourier]] in 1824. [[Mars]], Venus and other planets with atmospheres also have greenhouse effects. The effect on Venus is especially strong because Venus has so much CO<sub>2</sub>. This is why [[Venus]] is hotter than [[Mercury (planet)|Mercury]], even though Mercury is closer to the sun. The first person to predict that carbon dioxide from the burning of fossil fuels (and other combustion processes) could cause global warming was Nobel Prize winner [[Svante Arrhenius]]. ==Related pages== * [[Greenhouse gas]] * [[Greenhouse gas emissions]] *[[Global warming]] == References == {{Reflist}} == Other websites == * [http://www.climate-change-guide.com/greenhouse-effect.html The Climate Change Guide] easy-to-understand information on the greenhouse effect * [https://www.nytimes.com/2013/03/01/science/earth/at-ice-age-end-a-smaller-gap-in-warming-and-carbon-dioxide.html ''Study of Ice Age bolsters carbon and warming link''] February 28, 2013 [[The New York Times]] * [http://marine.rutgers.edu/mrs/education/class/yuri/erb.html Rutgers University: Earth Radiation Budget] {{Webarchive|url=https://web.archive.org/web/20060901043314/http://marine.rutgers.edu/mrs/education/class/yuri/erb.html |date=2006-09-01 }} * [http://en.citizendium.org/wiki/Greenhouse_effect Greenhouse effect] -Citizendium [[Category:Weather]] [[Category:Greenhouse gases]] [[Category:Ecology]]
The greenhouse effect occurs when greenhouse gases in a planet's atmosphere trap some of the heat radiated from the planet's surface, raising its temperature. This process happens because stars emit shortwave radiation that passes through greenhouse gases, but planets emit longwave radiation that is partly absorbed by greenhouse gases. That difference reduces the rate at which a planet can cool off in response to being warmed by its host star. Adding to greenhouse gases further reduces the rate a planet emits radiation to space, raising its average surface temperature. The Earth's average surface temperature would be about −18 °C (−0.4 °F) without the greenhouse effect, compared to Earth's 20th century average of about 14 °C (57 °F), or a more recent average of about 15 °C (59 °F). In addition to naturally present greenhouse gases, burning of fossil fuels has increased amounts of carbon dioxide and methane in the atmosphere. As a result, global warming of about 1.2 °C (2.2 °F) has occurred since the industrial revolution, with the global average surface temperature increasing at a rate of 0.18 °C (0.32 °F) per decade since 1981. The wavelengths of radiation emitted by the Sun and Earth differ because their surface temperatures are different. The Sun has a surface temperature of 5,500 °C (9,900 °F), so it emits most of its energy as shortwave radiation in near-infrared and visible wavelengths (as sunlight). In contrast, Earth's surface has a much lower temperature, so it emits longwave radiation at mid- and far-infrared wavelengths (sometimes called thermal radiation or radiated heat). A gas is a greenhouse gas if it absorbs longwave radiation. Earth's atmosphere absorbs only 23% of incoming shortwave radiation, but absorbs 90% of the longwave radiation emitted by the surface, thus accumulating energy and warming the Earth's surface. The existence of the greenhouse effect, while not named as such, was proposed as early as 1824 by Joseph Fourier. The argument and the evidence were further strengthened by Claude Pouillet in 1827 and 1838. In 1856 Eunice Newton Foote demonstrated that the warming effect of the sun is greater for air with water vapour than for dry air, and the effect is even greater with carbon dioxide. The term greenhouse was first applied to this phenomenon by Nils Gustaf Ekholm in 1901. The greenhouse effect on Earth is defined as: "The infrared radiative effect of all infrared absorbing constituents in the atmosphere. Greenhouse gases (GHGs), clouds, and some aerosols absorb terrestrial radiation emitted by the Earth’s surface and elsewhere in the atmosphere." The enhanced greenhouse effect describes the fact that by increasing the concentration of GHGs in the atmosphere (due to human action), the natural greenhouse effect is increased. The term greenhouse effect comes from an analogy to greenhouses. Both greenhouses and the greenhouse effect work by retaining heat from sunlight, but the way they retain heat differs. Greenhouses retain heat mainly by blocking convection (the movement of air). In contrast, the greenhouse effect retains heat by restricting radiative transfer through the air and reducing the rate at which heat escapes to space. Matter emits thermal radiation in an amount that is directly proportional to the fourth power of its temperature. Some of the radiation emitted by the Earth's surface is absorbed by greenhouse gases and clouds. Without this absorption, Earth's surface would have an average temperature of −18 °C (−0.4 °F). However, because some of the radiation is absorbed, Earth's average surface temperature is around 15 °C (59 °F). Thus, the Earth's greenhouse effect may be measured as a temperature change of 33 °C (59 °F). Thermal radiation is characterized by how much energy it carries, typically in watts per square meter (W/m). Scientists also measure the greenhouse effect based on how much more longwave thermal radiation leaves the Earth's surface than reaches space. Currently, longwave radiation leaves the surface at an average rate of 398 W/m, but only 239 W/m reaches space. Thus, the Earth's greenhouse effect can also be measured as an energy flow change of 159 W/m. The greenhouse effect can be expressed as a fraction (0.40) or percentage (40%) of the longwave thermal radiation that leaves Earth's surface but does not reach space. Whether the greenhouse effect is expressed as a change in temperature or as a change in longwave thermal radiation, the same effect is being measured. Strengthening of the greenhouse effect through additional greenhouse gases from human activities is known as the enhanced greenhouse effect. As well as being inferred from measurements by ARGO, CERES and other instruments throughout the 21st century, this increase in radiative forcing from human activity has been observed directly, and is attributable mainly to increased atmospheric carbon dioxide levels. CO2 is produced by fossil fuel burning and other activities such as cement production and tropical deforestation. Measurements of CO2 from the Mauna Loa Observatory show that concentrations have increased from about 313 parts per million (ppm) in 1960, passing the 400 ppm milestone in 2013. The current observed amount of CO2 exceeds the geological record maxima (≈300 ppm) from ice core data. Over the past 800,000 years, ice core data shows that carbon dioxide has varied from values as low as 180 ppm to the pre-industrial level of 270 ppm. Paleoclimatologists consider variations in carbon dioxide concentration to be a fundamental factor influencing climate variations over this time scale. Hotter matter emits shorter wavelengths of radiation. As a result, the Sun emits shortwave radiation as sunlight while the Earth and its atmosphere emit longwave radiation. Sunlight includes ultraviolet, visible light, and near-infrared radiation. Sunlight is reflected and absorbed by the Earth and its atmosphere. The atmosphere and clouds reflect about 23% and absorb 23%. The surface reflects 7% and absorbs 48%. Overall, Earth reflects about 30% of the incoming sunlight, and absorbs the rest (240 W/m). The Earth and its atmosphere emit longwave radiation, also known as thermal infrared or terrestrial radiation. Informally, longwave radiation is sometimes called thermal radiation. Outgoing longwave radiation (OLR) is the radiation from Earth and its atmosphere that passes through the atmosphere and into space. The greenhouse effect can be directly seen in graphs of Earth's outgoing longwave radiation as a function of frequency (or wavelength). The area between the curve for longwave radiation emitted by Earth's surface and the curve for outgoing longwave radiation indicates the size of the greenhouse effect. Different substances are responsible for reducing the radiation energy reaching space at different frequencies; for some frequencies, multiple substances play a role. Carbon dioxide is understood to be responsible for the dip in outgoing radiation (and associated rise in the greenhouse effect) at around 667 cm (equivalent to a wavelength of 15 microns). Each layer of the atmosphere with greenhouse gases absorbs some of the longwave radiation being radiated upwards from lower layers. It also emits longwave radiation in all directions, both upwards and downwards, in equilibrium with the amount it has absorbed. This results in less radiative heat loss and more warmth below. Increasing the concentration of the gases increases the amount of absorption and emission, and thereby causing more heat to be retained at the surface and in the layers below. The power of outgoing longwave radiation emitted by a planet corresponds to the effective temperature of the planet. The effective temperature is the temperature that a planet radiating with a uniform temperature (a blackbody) would need to have in order to radiate the same amount of energy. This concept may be used to compare the amount of longwave radiation emitted to space and the amount of longwave radiation emitted by the surface: Earth's surface temperature is often reported in terms of the average near-surface air temperature. This is about 15 °C (59 °F), a bit lower than the effective surface temperature. This value is 33 °C (59 °F) warmer than Earth's overall effective temperature. Energy flux is the rate of energy flow per unit area. Energy flux is expressed in units of W/m, which is the number of joules of energy that pass through a square meter each second. Most fluxes quoted in high-level discussions of climate are global values, which means they are the total flow of energy over the entire globe, divided by the surface area of the Earth, 5.1×10 m (5.1×10 km; 2.0×10 sq mi). The fluxes of radiation arriving at and leaving the Earth are important because radiative transfer is the only process capable of exchanging energy between Earth and the rest of the universe. The temperature of a planet depends on the balance between incoming radiation and outgoing radiation. If incoming radiation exceeds outgoing radiation, a planet will warm. If outgoing radiation exceeds incoming radiation, a planet will cool. A planet will tend towards a state of radiative equilibrium, in which the power of outgoing radiation equals the power of absorbed incoming radiation. Earth's energy imbalance is the amount by which the power of incoming sunlight absorbed by Earth's surface or atmosphere exceeds the power of outgoing longwave radiation emitted to space. Energy imbalance is the fundamental measurement that drives surface temperature. A UN presentation says "The EEI is the most critical number defining the prospects for continued global warming and climate change." One study argues, "The absolute value of EEI represents the most fundamental metric defining the status of global climate change." Earth's energy imbalance (EEI) was about 0.7 W/m as of around 2015, indicating that Earth as a whole is accumulating thermal energy and is in a process of becoming warmer. Over 90% of the retained energy goes into warming the oceans, with much smaller amounts going into heating the land, atmosphere, and ice. A simple picture assumes a steady state, but in the real world, the day/night (diurnal) cycle, as well as the seasonal cycle and weather disturbances, complicate matters. Solar heating applies only during daytime. At night the atmosphere cools somewhat, but not greatly because the thermal inertia of the climate system resists changes both day and night, as well as for longer periods. Diurnal temperature changes decrease with height in the atmosphere. In the lower portion of the atmosphere, the troposphere, the air temperature decreases (or "lapses") with increasing altitude. The rate at which temperature changes with altitude is called the lapse rate. On Earth, the air temperature decreases by about 6.5°C/km (3.6°F per 1000 ft), on average, although this varies. The temperature lapse is caused by convection. Air warmed by the surface rises. As it rises, air expands and cools. Simultaneously, other air descends, compresses, and warms. This process creates a vertical temperature gradient within the atmosphere. This vertical temperature gradient is essential to the greenhouse effect. If the lapse rate was zero (so that the atmospheric temperature did not vary with altitude and was the same as the surface temperature) then there would be no greenhouse effect (i.e., its value would be zero). Greenhouse gases make the atmosphere near Earth's surface mostly opaque to longwave radiation. The atmosphere only becomes transparent to longwave radiation at higher altitudes, where the air is less dense, there is less water vapor, and reduced pressure broadening of absorption lines limits the wavelengths that gas molecules can absorb. For any given wavelength, the longwave radiation that reaches space is emitted by a particular radiating layer of the atmosphere. The intensity of the emitted radiation is determined by the weighted average air temperature within that layer. So, for any given wavelength of radiation emitted to space, there is an associated effective emission temperature (or brightness temperature). A given wavelength of radiation may also be said to have an effective emission altitude, which is a weighted average of the altitudes within the radiating layer. The effective emission temperature and altitude vary by wavelength (or frequency). This phenomenon may be seen by examining plots of radiation emitted to space. Earth's surface radiates longwave radiation with wavelengths in the range of 4–100 microns. Greenhouse gases that were largely transparent to incoming solar radiation are more absorbent for some wavelengths in this range. The atmosphere near the Earth's surface is largely opaque to longwave radiation and most heat loss from the surface is by evaporation and convection. However radiative energy losses become increasingly important higher in the atmosphere, largely because of the decreasing concentration of water vapor, an important greenhouse gas. Rather than thinking of longwave radiation headed to space as coming from the surface itself, it is more realistic to think of this outgoing radiation as being emitted by a layer in the mid-troposphere, which is effectively coupled to the surface by a lapse rate. The difference in temperature between these two locations explains the difference between surface emissions and emissions to space, i.e., it explains the greenhouse effect. A greenhouse gas (GHG) is a gas which contributes to the trapping of heat by impeding the flow of longwave radiation out of a planet's atmosphere. Greenhouse gases contribute most of the greenhouse effect in Earth's energy budget. Gases which can absorb and emit longwave radiation are said to be infrared active and act as greenhouse gases. Most gases whose molecules have two different atoms (such as carbon monoxide, CO), and all gases with three or more atoms (including H2O and CO2), are infrared active and act as greenhouse gases. (Technically, this is because when these molecules vibrate, those vibrations modify the molecular dipole moment, or asymmetry in the distribution of electrical charge. See Infrared spectroscopy.) Gases with only one atom (such as argon, Ar) or with two identical atoms (such as nitrogen, N2, and oxygen, O2) are not infrared active. They are transparent to longwave radiation, and, for practical purposes, do not absorb or emit longwave radiation. (This is because their molecules are symmetrical and so do not have a dipole moment.) Such gases make up more than 99% of the dry atmosphere. Greenhouse gases absorb and emit longwave radiation within specific ranges of wavelengths (organized as spectral lines or bands). When greenhouse gases absorb radiation, they distribute the acquired energy to the surrounding air as thermal energy (i.e., kinetic energy of gas molecules). Energy is transferred from greenhouse gas molecules to other molecules via molecular collisions. Contrary to what is sometimes said, greenhouse gases do not "re-emit" photons after they are absorbed. Because each molecule experiences billions of collisions per second, any energy a greenhouse gas molecule receives by absorbing a photon will be redistributed to other molecules before there is a chance for a new photon to be emitted. In a separate process, greenhouse gases emit longwave radiation, at a rate determined by the air temperature. This thermal energy is either absorbed by other greenhouse gas molecules or leaves the atmosphere, cooling it. Effect on air: Air is warmed by latent heat (buoyant water vapor condensing into water droplets and releasing heat), thermals (warm air rising from below), and by sunlight being absorbed in the atmosphere. Air is cooled radiatively, by greenhouse gases and clouds emitting longwave thermal radiation. Within the troposphere, greenhouse gases typically have a net cooling effect on air, emitting more thermal radiation than they absorb. Warming and cooling of air are well balanced, on average, so that the atmosphere maintains a roughly stable average temperature. Effect on surface cooling: Longwave radiation flows both upward and downward due to absorption and emission in the atmosphere. These canceling energy flows reduce radiative surface cooling (net upward radiative energy flow). Latent heat transport and thermals provide non-radiative surface cooling which partially compensates for this reduction, but there is still a net reduction in surface cooling, for a given surface temperature. Effect on TOA energy balance: Greenhouse gases impact the top-of-atmosphere (TOA) energy budget by reducing the flux of longwave radiation emitted to space, for a given surface temperature. Thus, greenhouse gases alter the energy balance at TOA. This means that the surface temperature needs to be higher (than the planet's effective temperature, i.e., the temperature associated with emissions to space), in order for the outgoing energy emitted to space to balance the incoming energy from sunlight. It is important to focus on the top-of-atmosphere (TOA) energy budget (rather than the surface energy budget) when reasoning about the warming effect of greenhouse gases. Clouds and aerosols have both cooling effects, associated with reflecting sunlight back to space, and warming effects, associated with trapping thermal radiation. On average, clouds have a strong net cooling effect. However, the mix of cooling and warming effects varies, depending on detailed characteristics of particular clouds (including their type, height, and optical properties). Thin cirrus clouds can have a net warming effect. Clouds can absorb and emit infrared radiation and thus affect the radiative properties of the atmosphere. While the radiative forcing due to greenhouse gases may be determined to a reasonably high degree of accuracy... the uncertainties relating to aerosol radiative forcings remain large, and rely to a large extent on the estimates from global modeling studies that are difficult to verify at the present time. A given flux of thermal radiation has an associated effective radiating temperature or effective temperature. Effective temperature is the temperature that a black body (a perfect absorber/emitter) would need to be to emit that much thermal radiation. Thus, the overall effective temperature of a planet is given by where OLR is the average flux (power per unit area) of outgoing longwave radiation emitted to space and σ {\displaystyle \sigma } is the Stefan-Boltzmann constant. Similarly, the effective temperature of the surface is given by where SLR is the average flux of longwave radiation emitted by the surface. (OLR is a conventional abbreviation. SLR is used here to denote the flux of surface-emitted longwave radiation, although there is no standard abbreviation for this.) The IPCC reports the greenhouse effect, G, as being 159 W m, where G is the flux of longwave thermal radiation that leaves the surface minus the flux of outgoing longwave radiation that reaches space: Alternatively, the greenhouse effect can be described using the normalized greenhouse effect, g̃, defined as The normalized greenhouse effect is the fraction of the amount of thermal radiation emitted by the surface that does not reach space. Based on the IPCC numbers, g̃ = 0.40. In other words, 40 percent less thermal radiation reaches space than what leaves the surface. Sometimes the greenhouse effect is quantified as a temperature difference. This temperature difference is closely related to the quantities above. When the greenhouse effect is expressed as a temperature difference, Δ T G H E {\displaystyle \Delta T_{\mathrm {GHE} }} , this refers to the effective temperature associated with thermal radiation emissions from the surface minus the effective temperature associated with emissions to space: Informal discussions of the greenhouse effect often compare the actual surface temperature to the temperature that the planet would have if there were no greenhouse gases. However, in formal technical discussions, when the size of the greenhouse effect is quantified as a temperature, this is generally done using the above formula. The formula refers to the effective surface temperature rather than the actual surface temperature, and compares the surface with the top of the atmosphere, rather than comparing reality to a hypothetical situation. The temperature difference, Δ T G H E {\displaystyle \Delta T_{\mathrm {GHE} }} , indicates how much warmer a planet's surface is than the planet's overall effective temperature. Earth's top-of-atmosphere (TOA) energy imbalance (EEI) is the amount by which the power of incoming radiation exceeds the power of outgoing radiation: where ASR is the mean flux of absorbed solar radiation. ASR may be expanded as where A {\displaystyle A} is the albedo (reflectivity) of the planet and MSI is the mean solar irradiance incoming at the top of the atmosphere. The radiative equilibrium temperature of a planet can be expressed as A planet's temperature will tend to shift towards a state of radiative equilibrium, in which the TOA energy imbalance is zero, i.e., E E I = 0 {\displaystyle \mathrm {EEI} =0} . When the planet is in radiative equilibrium, the overall effective temperature of the planet is given by Thus, the concept of radiative equilibrium is important because it indicates what effective temperature a planet will tend towards having. If, in addition to knowing the effective temperature, T e f f {\displaystyle T_{\mathrm {eff} }} , we know the value of the greenhouse effect, then we know the mean (average) surface temperature of the planet. This is why the quantity known as the greenhouse effect is important: it is one of the few quantities that go into determining the planet's mean surface temperature. Typically, a planet will be close to radiative equilibrium, with the rates of incoming and outgoing energy being well-balanced. Under such conditions, the planet's equilibrium temperature is determined by the mean solar irradiance and the planetary albedo (how much sunlight is reflected back to space instead of being absorbed). The greenhouse effect measures how much warmer the surface is than the overall effective temperature of the planet. So, the effective surface temperature, T s u r f a c e , e f f {\displaystyle T_{\mathrm {surface,eff} }} , is, using the definition of Δ T G H E {\displaystyle \Delta T_{\mathrm {GHE} }} , One could also express the relationship between T s u r f a c e , e f f {\displaystyle T_{\mathrm {surface,eff} }} and T e f f {\displaystyle T_{\mathrm {eff} }} using G or g̃. So, the principle that a larger greenhouse effect corresponds to a higher surface temperature, if everything else (i.e., the factors that determine T e f f {\displaystyle T_{\mathrm {eff} }} ) is held fixed, is true as a matter of definition. Note that the greenhouse effect influences the temperature of the planet as a whole, in tandem with the planet's tendency to move toward radiative equilibrium. There are sometimes misunderstandings about how the greenhouse effect functions and raises temperatures. The surface budget fallacy is a common error in thinking. It involves thinking that an increased CO2 concentration could only cause warming by increasing the downward thermal radiation to the surface, as a result of making the atmosphere a better emitter. If the atmosphere near the surface is already nearly opaque to thermal radiation, this would mean that increasing CO2 could not lead to higher temperatures. However, it is a mistake to focus on the surface energy budget rather than the top-of-atmosphere energy budget. Regardless of what happens at the surface, increasing the concentration of CO2 tends to reduce the thermal radiation reaching space (OLR), leading to a TOA energy imbalance that leads to warming. Earlier researchers like Callendar (1938) and Plass (1959) focused on the surface budget, but the work of Manabe in the 1960s clarified the importance of the top-of-atmosphere energy budget. Among those who do not believe in the greenhouse effect, there is a fallacy that the greenhouse effect involves greenhouse gases sending heat from the cool atmosphere to the planet's warm surface, in violation of the Second Law of Thermodynamics. However, this idea reflects a misunderstanding. Radiation heat flow is the net energy flow after the flows of radiation in both directions have been taken into account. Radiation heat flow occurs in the direction from the surface to the atmosphere and space, as is to be expected given that the surface is warmer than the atmosphere and space. While greenhouse gases emit thermal radiation downward to the surface, this is part of the normal process of radiation heat transfer. The downward thermal radiation simply reduces the upward thermal radiation net energy flow (radiation heat flow), i.e., it reduces cooling. Simplified models are sometimes used to support understanding of how the greenhouse effect comes about and how this affects surface temperature. The greenhouse effect can be seen to occur in a simplified model in which the air is treated as if it is single uniform layer exchanging radiation with the ground and space. Slightly more complex models add additional layers, or introduce convection. One simplification is to treat all outgoing longwave radiation as being emitted from an altitude where the air temperature equals the overall effective temperature for planetary emissions, T e f f {\displaystyle T_{\mathrm {eff} }} . Some authors have referred to this altitude as the effective radiating level (ERL), and suggest that as the CO2 concentration increases, the ERL must rise to maintain the same mass of CO2 above that level. This approach is less accurate than accounting for variation in radiation wavelength by emission altitude. However, it can be useful in supporting a simplified understanding of the greenhouse effect. For instance, it can be used to explain how the greenhouse effect increases as the concentration of greenhouse gases increase. Earth's overall equivalent emission altitude has been increasing with a trend of 23 m (75 ft)/decade, which is said to be consistent with a global mean surface warming of 0.12 °C (0.22 °F)/decade over the period 1979–2011. The existence of the greenhouse effect, while not named as such, was proposed as early as 1824 by Joseph Fourier. The argument and the evidence were further strengthened by Claude Pouillet in 1827 and 1838. In 1856 Eunice Newton Foote demonstrated that the warming effect of the sun is greater for air with water vapour than for dry air, and the effect is even greater with carbon dioxide. She concluded that "An atmosphere of that gas would give to our earth a high temperature..." John Tyndall was the first to measure the infrared absorption and emission of various gases and vapors. From 1859 onwards, he showed that the effect was due to a very small proportion of the atmosphere, with the main gases having no effect, and was largely due to water vapor, though small percentages of hydrocarbons and carbon dioxide had a significant effect. The effect was more fully quantified by Svante Arrhenius in 1896, who made the first quantitative prediction of global warming due to a hypothetical doubling of atmospheric carbon dioxide. The term greenhouse was first applied to this phenomenon by Nils Gustaf Ekholm in 1901. Scientists have observed that, at times, there is a negative greenhouse effect over parts of Antarctica. In a location where there is a strong temperature inversion, so that the air is warmer than the surface, it is possible for the greenhouse effect to be reversed, so that the presence of greenhouse gases increases the rate of radiative cooling to space. In this case, the rate of thermal radiation emission to space is greater than the rate at which thermal radiation is emitted by the surface. Thus, the local value of the greenhouse effect is negative. In the solar system, apart from the Earth, at least two other planets and a moon also have a greenhouse effect. The greenhouse effect on Venus is particularly large, and it brings the surface temperature to as high as 735 K (462 °C; 863 °F). This is due to its very dense atmosphere which consists of about 97% carbon dioxide. Although Venus is about 30% closer to the Sun, it absorbs (and is warmed by) less sunlight than Earth, because Venus reflects 77% of incident sunlight while Earth reflects around 30%. In the absence of a greenhouse effect, the surface of Venus would be expected to have a temperature of 232 K (−41 °C; −42 °F). Thus, contrary to what one might think, being nearer to the Sun is not a reason why Venus is warmer than Earth. Due to its high pressure, the CO2 in the atmosphere of Venus exhibits continuum absorption (absorption over a broad range of wavelengths) and is not limited to absorption within the bands relevant to its absorption on Earth. A runaway greenhouse effect involving carbon dioxide and water vapor has for many years been hypothesized to have occurred on Venus; this idea is still largely accepted. The planet Venus experienced a runaway greenhouse effect, resulting in an atmosphere which is 96% carbon dioxide, and a surface atmospheric pressure roughly the same as found 900 m (3,000 ft) underwater on Earth. Venus may have had water oceans, but they would have boiled off as the mean surface temperature rose to the current 735 K (462 °C; 863 °F). Mars has about 70 times as much carbon dioxide as Earth, but experiences only a small greenhouse effect, about 6 K (11 °F). The greenhouse effect is small due to the lack of water vapor and the overall thinness of the atmosphere. The same radiative transfer calculations that predict warming on Earth accurately explain the temperature on Mars, given its atmospheric composition. Saturn's moon Titan has both a greenhouse effect and an anti-greenhouse effect. The presence of nitrogen (N2), methane (CH4), and hydrogen (H2) in the atmosphere contribute to a greenhouse effect, increasing the surface temperature by 21 K (38 °F) over the expected temperature of the body without these gases. While the gases N2 and H2 ordinarily do not absorb infrared radiation, these gases absorb thermal radiation on Titan due to pressure-induced collisions, the large mass and thickness of the atmosphere, and the long wavelengths of the thermal radiation from the cold surface. The existence of a high-altitude haze, which absorbs wavelengths of solar radiation but is transparent to infrared, contribute to an anti-greenhouse effect of approximately 9 K (16 °F). The net result of these two effects is a warming of 21 K − 9 K = 12 K (22 °F), so Titan's surface temperature of 94 K (−179 °C; −290 °F) is 12 K warmer than it would be if there were no atmosphere. One cannot predict the relative sizes of the greenhouse effects on different bodies simply by comparing the amount of greenhouse gases in their atmospheres. This is because factors other than the quantity of these gases also play a role in determining the size of the greenhouse effect. Overall atmospheric pressure affects how much thermal radiation each molecule of a greenhouse gas can absorb. High pressure leads to more absorption and low pressure leads to less. This is due to "pressure broadening" of spectral lines. When the total atmospheric pressure is higher, collisions between molecules occur at a higher rate. Collisions broaden the width of absorption lines, allowing a greenhouse gas to absorb thermal radiation over a broader range of wavelengths. Each molecule in the air near Earth's surface experiences about 7 billion collisions per second. This rate is lower at higher altitudes, where the pressure and temperature are both lower. This means that greenhouse gases are able to absorb more wavelengths in the lower atmosphere than they can in the upper atmosphere. On other planets, pressure broadening means that each molecule of a greenhouse gas is more effective at trapping thermal radiation if the total atmospheric pressure is high (as on Venus), and less effective at trapping thermal radiation if the atmospheric pressure is low (as on Mars).
[[Image:Radioactive.svg|thumb|Ionising radiation hazard symbol]] [[File:Logo iso radiation.svg|thumb|right|2007 ISO [[radioactivity]] danger logo. This logo was designed in part for long-term [[radioactive waste]] depositories which might survive into a far future time in which all knowledge of the meaning of present common radiation danger symbols and signs will have been lost.]] '''Ionizing radiation''' is a process in [[physics]] where something sends out [[particle]]s or [[electromagnetic radiation|waves]] that can [[ion]]ize an [[atom]] or [[molecule]] through atomic interactions. The strength of ionizing radiation depends on the energy of the individual particles or waves, and not a function of the number of particles or waves present. The material that the radiation can ionize depends on the [[ionization energy]] of the material. [[Ionization|Ionizing]] radiation can be [[electromagnetic radiation]] or [[subatomic particle]]s. Electromagnetic: * [[Gamma ray]]s are electromagnetic waves of highest energy. * [[X-ray]]s are less energetic. * [[Ultraviolet]] radiation only ionizes some materials. Subatomic particle radiation includes: * [[Alpha particle]] radiation, made of [[helium]] [[nucleus (physics)|nuclei]] * [[Beta particle]] radiation, made of energetic [[electron]]s or [[positrons]] * Neutron radiation, made of [[neutron]]s Some radiation can go through the [[human body]] and other objects. Usually when people use the term radiation, they are talking specifically about potentially harmful types of ionizing radiation. If something produces this sort of radiation, it is said to be '''radioactive'''. There is a little radiation all around us all the time, which people's bodies are used to, but larger amounts of radiation can make people sick or kill them. Natural ionizing radiation is produced by [[radioactive decay]] of some [[chemical element]]s, such as [[uranium]]. [[Star]]s and other things in [[outer space]] also make radiation. ''See [[cosmic ray]]''. Some [[isotope]]s that are radioactive only stay radioactive for much less than a [[second]]. Other things can stay radioactive for thousands of years. People and anything that lives gives off radiation naturally, because of the [[potassium]] and [[Carbon-14]] inside. Some of the machines that make radiation are called [[particle accelerator]]s. Scientists use these machines to make radiation so they can study it. [[X-ray]] machines also make radiation, so doctors can see the inside of the [[human body]] and help people. [[Nuclear weapon|Nuclear weapons]] ([[Nuclear weapon|atomic weapons]]) use a nuclear reaction to produce massive amounts of energy, in the form of heat, light, and radiation. This radiation is spread by the dust, ash, and smoke produced by the blast. [[Nuclear reactor|Nuclear reactors]] are used to make [[electricity]]. They make a lot of radiation, but the reactors are built carefully to keep the radiation inside the reactor. But many people are afraid that if there were a problem with the reactor, radioactive material could escape into the environment, harming or killing many animals and people. Also, the parts of the reactor stay radioactive, and can kill people, for hundreds or thousands of years, so people are not sure where they can keep parts of old reactors safely away from people. == Uses == [[X-ray]] machines also make radiation, so doctors can see the inside of the [[human body]] and help people. [[Nuclear weapon]]s ([[atomic weapon]]s) use a nuclear reaction to produce massive amounts of energy, in the form of heat, light, and radiation. [[Nuclear reactor]]s are used to make [[electricity]]. They make a lot of radiation, but the reactors are built carefully to keep the radiation inside the reactor. Alpha radiation is used in [[Static eliminator|static eliminators]] and [[Smoke detector|smoke detectors]]. Ionizing radiation are used to [[Sterilisation|sterilize]] [[Surgical instrument|medical instruments]]. It is also used in [[radiation therapy]], [[food irradiation]] and the [[sterile insect technique]]. == Related pages == *[[Acute radiation syndrome]] *[[Dirty bomb]] *[[Radiology]] == Other websites == *[http://www.nrc.gov The Nuclear Regulatory Commission] regulates most commercial radiation sources and non-medical exposures in the US: *[http://www.belleonline.com/ Biological Effects of Low Level Exposures: Radiation Hormesis] {{Webarchive|url=https://web.archive.org/web/20070418212709/http://www.belleonline.com/ |date=2007-04-18 }} *[http://books.nap.edu/catalog/11340.html?onpi_newsdoc062905 Health Risks from Exposure to Low Levels of Ionizing Radiation: BEIR VII Phase 2] *[http://toxnet.nlm.nih.gov/cgi-bin/sis/search/r?dbs+hsdb:@term+@na+@rel+ionizing+radiation NLM Hazardous Substances Databank &ndash; Ionizing Radiation] *[http://www.riscrad.org RISC-RAD] {{Webarchive|url=https://web.archive.org/web/20080827220834/http://www.riscrad.org/ |date=2008-08-27 }} is a European research project on assessment of low dose cancer risk *[http://www.unscear.org/unscear/en/publications/2000_1.html UNSCEAR 2000 Report, Volume 1: Sources] *[http://www.unscear.org/unscear/en/publications/2000_2.html UNSCEAR 2000 Report, Volume 2: Effects] *[http://www.npl.co.uk/publications/ionising_radiation/ Beginners Guide to Ionising Radiation Measurement] {{Webarchive|url=https://web.archive.org/web/20070711081147/http://www.npl.co.uk/publications/ionising_radiation/ |date=2007-07-11 }} *[http://www.npl.co.uk/ionrad/quantities.html Quantities, units and their relationships] {{Webarchive|url=https://web.archive.org/web/20060923191122/http://www.npl.co.uk/ionrad/quantities.html |date=2006-09-23 }} *[http://www.ki4u.com/free_book/s60p792.htm Plans for homemade ionizing radiation meter] *[http://www.pocketrad.com/sources2.html List of common household radioactive items] {{Webarchive|url=https://web.archive.org/web/20110324040428/http://www.pocketrad.com/sources2.html |date=2011-03-24 }} *[http://en.citizendium.org/wiki/Ionizing_radiation Ionizing radiation] -Citizendium <!--Categories--> [[Category:Nuclear physics]] <!--Interwiki-->
Ionizing radiation (or ionising radiation), including nuclear radiation, consists of subatomic particles or electromagnetic waves that have sufficient energy to ionize atoms or molecules by detaching electrons from them. Some particles can travel up to 99% of the speed of light, and the electromagnetic waves are on the high-energy portion of the electromagnetic spectrum. Gamma rays, X-rays, and the higher energy ultraviolet part of the electromagnetic spectrum are ionizing radiation, whereas the lower energy ultraviolet, visible light, nearly all types of laser light, infrared, microwaves, and radio waves are non-ionizing radiation. The boundary between ionizing and non-ionizing radiation in the ultraviolet area cannot be sharply defined, as different molecules and atoms ionize at different energies. The energy of ionizing radiation starts between 10 electronvolts (eV) and 33 eV. Typical ionizing subatomic particles include alpha particles, beta particles, and neutrons. These are typically created by radioactive decay, and almost all are energetic enough to ionize. There are also secondary cosmic particles produced after cosmic rays interact with Earth's atmosphere, including muons, mesons, and positrons. Cosmic rays may also produce radioisotopes on Earth (for example, carbon-14), which in turn decay and emit ionizing radiation. Cosmic rays and the decay of radioactive isotopes are the primary sources of natural ionizing radiation on Earth, contributing to background radiation. Ionizing radiation is also generated artificially by X-ray tubes, particle accelerators, and nuclear fission. Ionizing radiation is not immediately detectable by human senses, so instruments such as Geiger counters are used to detect and measure it. However, very high energy particles can produce visible effects on both organic and inorganic matter (e.g. water lighting in Cherenkov radiation) or humans (e.g. acute radiation syndrome). Ionizing radiation is used in a wide variety of fields such as medicine, nuclear power, research, and industrial manufacturing, but presents a health hazard if proper measures against excessive exposure are not taken. Exposure to ionizing radiation causes cell damage to living tissue and organ damage. In high acute doses, it will result in radiation burns and radiation sickness, and lower level doses over a protracted time can cause cancer. The International Commission on Radiological Protection (ICRP) issues guidance on ionizing radiation protection, and the effects of dose uptake on human health. Ionizing radiation may be grouped as directly or indirectly ionizing. Any charged particle with mass can ionize atoms directly by fundamental interaction through the Coulomb force if it carries sufficient kinetic energy. Such particles include atomic nuclei, electrons, muons, charged pions, protons, and energetic charged nuclei stripped of their electrons. When moving at relativistic speeds (near the speed of light, c) these particles have enough kinetic energy to be ionizing, but there is considerable speed variation. For example, a typical alpha particle moves at about 5% of c, but an electron with 33 eV (just enough to ionize) moves at about 1% of c. Two of the first types of directly ionizing radiation to be discovered are alpha particles which are helium nuclei ejected from the nucleus of an atom during radioactive decay, and energetic electrons, which are called beta particles. Natural cosmic rays are made up primarily of relativistic protons but also include heavier atomic nuclei like helium ions and HZE ions. In the atmosphere such particles are often stopped by air molecules, and this produces short-lived charged pions, which soon decay to muons, a primary type of cosmic ray radiation that reaches the surface of the earth. Pions can also be produced in large amounts in particle accelerators. Alpha particles consist of two protons and two neutrons bound together into a particle identical to a helium nucleus. Alpha particle emissions are generally produced in the process of alpha decay. Alpha particles are a strongly ionizing form of radiation, but when emitted by radioactive decay they have low penetration power and can be absorbed by a few centimeters of air, or by the top layer of human skin. More powerful alpha particles from ternary fission are three times as energetic, and penetrate proportionately farther in air. The helium nuclei that form 10–12% of cosmic rays, are also usually of much higher energy than those produced by radioactive decay and pose shielding problems in space. However, this type of radiation is significantly absorbed by the Earth's atmosphere, which is a radiation shield equivalent to about 10 meters of water. The alpha particle was named by Ernest Rutherford after the first letter in the Greek alphabet, α, when he ranked the known radioactive emissions in descending order of ionising effect in 1899. The symbol is α or α. Because they are identical to helium nuclei, they are also sometimes written as He or 2He indicating a Helium ion with a +2 charge (missing its two electrons). If the ion gains electrons from its environment, the alpha particle can be written as a normal (electrically neutral) helium atom 2He. Beta particles are high-energy, high-speed electrons or positrons emitted by certain types of radioactive nuclei, such as potassium-40. The production of beta particles is termed beta decay. They are designated by the Greek letter beta (β). There are two forms of beta decay, β and β, which respectively give rise to the electron and the positron. Beta particles are less penetrating than gamma radiation, but more penetrating than alpha particles. High-energy beta particles may produce X-rays known as bremsstrahlung ("braking radiation") or secondary electrons (delta ray) as they pass through matter. Both of these can cause an indirect ionization effect. Bremsstrahlung is of concern when shielding beta emitters, as the interaction of beta particles with some shielding materials produces Bremsstrahlung. The effect is greater with material having high atomic numbers, so material with low atomic numbers is used for beta source shielding. The positron or antielectron is the antiparticle or the antimatter counterpart of the electron. When a low-energy positron collides with a low-energy electron, annihilation occurs, resulting in their conversion into the energy of two or more gamma ray photons (see electron–positron annihilation). As positrons are positively charged particles they can directly ionize an atom through Coulomb interactions. Positrons can be generated by positron emission nuclear decay (through weak interactions), or by pair production from a sufficiently energetic photon. Positrons are common artificial sources of ionizing radiation used in medical positron emission tomography (PET) scans. Charged nuclei are characteristic of galactic cosmic rays and solar particle events and except for alpha particles (charged helium nuclei) have no natural sources on earth. In space, however, very high energy protons, helium nuclei, and HZE ions can be initially stopped by relatively thin layers of shielding, clothes, or skin. However, the resulting interaction will generate secondary radiation and cause cascading biological effects. If just one atom of tissue is displaced by an energetic proton, for example, the collision will cause further interactions in the body. This is called "linear energy transfer" (LET), which utilizes elastic scattering. LET can be visualized as a billiard ball hitting another in the manner of the conservation of momentum, sending both away with the energy of the first ball divided between the two unequally. When a charged nucleus strikes a relatively slow-moving nucleus of an object in space, LET occurs and neutrons, alpha particles, low-energy protons, and other nuclei will be released by the collisions and contribute to the total absorbed dose of tissue. Indirectly ionizing radiation is electrically neutral and does not interact strongly with matter, therefore the bulk of the ionization effects are due to secondary ionization. Even though photons are electrically neutral, they can ionize atoms indirectly through the photoelectric effect and the Compton effect. Either of those interactions will cause the ejection of an electron from an atom at relativistic speeds, turning that electron into a beta particle (secondary beta particle) that will ionize other atoms. Since most of the ionized atoms are due to the secondary beta particles, photons are indirectly ionizing radiation. Radiated photons are called gamma rays if they are produced by a nuclear reaction, subatomic particle decay, or radioactive decay within the nucleus. They are called x-rays if produced outside the nucleus. The generic term "photon" is used to describe both. X-rays normally have a lower energy than gamma rays, and an older convention was to define the boundary as a wavelength of 10 m (or a photon energy of 100 keV). That threshold was driven by historic limitations of older X-ray tubes and low awareness of isomeric transitions. Modern technologies and discoveries have shown an overlap between X-ray and gamma energies. In many fields they are functionally identical, differing for terrestrial studies only in origin of the radiation. In astronomy, however, where radiation origin often cannot be reliably determined, the old energy division has been preserved, with X-rays defined as being between about 120 eV and 120 keV, and gamma rays as being of any energy above 100 to 120 keV, regardless of source. Most astronomical "gamma-ray astronomy" are known not to originate in nuclear radioactive processes but, rather, result from processes like those that produce astronomical X-rays, except driven by much more energetic electrons. Photoelectric absorption is the dominant mechanism in organic materials for photon energies below 100 keV, typical of classical X-ray tube originated X-rays. At energies beyond 100 keV, photons ionize matter increasingly through the Compton effect, and then indirectly through pair production at energies beyond 5 MeV. The accompanying interaction diagram shows two Compton scatterings happening sequentially. In every scattering event, the gamma ray transfers energy to an electron, and it continues on its path in a different direction and with reduced energy. The lowest ionization energy of any element is 3.89 eV, for caesium. However, US Federal Communications Commission material defines ionizing radiation as that with a photon energy greater than 10 eV (equivalent to a far ultraviolet wavelength of 124 nanometers). Roughly, this corresponds to both the first ionization energy of oxygen, and the ionization energy of hydrogen, both about 14 eV. In some Environmental Protection Agency references, the ionization of a typical water molecule at an energy of 33 eV is referenced as the appropriate biological threshold for ionizing radiation: this value represents the so-called W-value, the colloquial name for the ICRU's mean energy expended in a gas per ion pair formed, which combines ionization energy plus the energy lost to other processes such as excitation. At 38 nanometers wavelength for electromagnetic radiation, 33 eV is close to the energy at the conventional 10 nm wavelength transition between extreme ultraviolet and X-ray radiation, which occurs at about 125 eV. Thus, X-ray radiation is always ionizing, but only extreme-ultraviolet radiation can be considered ionizing under all definitions. Neutrons have a neutral electrical charge often misunderstood as zero electrical charge and thus often do not directly cause ionization in a single step or interaction with matter. However, fast neutrons will interact with the protons in hydrogen via linear energy transfer, energy that a particle transfers to the material it is moving through. This mechanism scatters the nuclei of the materials in the target area, causing direct ionization of the hydrogen atoms. When neutrons strike the hydrogen nuclei, proton radiation (fast protons) results. These protons are themselves ionizing because they are of high energy, are charged, and interact with the electrons in matter. Neutrons that strike other nuclei besides hydrogen will transfer less energy to the other particle if linear energy transfer does occur. But, for many nuclei struck by neutrons, inelastic scattering occurs. Whether elastic or inelastic scatter occurs is dependent on the speed of the neutron, whether fast or thermal or somewhere in between. It is also dependent on the nuclei it strikes and its neutron cross section. In inelastic scattering, neutrons are readily absorbed in a type of nuclear reaction called neutron capture and attributes to the neutron activation of the nucleus. Neutron interactions with most types of matter in this manner usually produce radioactive nuclei. The abundant oxygen-16 nucleus, for example, undergoes neutron activation, rapidly decays by a proton emission forming nitrogen-16, which decays to oxygen-16. The short-lived nitrogen-16 decay emits a powerful beta ray. This process can be written as: O (n,p) N (fast neutron capture possible with >11 MeV neutron) N → O + β (Decay t1/2 = 7.13 s) This high-energy β further interacts rapidly with other nuclei, emitting high-energy γ via Bremsstrahlung While not a favorable reaction, the O (n,p) N reaction is a major source of X-rays emitted from the cooling water of a pressurized water reactor and contributes enormously to the radiation generated by a water-cooled nuclear reactor while operating. For the best shielding of neutrons, hydrocarbons that have an abundance of hydrogen are used. In fissile materials, secondary neutrons may produce nuclear chain reactions, causing a larger amount of ionization from the daughter products of fission. Outside the nucleus, free neutrons are unstable and have a mean lifetime of 14 minutes, 42 seconds. Free neutrons decay by emission of an electron and an electron antineutrino to become a proton, a process known as beta decay: In the adjacent diagram, a neutron collides with a proton of the target material, and then becomes a fast recoil proton that ionizes in turn. At the end of its path, the neutron is captured by a nucleus in an (n,γ)-reaction that leads to the emission of a neutron capture photon. Such photons always have enough energy to qualify as ionizing radiation. Neutron radiation, alpha radiation, and extremely energetic gamma (> ~20 MeV) can cause nuclear transmutation and induced radioactivity. The relevant mechanisms are neutron activation, alpha absorption, and photodisintegration. A large enough number of transmutations can change macroscopic properties and cause targets to become radioactive themselves, even after the original source is removed. Ionization of molecules can lead to radiolysis (breaking chemical bonds), and formation of highly reactive free radicals. These free radicals may then react chemically with neighbouring materials even after the original radiation has stopped. (e.g., ozone cracking of polymers by ozone formed by ionization of air). Ionizing radiation can also accelerate existing chemical reactions such as polymerization and corrosion, by contributing to the activation energy required for the reaction. Optical materials deteriorate under the effect of ionizing radiation. High-intensity ionizing radiation in air can produce a visible ionized air glow of telltale bluish-purple color. The glow can be observed, e.g., during criticality accidents, around mushroom clouds shortly after a nuclear explosion, or the inside of a damaged nuclear reactor like during the Chernobyl disaster. Monatomic fluids, e.g. molten sodium, have no chemical bonds to break and no crystal lattice to disturb, so they are immune to the chemical effects of ionizing radiation. Simple diatomic compounds with very negative enthalpy of formation, such as hydrogen fluoride will reform rapidly and spontaneously after ionization. Ionization of materials temporarily increases their conductivity, potentially permitting damaging current levels. This is a particular hazard in semiconductor microelectronics employed in electronic equipment, with subsequent currents introducing operation errors or even permanently damaging the devices. Devices intended for high radiation environments such as the nuclear industry and extra-atmospheric (space) applications may be made radiation hard to resist such effects through design, material selection, and fabrication methods. Proton radiation found in space can also cause single-event upsets in digital circuits. The electrical effects of ionizing radiation are exploited in gas-filled radiation detectors, e.g. the Geiger-Muller counter or the ion chamber. Most adverse health effects of exposure to ionizing radiation may be grouped in two general categories: The most common impact is stochastic induction of cancer with a latent period of years or decades after exposure. For example, ionizing radiation is one cause of chronic myelogenous leukemia, although most people with CML have not been exposed to radiation. The mechanism by which this occurs is well understood, but quantitative models predicting the level of risk remain controversial. The most widely accepted model, the Linear no-threshold model (LNT), holds that the incidence of cancers due to ionizing radiation increases linearly with effective radiation dose at a rate of 5.5% per sievert. If this is correct, then natural background radiation is the most hazardous source of radiation to general public health, followed by medical imaging as a close second. Other stochastic effects of ionizing radiation are teratogenesis, cognitive decline, and heart disease. Although DNA is always susceptible to damage by ionizing radiation, the DNA molecule may also be damaged by radiation with enough energy to excite certain molecular bonds to form pyrimidine dimers. This energy may be less than ionizing, but near to it. A good example is ultraviolet spectrum energy which begins at about 3.1 eV (400 nm) at close to the same energy level which can cause sunburn to unprotected skin, as a result of photoreactions in collagen and (in the UV-B range) also damage in DNA (for example, pyrimidine dimers). Thus, the mid and lower ultraviolet electromagnetic spectrum is damaging to biological tissues as a result of electronic excitation in molecules which falls short of ionization, but produces similar non-thermal effects. To some extent, visible light and also ultraviolet A (UVA) which is closest to visible energies, have been proven to result in formation of reactive oxygen species in skin, which cause indirect damage since these are electronically excited molecules which can inflict reactive damage, although they do not cause sunburn (erythema). Like ionization-damage, all these effects in skin are beyond those produced by simple thermal effects. The table below shows radiation and dose quantities in SI and non-SI units. Ionizing radiation has many industrial, military, and medical uses. Its usefulness must be balanced with its hazards, a compromise that has shifted over time. For example, at one time, assistants in shoe shops in the US used X-rays to check a child's shoe size, but this practice was halted when the risks of ionizing radiation were better understood. Neutron radiation is essential to the working of nuclear reactors and nuclear weapons. The penetrating power of x-ray, gamma, beta, and positron radiation is used for medical imaging, nondestructive testing, and a variety of industrial gauges. Radioactive tracers are used in medical and industrial applications, as well as biological and radiation chemistry. Alpha radiation is used in static eliminators and smoke detectors. The sterilizing effects of ionizing radiation are useful for cleaning medical instruments, food irradiation, and the sterile insect technique. Measurements of carbon-14, can be used to date the remains of long-dead organisms (such as wood that is thousands of years old). Ionizing radiation is generated through nuclear reactions, nuclear decay, by very high temperature, or via acceleration of charged particles in electromagnetic fields. Natural sources include the sun, lightning and supernova explosions. Artificial sources include nuclear reactors, particle accelerators, and x-ray tubes. The United Nations Scientific Committee on the Effects of Atomic Radiation (UNSCEAR) itemized types of human exposures. The International Commission on Radiological Protection manages the International System of Radiological Protection, which sets recommended limits for dose uptake. Background radiation comes from both natural and human-made sources. The global average exposure of humans to ionizing radiation is about 3 mSv (0.3 rem) per year, 80% of which comes from nature. The remaining 20% results from exposure to human-made radiation sources, primarily from medical imaging. Average human-made exposure is much higher in developed countries, mostly due to CT scans and nuclear medicine. Natural background radiation comes from five primary sources: cosmic radiation, solar radiation, external terrestrial sources, radiation in the human body, and radon. The background rate for natural radiation varies considerably with location, being as low as 1.5 mSv/a (1.5 mSv per year) in some areas and over 100 mSv/a in others. The highest level of purely natural radiation recorded on the Earth's surface is 90 µGy/h (0.8 Gy/a) on a Brazilian black beach composed of monazite. The highest background radiation in an inhabited area is found in Ramsar, primarily due to naturally radioactive limestone used as a building material. Some 2000 of the most exposed residents receive an average radiation dose of 10 mGy per year, (1 rad/yr) ten times more than the ICRP recommended limit for exposure to the public from artificial sources. Record levels were found in a house where the effective radiation dose due to external radiation was 135 mSv/a, (13.5 rem/yr) and the committed dose from radon was 640 mSv/a (64.0 rem/yr). This unique case is over 200 times higher than the world average background radiation. Despite the high levels of background radiation that the residents of Ramsar receive there is no compelling evidence that they experience a greater health risk. The ICRP recommendations are conservative limits and may represent an over representation of the actual health risk. Generally radiation safety organization recommend the most conservative limits assuming it is best to err on the side of caution. This level of caution is appropriate but should not be used to create fear about background radiation danger. Radiation danger from background radiation may be a serious threat but is more likely a small overall risk compared to all other factors in the environment. The Earth, and all living things on it, are constantly bombarded by radiation from outside our solar system. This cosmic radiation consists of relativistic particles: positively charged nuclei (ions) from 1 amu protons (about 85% of it) to 26 amu iron nuclei and even beyond. (The high-atomic number particles are called HZE ions.) The energy of this radiation can far exceed that which humans can create, even in the largest particle accelerators (see ultra-high-energy cosmic ray). This radiation interacts in the atmosphere to create secondary radiation that rains down, including x-rays, muons, protons, antiprotons, alpha particles, pions, electrons, positrons, and neutrons. The dose from cosmic radiation is largely from muons, neutrons, and electrons, with a dose rate that varies in different parts of the world and based largely on the geomagnetic field, altitude, and solar cycle. The cosmic-radiation dose rate on airplanes is so high that, according to the United Nations UNSCEAR 2000 Report (see links at bottom), airline flight crew workers receive more dose on average than any other worker, including those in nuclear power plants. Airline crews receive more cosmic rays if they routinely work flight routes that take them close to the North or South pole at high altitudes, where this type of radiation is maximal. Cosmic rays also include high-energy gamma rays, which are far beyond the energies produced by solar or human sources. Most materials on Earth contain some radioactive atoms, even if in small quantities. Most of the dose received from these sources is from gamma-ray emitters in building materials, or rocks and soil when outside. The major radionuclides of concern for terrestrial radiation are isotopes of potassium, uranium, and thorium. Each of these sources has been decreasing in activity since the formation of the Earth. All earthly materials that are the building blocks of life contain a radioactive component. As humans, plants, and animals consume food, air, and water, an inventory of radioisotopes builds up within the organism (see banana equivalent dose). Some radionuclides, like potassium-40, emit a high-energy gamma ray that can be measured by sensitive electronic radiation measurement systems. These internal radiation sources contribute to an individual's total radiation dose from natural background radiation. An important source of natural radiation is radon gas, which seeps continuously from bedrock but can, because of its high density, accumulate in poorly ventilated houses. Radon-222 is a gas produced by the α-decay of radium-226. Both are a part of the natural uranium decay chain. Uranium is found in soil throughout the world in varying concentrations. Radon is the largest cause of lung cancer among non-smokers and the second-leading cause overall. There are three standard ways to limit exposure: These can all be applied to natural and human-made sources. For human-made sources the use of Containment is a major tool in reducing dose uptake and is effectively a combination of shielding and isolation from the open environment. Radioactive materials are confined in the smallest possible space and kept out of the environment such as in a hot cell (for radiation) or glove box (for contamination). Radioactive isotopes for medical use, for example, are dispensed in closed handling facilities, usually gloveboxes, while nuclear reactors operate within closed systems with multiple barriers that keep the radioactive materials contained. Work rooms, hot cells and gloveboxes have slightly reduced air pressures to prevent escape of airborne material to the open environment. In nuclear conflicts or civil nuclear releases civil defense measures can help reduce exposure of populations by reducing ingestion of isotopes and occupational exposure. One is the issue of potassium iodide (KI) tablets, which blocks the uptake of radioactive iodine (one of the major radioisotope products of nuclear fission) into the human thyroid gland. Occupationally exposed individuals are controlled within the regulatory framework of the country they work in, and in accordance with any local nuclear licence constraints. These are usually based on the recommendations of the International Commission on Radiological Protection. The ICRP recommends limiting artificial irradiation. For occupational exposure, the limit is 50 mSv in a single year with a maximum of 100 mSv in a consecutive five-year period. The radiation exposure of these individuals is carefully monitored with the use of dosimeters and other radiological protection instruments which will measure radioactive particulate concentrations, area gamma dose readings and radioactive contamination. A legal record of dose is kept. Examples of activities where occupational exposure is a concern include: Some human-made radiation sources affect the body through direct radiation, known as effective dose (radiation) while others take the form of radioactive contamination and irradiate the body from within. The latter is known as committed dose. Medical procedures, such as diagnostic X-rays, nuclear medicine, and radiation therapy are by far the most significant source of human-made radiation exposure to the general public. Some of the major radionuclides used are I-131, Tc-99m, Co-60, Ir-192, and Cs-137. The public is also exposed to radiation from consumer products, such as tobacco (polonium-210), combustible fuels (gas, coal, etc.), televisions, luminous watches and dials (tritium), airport X-ray systems, smoke detectors (americium), electron tubes, and gas lantern mantles (thorium). Of lesser magnitude, members of the public are exposed to radiation from the nuclear fuel cycle, which includes the entire sequence from processing uranium to the disposal of the spent fuel. The effects of such exposure have not been reliably measured due to the extremely low doses involved. Opponents use a cancer per dose model to assert that such activities cause several hundred cases of cancer per year, an application of the widely accepted Linear no-threshold model (LNT). The International Commission on Radiological Protection recommends limiting artificial irradiation to the public to an average of 1 mSv (0.001 Sv) of effective dose per year, not including medical and occupational exposures. In a nuclear war, gamma rays from both the initial weapon explosion and fallout would be the sources of radiation exposure. Massive particles are a concern for astronauts outside the Earth's magnetic field who would receive solar particles from solar proton events (SPE) and galactic cosmic rays from cosmic sources. These high-energy charged nuclei are blocked by Earth's magnetic field but pose a major health concern for astronauts traveling to the Moon and to any distant location beyond the Earth orbit. Highly charged HZE ions in particular are known to be extremely damaging, although protons make up the vast majority of galactic cosmic rays. Evidence indicates past SPE radiation levels that would have been lethal for unprotected astronauts. Air travel exposes people on aircraft to increased radiation from space as compared to sea level, including cosmic rays and from solar flare events. Software programs such as Epcard, CARI, SIEVERT, PCAIRE are attempts to simulate exposure by aircrews and passengers. An example of a measured dose (not simulated dose) is 6 μSv per hour from London Heathrow to Tokyo Narita on a high-latitude polar route. However, dosages can vary, such as during periods of high solar activity. The United States FAA requires airlines to provide flight crew with information about cosmic radiation, and an International Commission on Radiological Protection recommendation for the general public is no more than 1 mSv per year. In addition, many airlines do not allow pregnant flightcrew members, to comply with a European Directive. The FAA has a recommended limit of 1 mSv total for a pregnancy, and no more than 0.5 mSv per month. Information originally based on Fundamentals of Aerospace Medicine published in 2008. Hazardous levels of ionizing radiation are signified by the trefoil sign on a yellow background. These are usually posted at the boundary of a radiation controlled area or in any place where radiation levels are significantly above background due to human intervention. The red ionizing radiation warning symbol (ISO 21482) was launched in 2007, and is intended for IAEA Category 1, 2 and 3 sources defined as dangerous sources capable of death or serious injury, including food irradiators, teletherapy machines for cancer treatment and industrial radiography units. The symbol is to be placed on the device housing the source, as a warning not to dismantle the device or to get any closer. It will not be visible under normal use, only if someone attempts to disassemble the device. The symbol will not be located on building access doors, transportation packages or containers.
{{Infobox person | name = Buzz Aldrin | image = Buzz Aldrin (S69-31743).jpg | caption = Aldrin in July 1969 | birth_name = Edwin Eugene Aldrin, Jr. | birth_date = {{birth date and age|1930|01|20}} | birth_place = [[Glen Ridge, New Jersey]], [[United States|U.S.]] | nationality = American | other_names = Dr. Rendezvous | education = [[Montclair High School]] | alma_mater = {{Plainlist| * [[United States Military Academy]], B.S. 1951 * [[Massachusetts Institute of Technology]], Sc.D. 1963 }} | occupation = {{hlist|[[Fighter pilot]]|[[astronaut]]}} <!-- | years_active = 1955–1971 --> | known_for = Second man on the [[Moon]] | spouse = {{Plainlist| * {{Marriage|Joan Ann Archer|1954|1974|reason=divorced}} * {{Marriage|Beverly Van Zile|1975|1978|reason=divorced}} * {{Marriage|Lois Driggs Cannon|1988|2012|reason=divorced}} * {{marriage|Anca V Faur|2023}} }} | children = 3 | mother = Marion Aldrin ([[née]] Moon) | father = Edwin Eugene Aldrin Sr. | module = {{Infobox astronaut |name = Space carrer |type = [[NASA]] [[astronaut]] |rank = [[File:US-O6 insignia.svg|30px]] [[Colonel (United States)|Colonel]], [[United States Air Force]] |time = 12 days, 1 hour and 53 minutes <!-- Another source says 12.08 days (12 days 1 hour and 55.2 minutes) <ref>{{Cite web |url=http://www.astronautix.com/a/aldrin.html |title=Aldrin, Edwin Eugene 'Buzz' |last=Wade |first=Mark |website=Encyclopedia Astronautica |archive-url=https://web.archive.org/web/20180525132518/http://www.astronautix.com/a/aldrin.html|archive-date=May 25, 2018|url-status=live}}</ref> --> |selection = [[NASA Astronaut Group 3]] |eva1 = 4 |eva2 = 7 hours and 52 minutes |mission = {{hlist|[[Gemini 12]]|[[Apollo 11]]}} |insignia = [[File:Gemini 12 insignia.png|45px]] [[File:Apollo 11 insignia.png|50px]] |Date of ret = July 1, 1971 }} | website = {{url|www.buzzaldrin.com}} | signature = Buzz Aldrin Autograph.svg }} Dr. '''Edwin Eugene "Buzz" Aldrin, Jr.''', retired Colonel (born January 20, 1930) is an [[Americans|American]] [[pilot]] and [[astronaut]]. He was one of the [[Apollo 11]] astronauts and the [[List of people who have walked on the Moon|second person]] to have ever walked on the [[moon]]. ==Early life== Aldrin was born in [[Glen Ridge, New Jersey]]. He earned a [[Bachelor of Science|BSc]] in Mechanical Engineering from the military academy [[West Point]] in 1951. After postgraduate studies, he received a [[Doctor of Science|ScD]] in Aeronautics and Astronautics from the [[Massachusetts Institute of Technology]] (MIT) in 1963.From his earlier activity as a [[fighter ace|fighter pilot]] in the [[Korean War]] he won several medals, including the [[Air Force Distinguished Service Medal]]. ==Apollo 11 mission== Aldrin was the [[List of people who have walked on the Moon|second person]] in history to set foot on [[the Moon]] (after [[Neil Armstrong]], during the [[Apollo 11]] mission). The Lunar Lander Module landed on the Moon on July 20, 1969.<ref>Buzz Aldrin, Ken Abraham: (2009) "Magnificent Desolation", Random House Publishers, p. 26</ref> He was 39 when he set foot on the moon. As a [[Presbyterian]] elder, Aldrin was the first and only person to hold a religious ceremony on the Moon.<ref>{{cite web|url=https://www.ncronline.org/news/5-faith-facts-about-moon-landing-space-communion-and-prayer-league-its-own|title=5 faith facts about the moon landing: Space Communion and a prayer league of its own|publisher=NCR Online|accessdate=June 17, 2024}}</ref> Using a kit given to him by his pastor, he took [[Eucharist|communion]] during his moon mission.<ref>{{cite web |url=https://www.guideposts.org/better-living/life-advice/finding-life-purpose/guideposts-classics-buzz-aldrin-on-communion-in-space |series=Guideposts Classics |title=Buzz Aldrin on Communion in Space |date=July 10, 2014 |orig-year=1970 |work=Guideposts |last1=Aldrin |first1=Buzz |access-date=January 21, 2019 |archive-date=April 17, 2019 |archive-url=https://web.archive.org/web/20190417181815/https://www.guideposts.org/better-living/life-advice/finding-life-purpose/guideposts-classics-buzz-aldrin-on-communion-in-space |url-status=dead }}</ref> He received the [[Presidential Medal of Freedom]] and the [[NASA]] Exceptional Service Medal. After the death of [[Michael Collins (astronaut)|Michael Collins]] in 2021, Aldrin became the last surviving member of the Apollo 11 crew. == Later life == Aldrin left [[NASA]] in March 1972. He has been [[marriage|married]] four times: to Joan Archer, to Beverly Zile, to Lois Driggs Cannon, and to Anca V Faur.<ref>{{Cite news|date=2023-01-21|title=Buzz Aldrin marries for the fourth time, aged 93|language=en-GB|work=BBC News|url=https://www.bbc.com/news/world-us-canada-64359316|access-date=2023-01-22}}</ref> He had three children with Joan, named James, Janice, and Andrew. He married Cannon on [[Valentine's Day]], 1988. He is the oldest contestant on the television series ''[[Dancing with the Stars]]''. He competed in 2010. == References == {{wikiquote}} {{Commons category|Buzz Aldrin}} {{reflist}} {{DEFAULTSORT:Aldrin, Buzz}} [[Category:1930 births]] [[Category:Living people]] [[Category:American astronauts]] [[Category:American aviators]] [[Category:Apollo astronauts]] [[Category:People from Glen Ridge, New Jersey]] [[Category:Scientists from New Jersey]]
Buzz Aldrin (/ˈɔːldrɪn/; born Edwin Eugene Aldrin Jr.; January 20, 1930) is an American former astronaut, engineer and fighter pilot. He made three spacewalks as pilot of the 1966 Gemini 12 mission. He was the Lunar Module Eagle pilot on the 1969 Apollo 11 mission and became the second person to walk on the Moon after mission commander Neil Armstrong. Born in Glen Ridge, New Jersey, Aldrin graduated third in the class of 1951 from the United States Military Academy at West Point with a degree in mechanical engineering. He was commissioned into the United States Air Force and served as a jet fighter pilot during the Korean War. He flew 66 combat missions and shot down two MiG-15 aircraft. After earning a Doctor of Science degree in astronautics from the Massachusetts Institute of Technology (MIT), Aldrin was selected as a member of NASA's Astronaut Group 3, making him the first astronaut with a doctoral degree. His doctoral thesis, Line-of-Sight Guidance Techniques for Manned Orbital Rendezvous, earned him the nickname "Dr. Rendezvous" from fellow astronauts. His first space flight was in 1966 on Gemini 12, during which he spent over five hours on extravehicular activity. Three years later, Aldrin set foot on the Moon at 03:15:16 on July 21, 1969 (UTC), nineteen minutes after Armstrong first touched the surface, while command module pilot Michael Collins remained in lunar orbit. A Presbyterian elder, Aldrin became the first person to hold a religious ceremony on the Moon when he privately took communion. After leaving NASA in 1971, Aldrin became Commandant of the U.S. Air Force Test Pilot School. He retired from the Air Force in 1972 after 21 years of service. His autobiographies Return to Earth (1973) and Magnificent Desolation (2009) recount his struggles with clinical depression and alcoholism in the years after leaving NASA. Aldrin continues to advocate for space exploration, particularly a human mission to Mars. He developed the Aldrin cycler, a special spacecraft trajectory that makes travel to Mars more efficient in terms of time and propellant. He has been accorded numerous honors, including the Presidential Medal of Freedom in 1969. Aldrin was born Edwin Eugene Aldrin Jr. on January 20, 1930, at Mountainside Hospital in Glen Ridge, New Jersey. His parents, Edwin Eugene Aldrin Sr. and Marion Aldrin (née Moon), lived in neighboring Montclair. His father was an Army aviator during World War I and the assistant commandant of the Army's test pilot school at McCook Field, Ohio, from 1919 to 1922, but left the Army in 1928 and became an executive at Standard Oil. Aldrin had two sisters: Madeleine, who was four years older, and Fay Ann, who was a year and a half older. His nickname, which became his legal first name in 1988, arose as a result of Fay's mispronouncing "brother" as "buzzer", which was then shortened to "Buzz". He was a Boy Scout, achieving the rank of Tenderfoot Scout. Aldrin did well in school, maintaining an A average. He played football and was the starting center for Montclair High School's undefeated 1946 state champion team. His father wanted him to go to the United States Naval Academy in Annapolis, Maryland, and enrolled him at nearby Severn School, a preparatory school for Annapolis, and even secured him a Naval Academy appointment from Albert W. Hawkes, one of the United States senators from New Jersey. Aldrin attended Severn School in 1946, but had other ideas about his future career. He suffered from seasickness and considered ships a distraction from flying airplanes. He faced down his father and told him to ask Hawkes to change the nomination to the United States Military Academy at West Point, New York. Aldrin entered West Point in 1947. He did well academically, finishing first in his class his plebe (first) year. Aldrin was also an excellent athlete, competing in pole vault for the academy track and field team. In 1950, he traveled with a group of West Point cadets to Japan and the Philippines to study the military government policies of Douglas MacArthur. During the trip, the Korean War broke out. On June 5, 1951, Aldrin graduated third in the class of 1951 with a Bachelor of Science degree in mechanical engineering. Among the top of his class, Aldrin had his choice of assignments. He chose the United States Air Force, which had become a separate service in 1947 while Aldrin was still at West Point and did not yet have its own academy. He was commissioned as a second lieutenant and underwent basic flight training in T-6 Texans at Bartow Air Base in Florida. His classmates included Sam Johnson, who later became a prisoner of war in Vietnam; the two became friends. At one point, Aldrin attempted a double Immelmann turn in a T-28 Trojan and suffered a grayout. He recovered in time to pull out at about 2,000 feet (610 m), averting what would have been a fatal crash. When Aldrin was deciding what sort of aircraft he should fly, his father advised him to choose bombers, because command of a bomber crew gave an opportunity to learn and hone leadership skills, which could open up better prospects for career advancement. Aldrin chose instead to fly fighters. He moved to Nellis Air Force Base in Las Vegas, where he learned to fly the F-80 Shooting Star and the F-86 Sabre. Like most jet fighter pilots of the era, he preferred the latter. In December 1952, Aldrin was assigned to the 16th Fighter-Interceptor Squadron, which was part of the 51st Fighter-Interceptor Wing. At the time it was based at Suwon Air Base, about 20 miles (32 km) south of Seoul, and was engaged in combat operations as part of the Korean War. During an acclimatization flight, his main fuel system froze at 100 percent power, which would have soon used up all his fuel. He was able to override the setting manually, but this required holding a button down, which in turn made it impossible to also use his radio. He barely managed to make it back under enforced radio silence. He flew 66 combat missions in F-86 Sabres in Korea and shot down two MiG-15 aircraft. The first MiG-15 he shot down was on May 14, 1953. Aldrin was flying about 5 miles (8.0 km) south of the Yalu River, when he saw two MiG-15 fighters below him. Aldrin opened fire on one of the MiGs, whose pilot may never have seen him coming. The June 8, 1953, issue of Life magazine featured gun camera footage taken by Aldrin of the pilot ejecting from his damaged aircraft. Aldrin's second aerial victory came on June 4, 1953, when he accompanied aircraft from the 39th Fighter-Interceptor Squadron in an attack on an airbase in North Korea. Their newer aircraft were faster than his and he had trouble keeping up. He then spotted a MiG approaching from above. This time, Aldrin and his opponent spotted each other at about the same time. They went through a series of scissor maneuvers, attempting to get behind the other. Aldrin was first to do so, but his gun sight jammed. He then manually sighted his gun and fired. He then had to pull out, as the two aircraft had gotten too low for the dogfight to continue. Aldrin saw the MiG's canopy open and the pilot eject, although Aldrin was uncertain whether there was sufficient time for a parachute to open. For his service in Korea, he was awarded two Distinguished Flying Crosses and three Air Medals. Aldrin's year-long tour ended in December 1953, by which time the fighting in Korea had ended. Aldrin was assigned as an aerial gunnery instructor at Nellis. In December 1954 he became an aide-de-camp to Brigadier General Don Z. Zimmerman, the Dean of Faculty at the nascent United States Air Force Academy, which opened in 1955. That same year, he graduated from the Squadron Officer School at Maxwell Air Force Base in Alabama. From 1956 to 1959 he flew F-100 Super Sabres equipped with nuclear weapons as a flight commander in the 22nd Fighter Squadron, 36th Fighter Wing, stationed at Bitburg Air Base in West Germany. Among his squadron colleagues was Ed White, who had been a year behind him at West Point. After White left West Germany to study for a master's degree at the University of Michigan in aeronautical engineering, he wrote to Aldrin encouraging him to do the same. Through the Air Force Institute of Technology, Aldrin enrolled as a graduate student at the Massachusetts Institute of Technology in 1959 intending to earn a master's degree. Richard Battin was the professor for his astrodynamics class. Two other USAF officers who later became astronauts, David Scott and Edgar Mitchell, took the course around this time. Another USAF officer, Charles Duke, also took the course and wrote his 1964 master's degree at MIT under the supervision of Laurence R. Young. Aldrin enjoyed the classwork and soon decided to pursue a doctorate instead. In January 1963, he earned a Sc.D. degree in astronautics. His doctoral thesis was Line-of-Sight Guidance Techniques for Manned Orbital Rendezvous, the dedication of which read: "In the hopes that this work may in some way contribute to their exploration of space, this is dedicated to the crew members of this country's present and future manned space programs. If only I could join them in their exciting endeavors!" Aldrin chose his doctoral thesis in the hope that it would help him be selected as an astronaut, although it meant foregoing test pilot training, which was a prerequisite at the time. After completing his doctorate Aldrin was assigned to the Gemini Target Office of the Air Force Space Systems Division in Los Angeles, working with the Lockheed Aircraft Corporation on enhancing the maneuver capabilities of the Agena target vehicle which was to be used by NASA's Project Gemini. He was then posted to the Space Systems Division's field office at NASA's Manned Spacecraft Center in Houston, where he was involved in integrating Department of Defense experiments into Project Gemini flights. Aldrin initially applied to join the astronaut corps when NASA's Astronaut Group 2 was selected in 1962. His application was rejected on the grounds that he was not a test pilot. Aldrin was aware of the requirement and asked for a waiver but the request was turned down. On May 15, 1963, NASA announced another round of selections, this time with the requirement that applicants had either test pilot experience or 1,000 hours of flying time in jet aircraft. Aldrin had over 2,500 hours of flying time, of which 2,200 was in jets. His selection as one of fourteen members of NASA's Astronaut Group 3 was announced on October 18, 1963. This made him the first astronaut with a doctoral degree which, combined with his expertise in orbital mechanics, earned him the nickname "Dr. Rendezvous" from his fellow astronauts. Although Aldrin was both the most educated and the rendezvous expert in the astronaut corps, he was aware that the nickname was not always intended as a compliment. Upon completion of initial training, each new astronaut was assigned a field of expertise; in Aldrin's case, it was mission planning, trajectory analysis, and flight plans. Jim Lovell and Aldrin were selected as the backup crew of Gemini 10, commander and pilot respectively. Backup crews usually became the prime crew of the third following mission, but the last scheduled mission in the program was Gemini 12. The February 28, 1966, deaths of the Gemini 9 prime crew, Elliot See and Charles Bassett, in an air crash, led to Lovell and Aldrin being moved up one mission to backup for Gemini 9, which put them in position as prime crew for Gemini 12. They were designated its prime crew on June 17, 1966, with Gordon Cooper and Gene Cernan as their backups. Initially, Gemini 12's mission objectives were uncertain. As the last scheduled mission, it was primarily intended to complete tasks that had not been successfully or fully carried out on earlier missions. While NASA had successfully performed rendezvous during Project Gemini, the gravity-gradient stabilization test on Gemini 11 was unsuccessful. NASA also had concerns about extravehicular activity (EVA). Cernan on Gemini 9 and Richard Gordon on Gemini 11 had suffered from fatigue carrying out tasks during EVA, but Michael Collins had a successful EVA on Gemini 10, which suggested that the order in which he had performed his tasks was an important factor. It therefore fell to Aldrin to complete Gemini's EVA goals. NASA formed a committee to give him a better chance of success. It dropped the test of the Air Force's astronaut maneuvering unit (AMU) that had given Gordon trouble on Gemini 11 so Aldrin could focus on EVA. NASA revamped the training program, opting for underwater training over parabolic flight. Aircraft flying a parabolic trajectory had given astronauts an experience of weightlessness in training, but there was a delay between each parabola which gave astronauts several minutes of rest. It also encouraged performing tasks quickly, whereas in space they had to be done slowly and deliberately. Training in a viscous, buoyant fluid gave a better simulation. NASA also placed additional handholds on the capsule, which were increased from nine on Gemini 9 to 44 on Gemini 12, and created workstations where he could anchor his feet. Gemini 12's main objectives were to rendezvous with a target vehicle, and fly the spacecraft and target vehicle together using gravity-gradient stabilization, perform docked maneuvers using the Agena propulsion system to change orbit, conduct a tethered stationkeeping exercise and three EVAs, and demonstrate an automatic reentry. Gemini 12 also carried 14 scientific, medical, and technological experiments. It was not a trailblazing mission; rendezvous from above had already been successfully performed by Gemini 9, and the tethered vehicle exercise by Gemini 11. Even gravity-gradient stabilization had been attempted by Gemini 11, albeit unsuccessfully. Gemini 12 was launched from Launch Complex 19 at Cape Canaveral on 20:46 UTC on November 11, 1966. The Gemini Agena Target Vehicle had been launched about an hour and a half before. The mission's first major objective was to rendezvous with this target vehicle. As the target and Gemini 12 capsule drew closer together, radar contact between the two deteriorated until it became unusable, forcing the crew to rendezvous manually. Aldrin used a sextant and rendezvous charts he helped create to give Lovell the right information to put the spacecraft in position to dock with the target vehicle. Gemini 12 achieved the fourth docking with an Agena target vehicle. The next task was to practice undocking and docking again. On undocking, one of the three latches caught, and Lovell had to use the Gemini's thrusters to free the spacecraft. Aldrin then docked again successfully a few minutes later. The flight plan then called for the Agena main engine to be fired to take the docked spacecraft into a higher orbit, but eight minutes after the Agena had been launched, it had suffered a loss of chamber pressure. The Mission and Flight Directors therefore decided not to risk the main engine. This would be the only mission objective that was not achieved. Instead, the Agena's secondary propulsion system was used to allow the spacecraft to view the solar eclipse of November 12, 1966, over South America, which Lovell and Aldrin photographed through the spacecraft windows. Aldrin performed three EVAs. The first was a standup EVA on November 12, in which the spacecraft door was opened and he stood up, but did not leave the spacecraft. The standup EVA mimicked some of the actions he would do during his free-flight EVA, so he could compare the effort expended between the two. It set an EVA record of two hours and twenty minutes. The next day Aldrin performed his free-flight EVA. He climbed across the newly installed hand-holds to the Agena and installed the cable needed for the gravity-gradient stabilization experiment. Aldrin performed numerous tasks, including installing electrical connectors and testing tools that would be needed for Project Apollo. A dozen two-minute rest periods prevented him from becoming fatigued. His second EVA concluded after two hours and six minutes. A third, 55-minute standup EVA was conducted on November 14, during which Aldrin took photographs, conducted experiments, and discarded some unneeded items. On November 15, the crew initiated the automatic reentry system and splashed down in the Atlantic Ocean, where they were picked up by a helicopter, which took them to the awaiting aircraft carrier USS Wasp. After the mission, his wife realized he had fallen into a depression, something she had not seen before. Lovell and Aldrin were assigned to an Apollo crew with Neil Armstrong as commander, Lovell as command module pilot (CMP), and Aldrin as lunar module pilot (LMP). Their assignment as the backup crew of Apollo 9 was announced on November 20, 1967. Due to design and manufacturing delays in the lunar module (LM), Apollo 8 and Apollo 9 swapped prime and backup crews, and Armstrong's crew became the backup for Apollo 8. Under the normal crew rotation scheme, Armstrong was expected to command Apollo 11. Michael Collins, the CMP on the Apollo 8 prime crew, required surgery to remove a bone spur on his spine. Lovell took his place on the Apollo 8 crew. When Collins recovered he joined Armstrong's crew as CMP. In the meantime, Fred Haise filled in as backup LMP, and Aldrin as backup CMP for Apollo 8. While the CMP usually occupied the center couch on launch, Aldrin occupied it rather than Collins, as he had already been trained to operate its console on liftoff before Collins arrived. Apollo 11 was the second American space mission made up entirely of astronauts who had already flown in space, the first being Apollo 10. The next would not be flown until STS-26 in 1988. Deke Slayton, who was responsible for astronaut flight assignments, gave Armstrong the option to replace Aldrin with Lovell, since some thought Aldrin was difficult to work with. Armstrong thought it over for a day before declining. He had no issues working with Aldrin, and thought Lovell deserved his own command. Early versions of the EVA checklist had the lunar module pilot as the first to step onto the lunar surface. However, when Aldrin learned that this might be amended, he lobbied within NASA for the original procedure to be followed. Multiple factors contributed to the final decision, including the physical positioning of the astronauts within the compact lunar lander, which made it easier for Armstrong to be the first to exit the spacecraft. Furthermore, there was little support for Aldrin's views among senior astronauts who would command later Apollo missions. Collins has commented that he thought Aldrin "resents not being first on the Moon more than he appreciates being second". Aldrin and Armstrong did not have time to perform much geological training. The first lunar landing focused more on landing on the Moon and making it safely back to Earth than the scientific aspects of the mission. The duo was briefed by NASA and USGS geologists. They made one geological field trip to West Texas. The press followed them, and a helicopter made it hard for Aldrin and Armstrong to hear their instructor. On the morning of July 16, 1969, an estimated one million spectators watched the launch of Apollo 11 from the highways and beaches in the vicinity of Cape Canaveral, Florida. The launch was televised live in 33 countries, with an estimated 25 million viewers in the United States alone. Millions more listened to radio broadcasts. Propelled by a Saturn V rocket, Apollo 11 lifted off from Launch Complex 39 at the Kennedy Space Center on July 16, 1969, at 13:32:00 UTC (9:32:00 EDT), and entered Earth orbit twelve minutes later. After one and a half orbits, the S-IVB third-stage engine pushed the spacecraft onto its trajectory toward the Moon. About thirty minutes later, the transposition, docking, and extraction maneuver was performed: this involved separating the command module Columbia from the spent S-IVB stage; turning around; and docking with, and extracting, the lunar module Eagle. The combined spacecraft then headed for the Moon, while the S-IVB stage continued on a trajectory past the Moon. On July 19 at 17:21:50 UTC, Apollo 11 passed behind the Moon and fired its service propulsion engine to enter lunar orbit. In the thirty orbits that followed, the crew saw passing views of their landing site in the southern Sea of Tranquillity about 12 miles (19 km) southwest of the crater Sabine D. At 12:52:00 UTC on July 20, Aldrin and Armstrong entered Eagle, and began the final preparations for lunar descent. At 17:44:00 Eagle separated from the Columbia. Collins, alone aboard Columbia, inspected Eagle as it pirouetted before him to ensure the craft was not damaged and that the landing gear had correctly deployed. Throughout the descent, Aldrin called out navigation data to Armstrong, who was busy piloting the Eagle. Five minutes into the descent burn, and 6,000 feet (1,800 m) above the surface of the Moon, the LM guidance computer (LGC) distracted the crew with the first of several unexpected alarms that indicated that it could not complete all its tasks in real time and had to postpone some of them. Due to the 1202/1201 program alarms caused by spurious rendezvous radar inputs to the LGC, Armstrong manually landed the Eagle instead of using the computer's autopilot. The Eagle landed at 20:17:40 UTC on Sunday July 20 with about 25 seconds of fuel left. As a Presbyterian elder, Aldrin was the first and only person to hold a religious ceremony on the Moon. He radioed Earth: "I'd like to take this opportunity to ask every person listening in, whoever and wherever they may be, to pause for a moment and contemplate the events of the past few hours, and to give thanks in his or her own way." Using a kit given to him by his pastor, he took communion and read Jesus's words from the New Testament's John 15:5, as Aldrin records it: "I am the vine. You are the branches. Whoever remains in me, and I in him, will bear much fruit; for you can do nothing without me." But he kept this ceremony secret because of a lawsuit over the reading of Genesis on Apollo 8. In 1970 he commented: "It was interesting to think that the very first liquid ever poured on the Moon, and the first food eaten there, were communion elements." On reflection in his 2009 book, Aldrin said, "Perhaps, if I had it to do over again, I would not choose to celebrate communion. Although it was a deeply meaningful experience for me, it was a Christian sacrament, and we had come to the moon in the name of all mankind – be they Christians, Jews, Muslims, animists, agnostics, or atheists. But at the time I could think of no better way to acknowledge the enormity of the Apollo 11 experience than by giving thanks to God." Aldrin shortly hit upon a more universally human reference on the voyage back to Earth by publicly broadcasting his reading of the Old Testament's Psalm 8:3–4, as Aldrin records: "When I considered the heavens, the work of Thy fingers, the moon and the stars which Thou hast ordained, what is man that Thou art mindful of him." Photos of these liturgical documents reveal the conflict's development as Aldrin expresses faith. Preparations for the EVA began at 23:43. Once Armstrong and Aldrin were ready to go outside, Eagle was depressurized, and the hatch was opened at 02:39:33 on July 21. Aldrin set foot on the Moon at 03:15:16 on July 21, 1969 (UTC), nineteen minutes after Armstrong first touched the surface. Armstrong and Aldrin became the first and second people, respectively, to walk on the Moon. Aldrin's first words after he set foot on the Moon were "Beautiful view", to which Armstrong asked "Isn't that something? Magnificent sight out here." Aldrin answered, "Magnificent desolation." Aldrin and Armstrong had trouble erecting the Lunar Flag Assembly, but with some effort secured it into the surface. Aldrin saluted the flag while Armstrong photographed the scene. Aldrin positioned himself in front of the video camera and began experimenting with different locomotion methods to move about the lunar surface to aid future moonwalkers. During these experiments, President Nixon called the duo to congratulate them on the successful landing. Nixon closed with, "Thank you very much, and all of us look forward to seeing you on the Hornet on Thursday." Aldrin replied, "I look forward to that very much, sir." After the call, Aldrin began photographing and inspecting the spacecraft to document and verify its condition before their flight. Aldrin and Armstrong then set up a seismometer, to detect moonquakes, and a laser beam reflector. While Armstrong inspected a crater, Aldrin began the difficult task of hammering a metal tube into the surface to obtain a core sample. Most of the iconic photographs of an astronaut on the Moon taken by the Apollo 11 astronauts are of Aldrin; Armstrong appears in just two color photographs. "As the sequence of lunar operations evolved," Aldrin explained, "Neil had the camera most of the time, and the majority of the pictures taken on the Moon that include an astronaut are of me. It wasn't until we were back on Earth and in the Lunar Receiving Laboratory looking over the pictures that we realized there were few pictures of Neil. My fault perhaps, but we had never simulated this during our training." Aldrin reentered Eagle first but, as he tells it, before ascending the module's ladder he became the first person to urinate on the Moon. With some difficulty they lifted film and two sample boxes containing 21.55 kilograms (47.5 lb) of lunar surface material to the hatch using a flat cable pulley device. Armstrong reminded Aldrin of a bag of memorial items in his sleeve pocket, and Aldrin tossed the bag down. It contained a mission patch for the Apollo 1 flight that Ed White never flew due to his death in a cabin fire during the launch rehearsal; medallions commemorating Yuri Gagarin, the first man in space (who had died the previous year in a test flight accident), and Vladimir Komarov, the first man to die in a space flight, and a silicon disk etched with goodwill messages from 73 nations. After transferring to LM life support, the explorers lightened the ascent stage for the return to lunar orbit by tossing out their backpacks, lunar overshoes, an empty Hasselblad camera, and other equipment. The hatch was closed again at 05:01, and they repressurized the lunar module and settled down to sleep. At 17:54 UTC, they lifted off in Eagle's ascent stage to rejoin Collins aboard Columbia in lunar orbit. After rendezvous with Columbia, the ascent stage was jettisoned into lunar orbit, and Columbia made its way back to Earth. It splashed down in the Pacific 2,660 km (1,440 nmi) east of Wake Island at 16:50 UTC (05:50 local time) on July 24. The total mission duration was 195 hours, 18 minutes, 35 seconds. Bringing back pathogens from the lunar surface was considered a possibility, albeit remote, so divers passed biological isolation garments (BIGs) to the astronauts, and assisted them into the life raft. The astronauts were winched on board the recovery helicopter, and flown to the aircraft carrier USS Hornet, where they spent the first part of the Earth-based portion of 21 days of quarantine. On August 13, the three astronauts rode in ticker-tape parades in their honor in New York and Chicago, attended by an estimated six million people. An official state dinner that evening in Los Angeles celebrated the flight. President Richard Nixon honored each of them with the highest American civilian award, the Presidential Medal of Freedom (with distinction). On September 16, 1969, the astronauts addressed a joint session of Congress where they thanked the representatives for their past support and implored them to continue funding the space effort. The astronauts embarked on a 38-day world tour on September 29 that brought the astronauts to 22 foreign countries and included visits with leaders of multiple countries. The last leg of the tour included Australia, South Korea, and Japan; the crew returned to the US on November 5, 1969. After Apollo 11, Aldrin was kept busy giving speeches and making public appearances. In October 1970, he joined Soviet cosmonauts Andriyan Nikolayev and Vitaly Sevastyanov on their tour of the NASA space centers. He was also involved in the design of the Space Shuttle. With the Apollo program coming to an end, Aldrin, now a colonel, saw few prospects at NASA, and decided to return to the Air Force on July 1, 1971. During his NASA career, he had spent 289 hours and 53 minutes in space, of which 7 hours and 52 minutes was in EVA. Aldrin hoped to become Commandant of Cadets at the United States Air Force Academy, but the job went to his West Point classmate Hoyt S. Vandenberg Jr. Aldrin was made Commandant of the USAF Aerospace Research Pilot School at Edwards Air Force Base, California. Aldrin had neither managerial nor test pilot experience, but a third of the training curriculum was devoted to astronaut training and students flew a modified F-104 Starfighter to the edge of space. Fellow Group 3 astronaut and moonwalker Alan Bean considered him well qualified for the job. Aldrin did not get along well with his superior, Brigadier General Robert M. White, who had earned his USAF astronaut wings flying the X-15. Aldrin's celebrity status led people to defer to him more than the higher-ranking general. There were two crashes at Edwards, of an A-7 Corsair II and a T-33. No people died, but the aircraft were destroyed and the accidents were attributed to insufficient supervision, which placed the blame on Aldrin. What he had hoped would be an enjoyable job became a highly stressful one. Aldrin went to see the base surgeon. In addition to signs of depression, he experienced neck and shoulder pains, and hoped that the latter might explain the former. He was hospitalized for depression at Wilford Hall Medical Center for four weeks. His mother had committed suicide in May 1968, and he was plagued with guilt that his fame after Gemini 12 had contributed. His mother's father had also committed suicide, and he believed he inherited depression from them. At the time there was great stigma related to mental illness and he was aware that it could not only be career-ending, but could result in his being ostracized socially. In February 1972, General George S. Brown paid a visit to Edwards and informed Aldrin that the USAF Aerospace Research Pilot School was being renamed the USAF Test Pilot School and the astronaut training was being dropped. With the Apollo program winding down, and Air Force budgets being cut, the Air Force's interest in space diminished. Aldrin elected to retire as a colonel on March 1, 1972, after 21 years of service. His father and General Jimmy Doolittle, a close friend of his father, attended the formal retirement ceremony. Aldrin's father died on December 28, 1974, from complications following a heart attack. Aldrin's autobiographies, Return to Earth (1973) and Magnificent Desolation (2009), recounted his struggles with clinical depression and alcoholism in the years after leaving NASA. Encouraged by a therapist to take a regular job, Aldrin worked selling used cars, at which he had no talent. Periods of hospitalization and sobriety alternated with bouts of heavy drinking. Eventually he was arrested for disorderly conduct. Finally, in October 1978, he quit drinking for good. Aldrin attempted to help others with drinking problems, including actor William Holden. Holden's girlfriend Stefanie Powers had portrayed Marianne, a woman with whom Aldrin had an affair, in the 1976 TV movie version of Return to Earth. Aldrin was saddened by Holden's alcohol-related death in 1981. On September 9, 2002, Aldrin was lured to a Beverly Hills hotel on the pretext of being interviewed for a Japanese children's television show on the subject of space. When he arrived, Moon landing conspiracy theorist Bart Sibrel accosted him with a film crew and demanded he swear on a Bible that the Moon landings were not faked. After a brief confrontation, during which Sibrel followed Aldrin despite being told to leave him alone, and called him "a coward, a liar, and a thief" the 72-year-old Aldrin punched Sibrel in the jaw, which was caught on camera by Sibrel's film crew. Aldrin said he had acted to defend himself and his stepdaughter. Witnesses said Sibrel had aggressively poked Aldrin with a Bible. Additional mitigating factors were that Sibrel sustained no visible injury and did not seek medical attention, and that Aldrin had no criminal record. The police declined to press charges against Aldrin. In 2005, while being interviewed for a Science Channel documentary titled First on the Moon: The Untold Story, Aldrin told an interviewer they had seen an unidentified flying object (UFO). The documentary makers omitted the crew's conclusion that they probably saw one of the four detached spacecraft adapter panels from the upper stage of the Saturn V rocket. The panels had been jettisoned before the separation maneuver so they closely followed the spacecraft until the first mid-course correction. When Aldrin appeared on The Howard Stern Show on August 15, 2007, Stern asked him about the supposed UFO sighting. Aldrin confirmed that there was no such sighting of anything deemed extraterrestrial and said they were, and are, "99.9 percent" sure the object was the detached panel. According to Aldrin his words had been taken out of context. He made a request to the Science Channel to make a correction, but was refused. In December 2016, Aldrin was part of a tourist group visiting the Amundsen–Scott South Pole Station in Antarctica when he fell ill and was evacuated, first to McMurdo Station and from there to Christchurch, New Zealand. At 86 years of age, Aldrin's visit made him the oldest person to reach the South Pole. He had traveled to the North Pole in 1998. After leaving NASA, Aldrin continued to advocate for space exploration. In 1985 he joined the University of North Dakota (UND)'s College of Aerospace Sciences at the invitation of John D. Odegard, the dean of the college. Aldrin helped to develop UND's Space Studies program and brought David Webb from NASA to serve as the department's first chair. To further promote space exploration, and to commemorate the 40th anniversary of the first lunar landing, Aldrin teamed up with Snoop Dogg, Quincy Jones, Talib Kweli, and Soulja Boy to create the rap single and video "Rocket Experience", proceeds from which were donated to Aldrin's non-profit foundation, ShareSpace. He is also a member of the Mars Society's Steering committee. In 1985, Aldrin proposed a special spacecraft trajectory now known as the Aldrin cycler. Cycler trajectories offer reduced cost of repeated travel to Mars by using less propellant. The Aldrin cycler provided a five and a half month journey from the Earth to Mars, with a return trip to Earth of the same duration on a twin cycler orbit. Aldrin continues to research this concept with engineers from Purdue University. In 1996 Aldrin founded Starcraft Boosters, Inc. (SBI) to design reusable rocket launchers. In December 2003, Aldrin published an opinion piece in The New York Times criticizing NASA's objectives. In it, he voiced concern about NASA's development of a spacecraft "limited to transporting four astronauts at a time with little or no cargo carrying capability" and declared the goal of sending astronauts back to the Moon was "more like reaching for past glory than striving for new triumphs". In a June 2013 opinion piece in The New York Times, Aldrin supported a human mission to Mars and which viewed the Moon "not as a destination but more a point of departure, one that places humankind on a trajectory to homestead Mars and become a two-planet species." In August 2015, Aldrin, in association with the Florida Institute of Technology, presented a master plan to NASA for consideration where astronauts, with a tour of duty of ten years, establish a colony on Mars before the year 2040. Aldrin was awarded the Air Force Distinguished Service Medal (DSM) in 1969 for his role as lunar module pilot on Apollo 11. He was awarded an oak leaf cluster in 1972 in lieu of a second DSM for his role in both the Korean War and in the space program, and the Legion of Merit for his role in the Gemini and Apollo programs. During a 1966 ceremony marking the end of the Gemini program, Aldrin was awarded the NASA Exceptional Service Medal by President Johnson at LBJ Ranch. He was awarded the NASA Distinguished Service Medal in 1970 for the Apollo 11 mission. Aldrin was one of ten Gemini astronauts inducted into the International Space Hall of Fame in 1982. He was also inducted into the U.S. Astronaut Hall of Fame in 1993, the National Aviation Hall of Fame in 2000, and the New Jersey Hall of Fame in 2008. The Toy Story character Buzz Lightyear was named in honor of Buzz Aldrin. In 1999, while celebrating the 30th anniversary of the lunar landing, Vice President Al Gore, who was also the vice-chancellor of the Smithsonian Institution's Board of Regents, presented the Apollo 11 crew with the Smithsonian Institution's Langley Gold Medal for aviation. After the ceremony, the crew went to the White House and presented President Bill Clinton with an encased Moon rock. The Apollo 11 crew was awarded the New Frontier Congressional Gold Medal in the Capitol Rotunda in 2011. During the ceremony, NASA administrator Charles Bolden said, "Those of us who have had the privilege to fly in space followed the trail they forged." The Apollo 11 crew were awarded the Collier Trophy in 1969. The National Aeronautic Association president awarded a duplicate trophy to Collins and Aldrin at a ceremony. The crew was awarded the 1969 General Thomas D. White USAF Space Trophy. The National Space Club named the crew the winners of the 1970 Dr. Robert H. Goddard Memorial Trophy, awarded annually for the greatest achievement in spaceflight. They received the international Harmon Trophy for aviators in 1970, conferred to them by Vice President Spiro Agnew in 1971. Agnew also presented them the Hubbard Medal of the National Geographic Society in 1970. He told them, "You've won a place alongside Christopher Columbus in American history". In 1970, the Apollo 11 team were co-winners of the Iven C. Kincheloe award from the Society of Experimental Test Pilots along with Darryl Greenamyer who broke the world speed record for piston engine airplanes. For contributions to the television industry, they were honored with round plaques on the Hollywood Walk of Fame. In 2001, President George W. Bush appointed Aldrin to the Commission on the Future of the United States Aerospace Industry. Aldrin received the 2003 Humanitarian Award from Variety, the Children's Charity, which, according to the organization, "is given to an individual who has shown unusual understanding, empathy, and devotion to mankind." In 2006, the Space Foundation awarded him its highest honor, the General James E. Hill Lifetime Space Achievement Award. Aldrin received honorary degrees from six colleges and universities, and was named as the Chancellor of the International Space University in 2015. He was a member of the National Space Society's Board of Governors, and has served as the organization's chairman. In 2016, his hometown middle school in Montclair, New Jersey, was renamed Buzz Aldrin Middle School. The Aldrin crater on the Moon near the Apollo 11 landing site and Asteroid 6470 Aldrin are named in his honor. In 2019, Aldrin was awarded the Starmus Festival's Stephen Hawking Medal for Science Communication for Lifetime Achievement. On his 93rd birthday he was honored by Living Legends of Aviation. On May 5, 2023, he received an honorary promotion to the rank of brigadier general in the United States Air Force, as well as being made an honorary Space Force guardian. On May 24, 2023, he received the 40 over 40 award by Monaco Voice, calling him "Guardian of the Galaxy". Aldrin has been married four times. His first marriage was on December 29, 1954, to Joan Archer, a Rutgers University and Columbia University alumna with a master's degree. They had three children, James, Janice and Andrew. They filed for divorce in 1974. His second wife was Beverly Van Zile, whom he married on December 31, 1975, and divorced in 1978. His third wife was Lois Driggs Cannon, whom he married on February 14, 1988. Their divorce was finalized in December 2012. The settlement included 50 percent of their $475,000 bank account and $9,500 a month plus 30 percent of his annual income, estimated at more than $600,000. As of 2017, he had one grandson, Jeffrey Schuss, born to his daughter Janice, and three great-grandsons and one great-granddaughter. In 2018, Aldrin was involved in a legal dispute with his children Andrew and Janice and former business manager Christina Korp over their claims that he was mentally impaired through dementia and Alzheimer's disease. His children alleged that he made new friends who were alienating him from the family and encouraging him to spend his savings at a high rate. They sought to be named legal guardians so they could control his finances. In June, Aldrin filed a lawsuit against Andrew, Janice, Korp, and businesses and foundations run by the family. Aldrin alleged that Janice was not acting in his financial interest and that Korp was exploiting the elderly. He sought to remove Andrew's control of Aldrin's social media accounts, finances, and businesses. The situation ended when his children withdrew their petition and he dropped the lawsuit in March 2019, several months before the 50th anniversary of the Apollo 11 mission. On January 20, 2023, his 93rd birthday, Aldrin announced on Twitter that he had married for the fourth time, to his 63-year-old companion, Anca Faur. Aldrin is an active supporter of the Republican Party, headlining fundraisers for its members of Congress and endorsing its candidates. He appeared at a rally for George W. Bush in 2004 and campaigned for Paul Rancatore in Florida in 2008, Mead Treadwell in Alaska in 2014 and Dan Crenshaw in Texas in 2018. He appeared at the 2019 State of the Union Address as a guest of President Donald Trump. Buzz Aldrin is the first Freemason to set foot on the Moon. Aldrin was initiated into Freemasonry at Oak Park Lodge No. 864 in Alabama and raised at Lawrence N. Greenleaf Lodge, No. 169 in Colorado. By the time Aldrin stepped onto the lunar surface, he was a member of two Masonic lodges: Montclair Lodge No. 144 in New Jersey and Clear Lake Lodge No. 1417 in Seabrook, Texas, where he was invited to serve on the High Council and was ordained in the 33rd degree of the Ancient and Accepted Scottish Rite. He is also a member of York Rite and Arabia Shrine Temple of Houston. In 2007, Aldrin confirmed to Time magazine that he had recently had a face-lift, joking that the g-forces he was exposed to in space "caused a sagging jowl that needed some attention." Following the 2012 death of his Apollo 11 colleague Neil Armstrong, Aldrin said he was ... deeply saddened by the passing ... I know I am joined by many millions of others from around the world in mourning the passing of a true American hero and the best pilot I ever knew ... I had truly hoped that on July 20, 2019, Neil, Mike and I would be standing together to commemorate the 50th Anniversary of our moon landing. Aldrin has primarily resided in the Los Angeles area, including Beverly Hills and Laguna Beach since 1985. In 2014, he sold his Westwood condominium; this was after his third divorce in 2012. He also lives in Satellite Beach, Florida. Aldrin has been portrayed by:
{{redirect|TB|the unit of information|Terabyte}} {{Infobox disease |Name = Tuberculosis |Image = Tuberculosis-x-ray-1.jpg |Caption = Chest [[X-ray]] of a person with advanced tuberculosis. White arrows point to infection in both lungs. Black arrows point to a cavity that has formed. |DiseasesDB = 8515 |ICD10 = {{ICD10|A|15||a|15}}–{{ICD10|A|19||a|15}} |ICD9 = {{ICD9|010}}–{{ICD9|018}} |ICDO = |OMIM = 607948 |MedlinePlus = 000077 |MedlinePlus_mult = |eMedicineSubj = med |eMedicineTopic = 2324 |MeshID = D014376}} [[Image:TB poster.jpg|thumb|Public health campaigns in the 1920s tried to halt the spread of TB.]] '''Tuberculosis (TB)''' is an [[infectious disease]] caused by [[bacteria]]. In the past, people called it '''consumption.''' TB is caused by several types of [[Mycobacterium|mycobacteria]], usually ''[[Mycobacterium tuberculosis]]''.<ref name="Robbins"/> The disease usually attacks the [[lung]]s, but it can also affect other parts of the body. == How it spreads == The bacteria can travel through the air and spread from one person to the next. This happens when infected people [[cough]], [[sneeze]], or spit.<ref name="AP">{{cite journal|author = Konstantinos A|year = 2010|title = Testing for tuberculosis|journal = Australian Prescriber|volume = 33|issue = 1|pages = 12–18|doi = 10.18773/austprescr.2010.005|url = http://www.australianprescriber.com/magazine/33/1/12/18/}}</ref> Of every 100 people with TB, between five and ten people show [[symptom]]s.<ref name="Pet2005">{{Cite book|url=https://www.worldcat.org/oclc/181343640|title=Evidence-based respiratory medicine|date=2005|publisher=BMJ Books/Blackwell Pub|isbn=978-0-470-98737-7|location=Malden, Mass.|oclc=181343640}}</ref> In these people, the disease is called ''active''. Tuberculosis kills more than half of the people who are infected if they do not get treatment. TB affects the a lot of people globally, but many do not show any symptoms so it goes underrepresented. ==Detection and treatment== [[File:TB Culture.jpg|thumb|A close up of a culture of ''[[Mycobacterium tuberculosis]]''. The patches that look like foam are the typical growth pattern of these bacteria|78x78px]] [[Diagnosis]] of active TB relies on [[radiology]]. Doctors often look at an [[X-ray]] of the [[chest]]. In addition, they check body [[fluid]]s. These fluids have [[microbe]]s in them, which are grown in [[cell culture]]s. The cell cultures are then analysed to see if the person is infected with TB. If the patient has TB, but does not show symptoms, the disease is 'latent'. Doctors use a [[skin]] test, called the [[Mantoux test]], to detect latent TB. They often do [[blood]] tests too. There is a [[vaccine]] against some forms of tuberculosis. It is called the [[bacillus Calmette-Guérin|Bacillus Calmette&ndash;Guérin]] vaccine. TB used to be easily treated and cured with [[antibiotics]]. However, the bacterium is now highly resistant to most antibiotics. This resistance makes treatment difficult. Many different kinds of antibiotics need to be given over a long period of time. There is a form of tuberculosis that is resistant to ''all'' [[drug]]s. == Symptoms == Tuberculosis can have many symptoms. The most common include: * A [[cough]] that does not go away, especially if the person is coughing up [[blood]] (this is called [[hemoptysis]]) * Chest pain * Not having any [[appetite]] * Weakness * Weight loss * Chills * Very pale skin * Listless eyes * [[Fever]] * [[Sweat]]ing a lot at night * [[Dyspnea|Difficulty breathing]] * [[Fatigue|Feeling very tired]] * People are also more likely to get tuberculosis if they live close to other people who have TB. For example, TB can spread easily in [[homeless shelter]]s, [[prison]]s, and [[Immigration|immigrant]] communities ==How common is TB?== [[File:Tuberculosis incidence (per 100,000 people),_OWID.svg|thumb|alt=World map with [[sub-Saharan Africa]] in darker shades of blue, marking prevalences above 250 per 100,000 people, and with North America, Australia, Europe, and Western Asia in white, marking prevalence below 20 per 100,000 people. Asia is light green and shades of blue, marking prevalence around 50 to 500 per 100,000 people. South America is mostly light green.|This world map shows the [[prevalence]] of TB, per 100,000 people, as of 2016. Countries with more cases are shown in darker colors, those with fewer cases are shown in lighter colors. The most cases were recorded in [[Southeast Asia]] and [[Sub-Saharan Africa]].<ref name="WHO2020">{{cite web | title=Tuberculosis (TB) | website=WHO | url=https://www.who.int/news-room/fact-sheets/detail/tuberculosis | access-date=20 February 2023}}</ref>]] Experts believe that one third of the [[world population]] is infected with ''M. tuberculosis''.<ref name=WHO2012data>{{cite web|url=http://www.who.int/mediacentre/factsheets/fs104/en/index.html|title=Tuberculosis Fact sheet N°104|publisher=[[World Health Organization]]|date=November 2010|access-date=26 July 2011}}</ref> New infections occur at a rate of one per second.<ref name=WHO2012data/> In 2007, about 13.7 million [[chronic]] cases were active globally.<ref name=WHO2009-Epidemiology>{{cite book |title=Global tuberculosis control: epidemiology, strategy, financing |author=World Health Organization |year=2009 |isbn=978-92-4-156380-2 |chapter=Epidemiology |chapter-url=http://who.int/entity/tb/publications/global_report/2009/pdf/chapter1.pdf |access-date=12 November 2009 |pages=6–33 }}{{Dead link|date=January 2021 |bot=InternetArchiveBot |fix-attempted=yes }}</ref> In 2010, about 8.8&nbsp;million new cases developed and nearly 1.5&nbsp;million people died from the disease, most of them in [[Developing country|developing countries]].<ref name=WHO2011>{{cite web|title=The sixteenth global report on tuberculosis|author=World Health Organization|url=http://www.who.int/tb/publications/global_report/2011/gtbr11_executive_summary.pdf|year=2011}}</ref> The number of tuberculosis cases has been decreasing since 2006, and new cases have decreased since 2002.<ref name=WHO2011/> Tuberculosis does not happen at the same rate around the world. About eighty percent of the population in many [[Asia]]n and [[Africa]]n countries test positive for TB, but only five to ten percent of people in the United States do.<ref name="Robbins">{{cite book |author = Kumar V. ''et al'' 2007|title = Robbins basic pathology|year = 2007|edition = 8th|publisher = Saunders Elsevier|pages = 516–522|isbn = 978-1-4160-2973-1}}</ref> People usually get tuberculosis because of a weakened [[immune system]]. Many people with [[HIV]] and [[AIDS]] can also get tuberculosis.<ref name="Lancet11">{{cite journal|title=Tuberculosis|journal=Lancet|date= 2011|volume=378|issue=9785|pages=57–72|pmid=21420161|doi=10.1016/S0140-6736(10)62173-3|s2cid=208791546|last1=Lawn|first1=Stephen D.|last2=Zumla|first2=Alimuddin I.}}</ref> ==References== {{Reflist|30em}} == Other websites == * [https://www.cdc.gov/tb/faqs/qa_introduction.htm Frequently asked Questions about TB] at CDC.gov * [http://www.explaintb.org ExplainTB]: Multilingual audiovisual information on tuberculosis [[Category:Diseases caused by bacteria]] [[Category:Pulmonology]]
Tuberculosis (TB), also known colloquially as the "white death", or historically as consumption, is an infectious disease usually caused by Mycobacterium tuberculosis (MTB) bacteria. Tuberculosis generally affects the lungs, but it can also affect other parts of the body. Most infections show no symptoms, in which case it is known as latent tuberculosis. Around 10% of latent infections progress to active disease which, if left untreated, kill about half of those affected. Typical symptoms of active TB are chronic cough with blood-containing mucus, fever, night sweats, and weight loss. Infection of other organs can cause a wide range of symptoms. Tuberculosis is spread from one person to the next through the air when people who have active TB in their lungs cough, spit, speak, or sneeze. People with latent TB do not spread the disease. Active infection occurs more often in people with HIV/AIDS and in those who smoke. Diagnosis of active TB is based on chest X-rays, as well as microscopic examination and culture of body fluids. Diagnosis of Latent TB relies on the tuberculin skin test (TST) or blood tests. Prevention of TB involves screening those at high risk, early detection and treatment of cases, and vaccination with the bacillus Calmette-Guérin (BCG) vaccine. Those at high risk include household, workplace, and social contacts of people with active TB. Treatment requires the use of multiple antibiotics over a long period of time. Antibiotic resistance is a growing problem, with increasing rates of multiple drug-resistant tuberculosis (MDR-TB). In 2018, one quarter of the world's population was thought to have a latent infection of TB. New infections occur in about 1% of the population each year. In 2022, an estimated 10.6 million people developed active TB, resulting in 1.3 million deaths, making it the second leading cause of death from an infectious disease after COVID-19. As of 2018, most TB cases occurred in the regions of South-East Asia (44%), Africa (24%), and the Western Pacific (18%), with more than 50% of cases being diagnosed in seven countries: India (27%), China (9%), Indonesia (8%), the Philippines (6%), Pakistan (6%), Nigeria (4%), and Bangladesh (4%). By 2021, the number of new cases each year was decreasing by around 2% annually. About 80% of people in many Asian and African countries test positive, while 5–10% of people in the United States test positive via the tuberculin test. Tuberculosis has been present in humans since ancient times. Tuberculosis may infect any part of the body, but most commonly occurs in the lungs (known as pulmonary tuberculosis). Extrapulmonary TB occurs when tuberculosis develops outside of the lungs, although extrapulmonary TB may coexist with pulmonary TB. General signs and symptoms include fever, chills, night sweats, loss of appetite, weight loss, and fatigue. Significant nail clubbing may also occur. If a tuberculosis infection does become active, it most commonly involves the lungs (in about 90% of cases). Symptoms may include chest pain and a prolonged cough producing sputum. About 25% of people may not have any symptoms (i.e., they remain asymptomatic). Occasionally, people may cough up blood in small amounts, and in very rare cases, the infection may erode into the pulmonary artery or a Rasmussen's aneurysm, resulting in massive bleeding. Tuberculosis may become a chronic illness and cause extensive scarring in the upper lobes of the lungs. The upper lung lobes are more frequently affected by tuberculosis than the lower ones. The reason for this difference is not clear. It may be due to either better air flow, or poor lymph drainage within the upper lungs. In 15–20% of active cases, the infection spreads outside the lungs, causing other kinds of TB. These are collectively denoted as extrapulmonary tuberculosis. Extrapulmonary TB occurs more commonly in people with a weakened immune system and young children. In those with HIV, this occurs in more than 50% of cases. Notable extrapulmonary infection sites include the pleura (in tuberculous pleurisy), the central nervous system (in tuberculous meningitis), the lymphatic system (in scrofula of the neck), the genitourinary system (in urogenital tuberculosis), and the bones and joints (in Pott disease of the spine), among others. A potentially more serious, widespread form of TB is called "disseminated tuberculosis", it is also known as miliary tuberculosis. Miliary TB currently makes up about 10% of extrapulmonary cases. The main cause of TB is Mycobacterium tuberculosis (MTB), a small, aerobic, nonmotile bacillus. The high lipid content of this pathogen accounts for many of its unique clinical characteristics. It divides every 16 to 20 hours, which is an extremely slow rate compared with other bacteria, which usually divide in less than an hour. Mycobacteria have an outer membrane lipid bilayer. If a Gram stain is performed, MTB either stains very weakly "Gram-positive" or does not retain dye as a result of the high lipid and mycolic acid content of its cell wall. MTB can withstand weak disinfectants and survive in a dry state for weeks. In nature, the bacterium can grow only within the cells of a host organism, but M. tuberculosis can be cultured in the laboratory. Using histological stains on expectorated samples from phlegm (also called sputum), scientists can identify MTB under a microscope. Since MTB retains certain stains even after being treated with acidic solution, it is classified as an acid-fast bacillus. The most common acid-fast staining techniques are the Ziehl–Neelsen stain and the Kinyoun stain, which dye acid-fast bacilli a bright red that stands out against a blue background. Auramine-rhodamine staining and fluorescence microscopy are also used. The M. tuberculosis complex (MTBC) includes four other TB-causing mycobacteria: M. bovis, M. africanum, M. canettii, and M. microti. M. africanum is not widespread, but it is a significant cause of tuberculosis in parts of Africa. M. bovis was once a common cause of tuberculosis, but the introduction of pasteurized milk has almost eliminated this as a public health problem in developed countries. M. canettii is rare and seems to be limited to the Horn of Africa, although a few cases have been seen in African emigrants. M. microti is also rare and is seen almost only in immunodeficient people, although its prevalence may be significantly underestimated. Other known pathogenic mycobacteria include M. leprae, M. avium, and M. kansasii. The latter two species are classified as "nontuberculous mycobacteria" (NTM) or atypical mycobacteria. NTM cause neither TB nor leprosy, but they do cause lung diseases that resemble TB. When people with active pulmonary TB cough, sneeze, speak, sing, or spit, they expel infectious aerosol droplets 0.5 to 5.0 µm in diameter. A single sneeze can release up to 40,000 droplets. Each one of these droplets may transmit the disease, since the infectious dose of tuberculosis is very small (the inhalation of fewer than 10 bacteria may cause an infection). People with prolonged, frequent, or close contact with people with TB are at particularly high risk of becoming infected, with an estimated 22% infection rate. A person with active but untreated tuberculosis may infect 10–15 (or more) other people per year. Transmission should occur from only people with active TB – those with latent infection are not thought to be contagious. The probability of transmission from one person to another depends upon several factors, including the number of infectious droplets expelled by the carrier, the effectiveness of ventilation, the duration of exposure, the virulence of the M. tuberculosis strain, the level of immunity in the uninfected person, and others. The cascade of person-to-person spread can be circumvented by segregating those with active ("overt") TB and putting them on anti-TB drug regimens. After about two weeks of effective treatment, subjects with nonresistant active infections generally do not remain contagious to others. If someone does become infected, it typically takes three to four weeks before the newly infected person becomes infectious enough to transmit the disease to others. A number of factors make individuals more susceptible to TB infection and/or disease. The most important risk factor globally for developing active TB is concurrent HIV infection; 13% of those with TB are also infected with HIV. This is a particular problem in sub-Saharan Africa, where HIV infection rates are high. Of those without HIV infection who are infected with tuberculosis, about 5–10% develop active disease during their lifetimes; in contrast, 30% of those co-infected with HIV develop the active disease. Use of certain medications, such as corticosteroids and infliximab (an anti-αTNF monoclonal antibody), is another important risk factor, especially in the developed world. Other risk factors include: alcoholism, diabetes mellitus (3-fold increased risk), silicosis (30-fold increased risk), tobacco smoking (2-fold increased risk), indoor air pollution, malnutrition, young age, recently acquired TB infection, recreational drug use, severe kidney disease, low body weight, organ transplant, head and neck cancer, and genetic susceptibility (the overall importance of genetic risk factors remains undefined). Tobacco smoking increases the risk of infections (in addition to increasing the risk of active disease and death). Additional factors increasing infection susceptibility include young age. About 90% of those infected with M. tuberculosis have asymptomatic, latent TB infections (sometimes called LTBI), with only a 10% lifetime chance that the latent infection will progress to overt, active tuberculous disease. In those with HIV, the risk of developing active TB increases to nearly 10% a year. If effective treatment is not given, the death rate for active TB cases is up to 66%. TB infection begins when the mycobacteria reach the alveolar air sacs of the lungs, where they invade and replicate within endosomes of alveolar macrophages. Macrophages identify the bacterium as foreign and attempt to eliminate it by phagocytosis. During this process, the bacterium is enveloped by the macrophage and stored temporarily in a membrane-bound vesicle called a phagosome. The phagosome then combines with a lysosome to create a phagolysosome. In the phagolysosome, the cell attempts to use reactive oxygen species and acid to kill the bacterium. However, M. tuberculosis has a thick, waxy mycolic acid capsule that protects it from these toxic substances. M. tuberculosis is able to reproduce inside the macrophage and will eventually kill the immune cell. The primary site of infection in the lungs, known as the Ghon focus, is generally located in either the upper part of the lower lobe, or the lower part of the upper lobe. Tuberculosis of the lungs may also occur via infection from the blood stream. This is known as a Simon focus and is typically found in the top of the lung. This hematogenous transmission can also spread infection to more distant sites, such as peripheral lymph nodes, the kidneys, the brain, and the bones. All parts of the body can be affected by the disease, though for unknown reasons it rarely affects the heart, skeletal muscles, pancreas, or thyroid. Tuberculosis is classified as one of the granulomatous inflammatory diseases. Macrophages, epithelioid cells, T lymphocytes, B lymphocytes, and fibroblasts aggregate to form granulomas, with lymphocytes surrounding the infected macrophages. When other macrophages attack the infected macrophage, they fuse together to form a giant multinucleated cell in the alveolar lumen. The granuloma may prevent dissemination of the mycobacteria and provide a local environment for interaction of cells of the immune system. However, more recent evidence suggests that the bacteria use the granulomas to avoid destruction by the host's immune system. Macrophages and dendritic cells in the granulomas are unable to present antigen to lymphocytes; thus the immune response is suppressed. Bacteria inside the granuloma can become dormant, resulting in latent infection. Another feature of the granulomas is the development of abnormal cell death (necrosis) in the center of tubercles. To the naked eye, this has the texture of soft, white cheese and is termed caseous necrosis. If TB bacteria gain entry to the blood stream from an area of damaged tissue, they can spread throughout the body and set up many foci of infection, all appearing as tiny, white tubercles in the tissues. This severe form of TB disease, most common in young children and those with HIV, is called miliary tuberculosis. People with this disseminated TB have a high fatality rate even with treatment (about 30%). In many people, the infection waxes and wanes. Tissue destruction and necrosis are often balanced by healing and fibrosis. Affected tissue is replaced by scarring and cavities filled with caseous necrotic material. During active disease, some of these cavities are joined to the air passages (bronchi) and this material can be coughed up. It contains living bacteria and thus can spread the infection. Treatment with appropriate antibiotics kills bacteria and allows healing to take place. Upon cure, affected areas are eventually replaced by scar tissue. Diagnosing active tuberculosis based only on signs and symptoms is difficult, as is diagnosing the disease in those who have a weakened immune system. A diagnosis of TB should, however, be considered in those with signs of lung disease or constitutional symptoms lasting longer than two weeks. A chest X-ray and multiple sputum cultures for acid-fast bacilli are typically part of the initial evaluation. Interferon-γ release assays (IGRA) and tuberculin skin tests are of little use in most of the developing world. IGRA have similar limitations in those with HIV. A definitive diagnosis of TB is made by identifying M. tuberculosis in a clinical sample (e.g., sputum, pus, or a tissue biopsy). However, the difficult culture process for this slow-growing organism can take two to six weeks for blood or sputum culture. Thus, treatment is often begun before cultures are confirmed. Nucleic acid amplification tests and adenosine deaminase testing may allow rapid diagnosis of TB. Blood tests to detect antibodies are not specific or sensitive, so they are not recommended. The Mantoux tuberculin skin test is often used to screen people at high risk for TB. Those who have been previously immunized with the Bacille Calmette-Guerin vaccine may have a false-positive test result. The test may be falsely negative in those with sarcoidosis, Hodgkin's lymphoma, malnutrition, and most notably, active tuberculosis. Interferon gamma release assays, on a blood sample, are recommended in those who are positive to the Mantoux test. These are not affected by immunization or most environmental mycobacteria, so they generate fewer false-positive results. However, they are affected by M. szulgai, M. marinum, and M. kansasii. IGRAs may increase sensitivity when used in addition to the skin test, but may be less sensitive than the skin test when used alone. The US Preventive Services Task Force (USPSTF) has recommended screening people who are at high risk for latent tuberculosis with either tuberculin skin tests or interferon-gamma release assays. While some have recommend testing health care workers, evidence of benefit for this is poor as of 2019. The Centers for Disease Control and Prevention (CDC) stopped recommending yearly testing of health care workers without known exposure in 2019. Tuberculosis prevention and control efforts rely primarily on the vaccination of infants and the detection and appropriate treatment of active cases. The World Health Organization (WHO) has achieved some success with improved treatment regimens, and a small decrease in case numbers. Some countries have legislation to involuntarily detain or examine those suspected to have tuberculosis, or involuntarily treat them if infected. The only available vaccine as of 2021 is bacillus Calmette-Guérin (BCG). In children it decreases the risk of getting the infection by 20% and the risk of infection turning into active disease by nearly 60%. It is the most widely used vaccine worldwide, with more than 90% of all children being vaccinated. The immunity it induces decreases after about ten years. As tuberculosis is uncommon in most of Canada, Western Europe, and the United States, BCG is administered to only those people at high risk. Part of the reasoning against the use of the vaccine is that it makes the tuberculin skin test falsely positive, reducing the test's usefulness as a screening tool. Several vaccines are being developed. Intradermal MVA85A vaccine in addition to BCG injection is not effective in preventing tuberculosis. Public health campaigns which have focused on overcrowding, public spitting and regular sanitation (including hand washing) during the 1800s helped to either interrupt or slow spread which when combined with contact tracing, isolation and treatment helped to dramatically curb the transmission of both tuberculosis and other airborne diseases which led to the elimination of tuberculosis as a major public health issue in most developed economies. Other risk factors which worsened TB spread such as malnutrition were also ameliorated, but since the emergence of HIV a new population of immunocompromised individuals was available for TB to infect. The World Health Organization (WHO) declared TB a "global health emergency" in 1993, and in 2006, the Stop TB Partnership developed a Global Plan to Stop Tuberculosis that aimed to save 14 million lives between its launch and 2015. A number of targets they set were not achieved by 2015, mostly due to the increase in HIV-associated tuberculosis and the emergence of multiple drug-resistant tuberculosis. A tuberculosis classification system developed by the American Thoracic Society is used primarily in public health programs. In 2015, it launched the End TB Strategy to reduce deaths by 95% and incidence by 90% before 2035. The goal of tuberculosis elimination is hampered by the lack of rapid testing, of short and effective treatment courses, and of completely effective vaccines. The benefits and risks of giving anti-tubercular drugs in those exposed to MDR-TB is unclear. Making HAART therapy available to HIV-positive individuals significantly reduces the risk of progression to an active TB infection by up to 90% and can mitigate the spread through this population. Treatment of TB uses antibiotics to kill the bacteria. Effective TB treatment is difficult, due to the unusual structure and chemical composition of the mycobacterial cell wall, which hinders the entry of drugs and makes many antibiotics ineffective. Active TB is best treated with combinations of several antibiotics to reduce the risk of the bacteria developing antibiotic resistance. The routine use of rifabutin instead of rifampicin in HIV-positive people with tuberculosis is of unclear benefit as of 2007. Acetylsalicylic acid (aspirin) at a dose of 100 mg per day has been shown to improve clinical signs and symptoms, reduce cavitary lesions, lower inflammatory markers, and increase the rate of sputum-negative conversion in patients with pulmonary tuberculosis. Latent TB is treated with either isoniazid or rifampin alone, or a combination of isoniazid with either rifampicin or rifapentine. The treatment takes three to nine months depending on the medications used. People with latent infections are treated to prevent them from progressing to active TB disease later in life. Education or counselling may improve the latent tuberculosis treatment completion rates. The recommended treatment of new-onset pulmonary tuberculosis, as of 2010, is six months of a combination of antibiotics containing rifampicin, isoniazid, pyrazinamide, and ethambutol for the first two months, and only rifampicin and isoniazid for the last four months. Where resistance to isoniazid is high, ethambutol may be added for the last four months as an alternative. Treatment with anti-TB drugs for at least 6 months results in higher success rates when compared with treatment less than 6 months, even though the difference is small. Shorter treatment regimen may be recommended for those with compliance issues. There is also no evidence to support shorter anti-tuberculosis treatment regimens when compared to a 6-month treatment regimen. However recently, results from an international, randomized, controlled clinical trial indicate that a four-month daily treatment regimen containing high-dose, or "optimized", rifapentine with moxifloxacin (2PHZM/2PHM) is as safe and effective as the existing standard six-month daily regimen at curing drug-susceptible tuberculosis (TB) disease. If tuberculosis recurs, testing to determine which antibiotics it is sensitive to is important before determining treatment. If multiple drug-resistant TB (MDR-TB) is detected, treatment with at least four effective antibiotics for 18 to 24 months is recommended. Directly observed therapy, i.e., having a health care provider watch the person take their medications, is recommended by the World Health Organization (WHO) in an effort to reduce the number of people not appropriately taking antibiotics. The evidence to support this practice over people simply taking their medications independently is of poor quality. There is no strong evidence indicating that directly observed therapy improves the number of people who were cured or the number of people who complete their medicine. Moderate quality evidence suggests that there is also no difference if people are observed at home versus at a clinic, or by a family member versus a health care worker. Methods to remind people of the importance of treatment and appointments may result in a small but important improvement. There is also not enough evidence to support intermittent rifampicin-containing therapy given two to three times a week has equal effectiveness as daily dose regimen on improving cure rates and reducing relapsing rates. There is also not enough evidence on effectiveness of giving intermittent twice or thrice weekly short course regimen compared to daily dosing regimen in treating children with tuberculosis. Primary resistance occurs when a person becomes infected with a resistant strain of TB. A person with fully susceptible MTB may develop secondary (acquired) resistance during therapy because of inadequate treatment, not taking the prescribed regimen appropriately (lack of compliance), or using low-quality medication. Drug-resistant TB is a serious public health issue in many developing countries, as its treatment is longer and requires more expensive drugs. MDR-TB is defined as resistance to the two most effective first-line TB drugs: rifampicin and isoniazid. Extensively drug-resistant TB is also resistant to three or more of the six classes of second-line drugs. Totally drug-resistant TB is resistant to all currently used drugs. It was first observed in 2003 in Italy, but not widely reported until 2012, and has also been found in Iran and India. There is some efficacy for linezolid to treat those with XDR-TB but side effects and discontinuation of medications were common. Bedaquiline is tentatively supported for use in multiple drug-resistant TB. XDR-TB is a term sometimes used to define extensively resistant TB, and constitutes one in ten cases of MDR-TB. Cases of XDR TB have been identified in more than 90% of countries. For those with known rifampicin or MDR-TB, molecular tests such as the Genotype MTBDRsl Assay (performed on culture isolates or smear positive specimens) may be useful to detect second-line anti-tubercular drug resistance. Progression from TB infection to overt TB disease occurs when the bacilli overcome the immune system defenses and begin to multiply. In primary TB disease (some 1–5% of cases), this occurs soon after the initial infection. However, in the majority of cases, a latent infection occurs with no obvious symptoms. These dormant bacilli produce active tuberculosis in 5–10% of these latent cases, often many years after infection. The risk of reactivation increases with immunosuppression, such as that caused by infection with HIV. In people coinfected with M. tuberculosis and HIV, the risk of reactivation increases to 10% per year. Studies using DNA fingerprinting of M. tuberculosis strains have shown reinfection contributes more substantially to recurrent TB than previously thought, with estimates that it might account for more than 50% of reactivated cases in areas where TB is common. The chance of death from a case of tuberculosis is about 4% as of 2008, down from 8% in 1995. In people with smear-positive pulmonary TB (without HIV co-infection), after 5 years without treatment, 50-60% die while 20-25% achieve spontaneous resolution (cure). TB is almost always fatal in those with untreated HIV co-infection and death rates are increased even with antiretroviral treatment of HIV. Roughly one-quarter of the world's population has been infected with M. tuberculosis, with new infections occurring in about 1% of the population each year. However, most infections with M. tuberculosis do not cause disease, and 90–95% of infections remain asymptomatic. In 2012, an estimated 8.6 million chronic cases were active. In 2010, 8.8 million new cases of tuberculosis were diagnosed, and 1.20–1.45 million deaths occurred (most of these occurring in developing countries). Of these, about 0.35 million occur in those also infected with HIV. In 2018, tuberculosis was the leading cause of death worldwide from a single infectious agent. The total number of tuberculosis cases has been decreasing since 2005, while new cases have decreased since 2002. Tuberculosis incidence is seasonal, with peaks occurring every spring and summer. The reasons for this are unclear, but may be related to vitamin D deficiency during the winter. There are also studies linking tuberculosis to different weather conditions like low temperature, low humidity and low rainfall. It has been suggested that tuberculosis incidence rates may be connected to climate change. Tuberculosis is closely linked to both overcrowding and malnutrition, making it one of the principal diseases of poverty. Those at high risk thus include: people who inject illicit drugs, inhabitants and employees of locales where vulnerable people gather (e.g., prisons and homeless shelters), medically underprivileged and resource-poor communities, high-risk ethnic minorities, children in close contact with high-risk category patients, and health-care providers serving these patients. The rate of tuberculosis varies with age. In Africa, it primarily affects adolescents and young adults. However, in countries where incidence rates have declined dramatically (such as the United States), tuberculosis is mainly a disease of the elderly and immunocompromised (risk factors are listed above). Worldwide, 22 "high-burden" states or countries together experience 80% of cases as well as 83% of deaths. In Canada and Australia, tuberculosis is many times more common among the Indigenous peoples, especially in remote areas. Factors contributing to this include higher prevalence of predisposing health conditions and behaviours, and overcrowding and poverty. In some Canadian Indigenous groups, genetic susceptibility may play a role. Socioeconomic status (SES) strongly affects TB risk. People of low SES are both more likely to contract TB and to be more severely affected by the disease. Those with low SES are more likely to be affected by risk factors for developing TB (e.g., malnutrition, indoor air pollution, HIV co-infection, etc.), and are additionally more likely to be exposed to crowded and poorly ventilated spaces. Inadequate healthcare also means that people with active disease who facilitate spread are not diagnosed and treated promptly; sick people thus remain in the infectious state and (continue to) spread the infection. The distribution of tuberculosis is not uniform across the globe; about 80% of the population in many African, Caribbean, South Asian, and eastern European countries test positive in tuberculin tests, while only 5–10% of the U.S. population test positive. Hopes of totally controlling the disease have been dramatically dampened because of many factors, including the difficulty of developing an effective vaccine, the expensive and time-consuming diagnostic process, the necessity of many months of treatment, the increase in HIV-associated tuberculosis, and the emergence of drug-resistant cases in the 1980s. In developed countries, tuberculosis is less common and is found mainly in urban areas. In Europe, deaths from TB fell from 500 out of 100,000 in 1850 to 50 out of 100,000 by 1950. Improvements in public health were reducing tuberculosis even before the arrival of antibiotics, although the disease remained a significant threat to public health, such that when the Medical Research Council was formed in Britain in 1913 its initial focus was tuberculosis research. In 2010, rates per 100,000 people in different areas of the world were: globally 178, Africa 332, the Americas 36, Eastern Mediterranean 173, Europe 63, Southeast Asia 278, and Western Pacific 139. Russia has achieved particularly dramatic progress with a decline in its TB mortality rate—from 61.9 per 100,000 in 1965 to 2.7 per 100,000 in 1993; however, mortality rate increased to 24 per 100,000 in 2005 and then recoiled to 11 per 100,000 by 2015. China has achieved particularly dramatic progress, with about an 80% reduction in its TB mortality rate between 1990 and 2010. The number of new cases has declined by 17% between 2004 and 2014. In 2007, the country with the highest estimated incidence rate of TB was Eswatini, with 1,200 cases per 100,000 people. In 2017, the country with the highest estimated incidence rate as a % of the population was Lesotho, with 665 cases per 100,000 people. As of 2017, India had the largest total incidence, with an estimated 2,740,000 cases. According to the World Health Organization (WHO), in 2000–2015, India's estimated mortality rate dropped from 55 to 36 per 100,000 population per year with estimated 480 thousand people died of TB in 2015. In India a major proportion of tuberculosis patients are being treated by private partners and private hospitals. Evidence indicates that the tuberculosis national survey does not represent the number of cases that are diagnosed and recorded by private clinics and hospitals in India. In the United States, Native Americans have a fivefold greater mortality from TB, and racial and ethnic minorities accounted for 84% of all reported TB cases. The overall tuberculosis case rate in the United States was 3 per 100,000 persons in 2017. In Canada, tuberculosis was endemic in some rural areas as of 1998. In 2017, in the United Kingdom, the national average was 9 per 100,000 and the highest incidence rates in Western Europe were 20 per 100,000 in Portugal. Tuberculosis has existed since antiquity. The oldest unambiguously detected M. tuberculosis gives evidence of the disease in the remains of bison in Wyoming dated to around 17,000 years ago. However, whether tuberculosis originated in bovines, then transferred to humans, or whether both bovine and human tuberculosis diverged from a common ancestor, remains unclear. A comparison of the genes of M. tuberculosis complex (MTBC) in humans to MTBC in animals suggests humans did not acquire MTBC from animals during animal domestication, as researchers previously believed. Both strains of the tuberculosis bacteria share a common ancestor, which could have infected humans even before the Neolithic Revolution. Skeletal remains show some prehistoric humans (4000 BC) had TB, and researchers have found tubercular decay in the spines of Egyptian mummies dating from 3000 to 2400 BC. Genetic studies suggest the presence of TB in the Americas from about AD 100. Before the Industrial Revolution, folklore often associated tuberculosis with vampires. When one member of a family died from the disease, the other infected members would lose their health slowly. People believed this was caused by the original person with TB draining the life from the other family members. Although Richard Morton established the pulmonary form associated with tubercles as a pathology in 1689, due to the variety of its symptoms, TB was not identified as a single disease until the 1820s. Benjamin Marten conjectured in 1720 that consumptions were caused by microbes which were spread by people living close to each other. In 1819, René Laennec claimed that tubercles were the cause of pulmonary tuberculosis. J. L. Schönlein first published the name "tuberculosis" (German: Tuberkulose) in 1832. Between 1838 and 1845, John Croghan, the owner of Mammoth Cave in Kentucky from 1839 onwards, brought a number of people with tuberculosis into the cave in the hope of curing the disease with the constant temperature and purity of the cave air; each died within a year. Hermann Brehmer opened the first TB sanatorium in 1859 in Görbersdorf (now Sokołowsko) in Silesia. In 1865, Jean Antoine Villemin demonstrated that tuberculosis could be transmitted, via inoculation, from humans to animals and among animals. (Villemin's findings were confirmed in 1867 and 1868 by John Burdon-Sanderson.) Robert Koch identified and described the bacillus causing tuberculosis, M. tuberculosis, on 24 March 1882. In 1905, he was awarded the Nobel Prize in Physiology or Medicine for this discovery. In Europe, rates of tuberculosis began to rise in the early 1600s to a peak level in the 1800s, when it caused nearly 25% of all deaths. In the 18th and 19th century, tuberculosis had become epidemic in Europe, showing a seasonal pattern. Tuberculosis caused widespread public concern in the 19th and early 20th centuries as the disease became common among the urban poor. In 1815, one in four deaths in England was due to "consumption". By 1918, TB still caused one in six deaths in France. After TB was determined to be contagious, in the 1880s, it was put on a notifiable-disease list in Britain; campaigns started to stop people from spitting in public places, and the infected poor were "encouraged" to enter sanatoria that resembled prisons (the sanatoria for the middle and upper classes offered excellent care and constant medical attention). Whatever the benefits of the "fresh air" and labor in the sanatoria, even under the best conditions, 50% of those who entered died within five years (c. 1916). Robert Koch did not believe the cattle and human tuberculosis diseases were similar, which delayed the recognition of infected milk as a source of infection. During the first half of the 1900s, the risk of transmission from this source was dramatically reduced after the application of the pasteurization process. Koch announced a glycerine extract of the tubercle bacilli as a "remedy" for tuberculosis in 1890, calling it "tuberculin". Although it was not effective, it was later successfully adapted as a screening test for the presence of pre-symptomatic tuberculosis. World Tuberculosis Day is marked on 24 March each year, the anniversary of Koch's original scientific announcement. When the Medical Research Council formed in Britain in 1913, it initially focused on tuberculosis research. Albert Calmette and Camille Guérin achieved the first genuine success in immunization against tuberculosis in 1906, using attenuated bovine-strain tuberculosis. It was called bacille Calmette–Guérin (BCG). The BCG vaccine was first used on humans in 1921 in France, but achieved widespread acceptance in the US, Great Britain, and Germany only after World War II. By the 1950s mortality in Europe had decreased about 90%. Improvements in sanitation, vaccination, and other public-health measures began significantly reducing rates of tuberculosis even before the arrival of streptomycin and other antibiotics, although the disease remained a significant threat. In 1946, the development of the antibiotic streptomycin made effective treatment and cure of TB a reality. Prior to the introduction of this medication, the only treatment was surgical intervention, including the "pneumothorax technique", which involved collapsing an infected lung to "rest" it and to allow tuberculous lesions to heal. Because of the emergence of multidrug-resistant tuberculosis (MDR-TB), surgery has been re-introduced for certain cases of TB infections. It involves the removal of infected chest cavities ("bullae") in the lungs to reduce the number of bacteria and to increase exposure of the remaining bacteria to antibiotics in the bloodstream. Hopes of eliminating TB ended with the rise of drug-resistant strains in the 1980s. The subsequent resurgence of tuberculosis resulted in the declaration of a global health emergency by the World Health Organization (WHO) in 1993. Tuberculosis has been known by many names from the technical to the familiar. Phthisis (Φθισις) is a Greek word for consumption, an old term for pulmonary tuberculosis; around 460 BCE, Hippocrates described phthisis as a disease of dry seasons. The abbreviation TB is short for tubercle bacillus. Consumption was the most common nineteenth century English word for the disease, and was also in use well into the twentieth century. The Latin root con meaning 'completely' is linked to sumere meaning 'to take up from under'. In The Life and Death of Mr Badman by John Bunyan, the author calls consumption "the captain of all these men of death." "Great white plague" has also been used. Tuberculosis was for centuries associated with poetic and artistic qualities among those infected, and was also known as "the romantic disease". Major artistic figures such as the poets John Keats, Percy Bysshe Shelley, and Edgar Allan Poe, the composer Frédéric Chopin, the playwright Anton Chekhov, the novelists Franz Kafka, Katherine Mansfield, Charlotte Brontë, Fyodor Dostoevsky, Thomas Mann, W. Somerset Maugham, George Orwell, and Robert Louis Stevenson, and the artists Alice Neel, Jean-Antoine Watteau, Elizabeth Siddal, Marie Bashkirtseff, Edvard Munch, Aubrey Beardsley and Amedeo Modigliani either had the disease or were surrounded by people who did. A widespread belief was that tuberculosis assisted artistic talent. Physical mechanisms proposed for this effect included the slight fever and toxaemia that it caused, allegedly helping them to see life more clearly and to act decisively. Tuberculosis formed an often-reused theme in literature, as in Thomas Mann's The Magic Mountain, set in a sanatorium; in music, as in Van Morrison's song "T.B. Sheets"; in opera, as in Puccini's La bohème and Verdi's La Traviata; in art, as in Monet's painting of his first wife Camille on her deathbed; and in film, such as the 1945 The Bells of St. Mary's starring Ingrid Bergman as a nun with tuberculosis. In 2014, the WHO adopted the "End TB" strategy which aims to reduce TB incidence by 80% and TB deaths by 90% by 2030. The strategy contains a milestone to reduce TB incidence by 20% and TB deaths by 35% by 2020. However, by 2020 only a 9% reduction in incidence per population was achieved globally, with the European region achieving 19% and the African region achieving 16% reductions. Similarly, the number of deaths only fell by 14%, missing the 2020 milestone of a 35% reduction, with some regions making better progress (31% reduction in Europe and 19% in Africa). Correspondingly, also treatment, prevention and funding milestones were missed in 2020, for example only 6.3 million people were started on TB prevention short of the target of 30 million. The World Health Organization (WHO), the Bill and Melinda Gates Foundation, and the U.S. government are subsidizing a fast-acting diagnostic tuberculosis test for use in low- and middle-income countries as of 2012. In addition to being fast-acting, the test can determine if there is resistance to the antibiotic rifampicin which may indicate multi-drug resistant tuberculosis and is accurate in those who are also infected with HIV. Many resource-poor places as of 2011 have access to only sputum microscopy. India had the highest total number of TB cases worldwide in 2010, in part due to poor disease management within the private and public health care sector. Programs such as the Revised National Tuberculosis Control Program are working to reduce TB levels among people receiving public health care. A 2014 EIU-healthcare report finds there is a need to address apathy and urges for increased funding. The report cites among others Lucica Ditui "[TB] is like an orphan. It has been neglected even in countries with a high burden and often forgotten by donors and those investing in health interventions." Slow progress has led to frustration, expressed by the executive director of the Global Fund to Fight AIDS, Tuberculosis and Malaria – Mark Dybul: "we have the tools to end TB as a pandemic and public health threat on the planet, but we are not doing it." Several international organizations are pushing for more transparency in treatment, and more countries are implementing mandatory reporting of cases to the government as of 2014, although adherence is often variable. Commercial treatment providers may at times overprescribe second-line drugs as well as supplementary treatment, promoting demands for further regulations. The government of Brazil provides universal TB care, which reduces this problem. Conversely, falling rates of TB infection may not relate to the number of programs directed at reducing infection rates but may be tied to an increased level of education, income, and health of the population. Costs of the disease, as calculated by the World Bank in 2009 may exceed US$150 billion per year in "high burden" countries. Lack of progress eradicating the disease may also be due to lack of patient follow-up – as among the 250 million rural migrants in China. There is insufficient data to show that active contact tracing helps to improve case detection rates for tuberculosis. Interventions such as house-to-house visits, educational leaflets, mass media strategies, educational sessions may increase tuberculosis detection rates in short-term. There is no study that compares new methods of contact tracing such as social network analysis with existing contact tracing methods. Slow progress in preventing the disease may in part be due to stigma associated with TB. Stigma may be due to the fear of transmission from affected individuals. This stigma may additionally arise due to links between TB and poverty, and in Africa, AIDS. Such stigmatization may be both real and perceived; for example, in Ghana, individuals with TB are banned from attending public gatherings. Stigma towards TB may result in delays in seeking treatment, lower treatment compliance, and family members keeping cause of death secret – allowing the disease to spread further. In contrast, in Russia stigma was associated with increased treatment compliance. TB stigma also affects socially marginalized individuals to a greater degree and varies between regions. One way to decrease stigma may be through the promotion of "TB clubs", where those infected may share experiences and offer support, or through counseling. Some studies have shown TB education programs to be effective in decreasing stigma, and may thus be effective in increasing treatment adherence. Despite this, studies on the relationship between reduced stigma and mortality are lacking as of 2010, and similar efforts to decrease stigma surrounding AIDS have been minimally effective. Some have claimed the stigma to be worse than the disease, and healthcare providers may unintentionally reinforce stigma, as those with TB are often perceived as difficult or otherwise undesirable. A greater understanding of the social and cultural dimensions of tuberculosis may also help with stigma reduction. The BCG vaccine has limitations, and research to develop new TB vaccines is ongoing. A number of potential candidates are currently in phase I and II clinical trials. Two main approaches are used to attempt to improve the efficacy of available vaccines. One approach involves adding a subunit vaccine to BCG, while the other strategy is attempting to create new and better live vaccines. MVA85A, an example of a subunit vaccine, is in trials in South Africa as of 2006, is based on a genetically modified vaccinia virus. Vaccines are hoped to play a significant role in treatment of both latent and active disease. To encourage further discovery, researchers and policymakers are promoting new economic models of vaccine development as of 2006, including prizes, tax incentives, and advance market commitments. A number of groups, including the Stop TB Partnership, the South African Tuberculosis Vaccine Initiative, and the Aeras Global TB Vaccine Foundation, are involved with research. Among these, the Aeras Global TB Vaccine Foundation received a gift of more than $280 million (US) from the Bill and Melinda Gates Foundation to develop and license an improved vaccine against tuberculosis for use in high burden countries. A number of medications are being studied as of 2012 for multidrug-resistant tuberculosis, including bedaquiline and delamanid. Bedaquiline received U.S. Food and Drug Administration (FDA) approval in late 2012. The safety and effectiveness of these new agents are uncertain as of 2012, because they are based on the results of relatively small studies. However, existing data suggest that patients taking bedaquiline in addition to standard TB therapy are five times more likely to die than those without the new drug, which has resulted in medical journal articles raising health policy questions about why the FDA approved the drug and whether financial ties to the company making bedaquiline influenced physicians' support for its use. Steroids add-on therapy has not shown any benefits for active pulmonary tuberculosis infection. Mycobacteria infect many different animals, including birds, fish, rodents, and reptiles. The subspecies Mycobacterium tuberculosis, though, is rarely present in wild animals. An effort to eradicate bovine tuberculosis caused by Mycobacterium bovis from the cattle and deer herds of New Zealand has been relatively successful. Efforts in Great Britain have been less successful. As of 2015, tuberculosis appears to be widespread among captive elephants in the US. It is believed that the animals originally acquired the disease from humans, a process called reverse zoonosis. Because the disease can spread through the air to infect both humans and other animals, it is a public health concern affecting circuses and zoos.
{{italic title}} The word '''American''' is used to mean a person or a thing from the [[United States]] or any [[country]] in the [[Americas]] - [[North America]], [[Central America]], and [[South America]]. In [[English language|English]], the most common use of this word is to mean a person or a thing from the [[United States]]. Even though the [[island]]s in the [[Caribbean Sea]] are close to the Americas, people who speak English do not usually use the word "American" for people or things from these islands. Because the word 'American' can be unclear, some languages don’t use it to talk about people or things from the United States. Instead, they use more specific words, like ''estadounidense'' in [[Spanish language|Spanish]], ''usonano'' in [[Esperanto]], ''yhdysvaltalainen'' in [[Finnish language|Finnish]], and ''US-Amerikaner'' in [[German language|German]]. In English, people sometimes use the word '[[Yankee]]', though this usually means someone from [[New England]]. The word ''Usonian'' (from [[Usonia]]) was suggested by architect [[Frank Lloyd Wright]] as another way to say 'American'. A [[Native American]] is someone who is mostly descended from the people who lived in the Americas before the [[Europe]]ans arrived. Native Americans are also called [[First Nations]] and ''Indians''. There is no [[language]] "American." Some important languages used in the Americas are English, [[French language|French]], [[Portuguese language|Portuguese]], and [[Spanish language|Spanish]]. Some people still use Native American languages. [[Quechua]] is the largest. [[Category:North America]] [[Category:South America]] {{US-stub}}
The meaning of the word American in the English language varies according to the historical, geographical, and political context in which it is used. American is derived from America, a term originally denoting all of the Americas (also called the Western Hemisphere). In some expressions, it retains this Pan-American sense, but its usage has evolved over time and, for various historical reasons, the word came to denote people or things specifically from the United States of America. In modern English, American generally refers to persons or things related to the United States of America; among native English speakers this usage is almost universal, with any other use of the term requiring specification. However, some linguists in the past have argued that "American" should be widened to also include people or things from anywhere in the American continents. The word can be used as either an adjective or a noun (viz. a demonym). In adjectival use, it means "of or relating to the United States"; for example, "Elvis Presley was an American singer" or "the man prefers American English". In its noun form, the word generally means a resident or citizen of the U.S., but is also used for someone whose ethnic identity is simply "American". The noun is rarely used in English to refer to people not connected to the United States when intending a geographical meaning. When used with a grammatical qualifier, the adjective American can mean "of or relating to the Americas", as in Latin American or Indigenous American. Less frequently, the adjective can take this meaning without a qualifier, as in "American Spanish dialects and pronunciation differ by country" or the names of the Organization of American States and the American Registry for Internet Numbers (ARIN). A third use of the term pertains specifically to the indigenous peoples of the Americas, for instance, "In the 16th century, many Americans died from imported diseases during the European conquest", though this usage is rare, as "indigenous", "First Nations" or "Amerindian" are considered less confusing and generally more appropriate. Compound constructions which indicate a minority ethnic group, such as "African-Americans" likewise refer exclusively to people in or from the United States of America, as does the prefix "Americo-". For instance, the Americo-Liberians and their language Merico derive their name from the fact that they are descended from African-American settlers, i.e. Blacks who were formerly enslaved in the United States of America. French, German, Italian, Japanese, Hebrew, Arabic, and Russian speakers may use cognates of American to refer to inhabitants of the Americas or to U.S. nationals. They generally have other terms specific to U.S. nationals, such as the German US-Amerikaner, French étatsunien, Japanese beikokujin (米国人), and Italian statunitense. These specific terms may be less common than the term American. In French, états-unien, étas-unien or étasunien, from États-Unis d'Amérique ("United States of America"), is a rarely used word that distinguishes U.S. things and persons from the adjective américain, which denotes persons and things from the United States, but may also refer to "the Americas". Likewise, German's use of U.S.-amerikanisch and U.S.-Amerikaner observe this cultural distinction, solely denoting U.S. things and people. In normal parlance, the adjective "American" and its direct cognates are usually used if the context renders the nationality of the person clear. This differentiation is prevalent in German-speaking countries, as indicated by the style manual of the Neue Zürcher Zeitung (one of the leading German-language newspapers in Switzerland) which dismisses the term U.S.-amerikanisch as both 'unnecessary' and 'artificial' and recommends replacing it with amerikanisch. The respective guidelines of the foreign ministries of Austria, Germany and Switzerland all prescribe Amerikaner and amerikanisch in reference to the United States for official usage, making no mention of U.S.-Amerikaner or U.S.-amerikanisch. Portuguese has americano, denoting both a person or thing from the Americas and a U.S. national. For referring specifically to a U.S. national and things, some words used are estadunidense (also spelled estado-unidense, "United States person"), from Estados Unidos da América, and ianque ("Yankee")—both usages exist in Brazil, but are uncommon in Portugal—but the term most often used, and the only one in Portugal, is norte-americano, even though it could, as with its Spanish equivalent, apply to Canadians and Mexicans as well. In Spanish, americano denotes geographic and cultural origin in the New World, as well as (infrequently) a U.S. citizen; the more common term is estadounidense ("United States person"), which derives from Estados Unidos de América ("United States of America"). The Spanish term norteamericano ("North American") is frequently used to refer things and persons from the United States, but this term can also denote people and things from Canada and Mexico. Among Spanish-speakers, North America generally does not include Central America or the Caribbean. Conversely, in Czech, there is no possibility for disambiguation. Američan (m.) and američanka (f.) can refer to persons from the United States or from the continents of the Americas, and there is no specific word capable of distinguishing the two meanings. For this reason, the latter meaning is very rarely used, and word američan(ka) is used almost exclusively to refer to persons from the United States. The usage is exactly parallel to the English word. In other languages, however, there is no possibility for confusion. For example, the Chinese word for "U.S. national" is měiguórén (simplified Chinese: 美国人; traditional Chinese: 美國人) is derived from a word for the United States, měiguó, where měi is an abbreviation for Yàměilìjiā ("America") and guó is "country". The name for the American continents is měizhōu, from měi plus zhōu ("continent"). Thus, a měizhōurén is an American in the continent sense, and a měiguórén is an American in the U.S. sense. Korean and Vietnamese also use unambiguous terms, with Korean having Migug (미국(인)) for the country versus Amerika (아메리카) for the continents, and Vietnamese having Hoa Kỳ for the country versus Châu Mỹ for the continents. Japanese has such terms as well (beikoku(jin) [米国(人) versus beishū(jin) [米洲人]), but they are found more in newspaper headlines than in speech, where amerikajin predominates. In Swahili, Marekani means specifically the United States, and Mmarekani is a U.S. national, whereas the international form Amerika refers to the continents, and Mwamerika would be an inhabitant thereof. Likewise, the Esperanto word Ameriko refers to the continents. For the country there is the term Usono. Thus, a citizen of the United States is an usonano, whereas an amerikano is an inhabitant of the Americas. The name America was coined by Martin Waldseemüller from Americus Vespucius, the Latinized version of the name of Amerigo Vespucci (1454–1512), the Italian explorer who mapped South America's east coast and the Caribbean Sea in the early 16th century. Later, Vespucci's published letters were the basis of Waldseemüller's 1507 map, which is the first usage of America. The adjective American subsequently denoted the New World. In the 16th century, European usage of American denoted the native inhabitants of the New World. The earliest recorded use of this term in English is in Thomas Hacket's 1568 translation of André Thévet's book France Antarctique; Thévet himself had referred to the natives as Ameriques. In the following century, the term was extended to European settlers and their descendants in the Americas. The earliest recorded use of "English-American" dates to 1648, in Thomas Gage's The English-American his travail by sea and land: or, a new survey of the West India's. In English, American was used especially for people in British America. Samuel Johnson, the leading English lexicographer, wrote in 1775, before the United States declared independence: "That the Americans are able to bear taxation is indubitable." The Declaration of Independence of July 1776 refers to "[the] unanimous Declaration of the thirteen united States of America" adopted by the "Representatives of the united States of America" on July 4, 1776. The official name of the country was reaffirmed on November 15, 1777, when the Second Continental Congress adopted the Articles of Confederation, the first of which says, "The Stile of this Confederacy shall be 'The United States of America'". The Articles further state: In Witness whereof we have hereunto set our hands in Congress. Done at Philadelphia in the State of Pennsylvania the ninth day of July in the Year of our Lord One Thousand Seven Hundred and Seventy-Eight, and in the Third Year of the independence of America. Thomas Jefferson, newly elected president in May 1801 wrote, "I am sure the measures I mean to pursue are such as would in their nature be approved by every American who can emerge from preconceived prejudices; as for those who cannot, we must take care of them as of the sick in our hospitals. The medicine of time and fact may cure some of them." In The Federalist Papers (1787–88), Alexander Hamilton and James Madison used the adjective American with two different meanings: one political and one geographic; "the American republic" in Federalist No. 51 and in Federalist No. 70, and, in Federalist No. 24, Hamilton used American to denote the lands beyond the U.S.'s political borders. Early official U.S. documents show inconsistent usage; the 1778 Treaty of Alliance with France used "the United States of North America" in the first sentence, then "the said united States" afterwards; "the United States of America" and "the United States of North America" derive from "the United Colonies of America" and "the United Colonies of North America". The Treaty of Peace and Amity of September 5, 1795, between the United States and the Barbary States contains the usages "the United States of North America", "citizens of the United States", and "American Citizens". U.S. President George Washington, in his 1796 Farewell Address, declaimed that "The name of American, which belongs to you in your national capacity, must always exalt the just pride of patriotism more than any appellation." Political scientist Virginia L. Arbery notes that, in his Farewell Address: "...Washington invites his fellow citizens to view themselves now as Americans who, out of their love for the truth of liberty, have replaced their maiden names (Virginians, South Carolinians, New Yorkers, etc.) with that of “American”. Get rid of, he urges, “any appellation derived from local discriminations.” By defining himself as an American rather than as a Virginian, Washington set the national standard for all citizens. "Over and over, Washington said that America must be something set apart. As he put it to Patrick Henry, 'In a word, I want an American character, that the powers of Europe may be convinced we act for ourselves and not for others.'" As the historian Garry Wills has noted: "This was a theme dear to Washington. He wrote to Timothy Pickering that the nation 'must never forget that we are Americans; the remembrance of which will convince us we ought not to be French or English'." Washington's countrymen subsequently embraced his exhortation with notable enthusiasm. This semantic divergence among North American anglophones, however, remained largely unknown in the Spanish-American colonies. In 1801, the document titled Letter to American Spaniards—published in French (1799), in Spanish (1801), and in English (1808)—might have influenced Venezuela's Act of Independence and its 1811 constitution. The Latter-day Saints' Articles of Faith refer to the American continents as where they are to build Zion. Common short forms and abbreviations are the United States, the U.S., the U.S.A., and America; colloquial versions include the U.S. of A. and the States. The term Columbia (from the Columbus surname) was a popular name for the U.S. and for the entire geographic Americas; its usage is present today in the District of Columbia's name. Moreover, the womanly personification of Columbia appears in some official documents, including editions of the U.S. dollar. Use of the term American for U.S. nationals is common at the United Nations, and financial markets in the United States are referred to as "American financial markets". American Samoa, an unincorporated territory of the United States, is a recognized territorial name at the United Nations. Modern Canadians typically refer to people from the United States as Americans, though they seldom refer to the United States as America; they use the terms the United States, the U.S., or (informally) the States instead. Because of anti-American sentiment or simply national pride, Canadians never apply the term American to themselves. Not being an "American" is a part of Canadian identity, with many Canadians resenting being referred to as Americans or mistaken for U.S. citizens. This is often due to others' inability, particularly overseas, to distinguish Canadians from Americans, by their accent or other cultural attributes. Some Canadians have protested the use of American as a national demonym. People of U.S. ethnic origin in Canada are categorized as "Other North American origins" by Statistics Canada for purposes of census counts. The use of American as a national demonym for U.S. nationals is challenged, primarily by Latin Americans. Spanish speakers in Spain and Latin America use the term estadounidense to refer to people and things from the United States (from Estados Unidos), while americano refers to the continents as a whole. The term gringo is also accepted in many parts of Latin America to refer to a person or something from the United States; however, this term may be ambiguous in certain parts. Up to and including the 1992 edition, the Diccionario de la lengua española, published by the Real Academia Española, did not include the United States definition in the entry for americano; this was added in the 2001 edition. The Real Academia Española advised against using americanos exclusively for U.S. nationals: [Translated] It is common, and thus acceptable, to use norteamericano as a synonym of estadounidense, even though strictly speaking, the term norteamericano can equally be used to refer to the inhabitants of any country in North America, it normally applies to the inhabitants of the United States. But americano should not be used to refer exclusively to the inhabitants of the United States, an abusive usage which can be explained by the fact that in the United States, they frequently abbreviate the name of the country to "America" (in English, with no accent). Generally, americano denotes "U.S. citizen" in Portugal. Usage of americano to exclusively denote people and things of the U.S. is discouraged by the Lisbon Academy of Sciences, because the specific word estado-unidense (also estadunidense) clearly denotes a person from the United States. The term currently used by the Portuguese press is norte-americano. In Brazil, the term americano is used to address both that which pertains to the Americas and that which pertains to the U.S.; the particular meaning is deduced from context. Alternatively, the term norte-americano ("North American") is also used in more informal contexts, while estadunidense (of the U.S.) is the preferred form in academia. Use of the three terms is common in schools, government, and media. The term América is used exclusively for the whole continent, and the U.S. is called Estados Unidos ("United States") or Estados Unidos da América ("United States of America"), often abbreviated EUA. "American" in the 1994 Associated Press Stylebook was defined as, "An acceptable description for a resident of the United States. It also may be applied to any resident or citizen of nations in North or South America." Elsewhere, the AP Stylebook indicates that "United States" must "be spelled out when used as a noun. Use U.S. (no space) only as an adjective." The entry for "America" in The New York Times Manual of Style and Usage from 1999 reads: [the] terms "America", "American(s)" and "Americas" refer not only to the United States, but to all of North America and South America. They may be used in any of their senses, including references to just the United States, if the context is clear. The countries of the Western Hemisphere are collectively 'the Americas'. Media releases from the Pope and Holy See frequently use "America" to refer to the United States, and "American" to denote something or someone from the United States. At least one international law uses U.S. citizen in defining a citizen of the United States rather than American citizen; for example, the English version of the North American Free Trade Agreement includes: Only air carriers that are "citizens of the United States" may operate aircraft in domestic air service (cabotage) and may provide international scheduled and non-scheduled air service as U.S. air carriers... Under the Federal Aviation Act of 1958, a "citizen of the United States" means: Many international treaties use the terms American and American citizen: Products that are labeled, advertised, and marketed in the U.S. as "Made in the USA" must be, as set by the Federal Trade Commission (FTC), "all or virtually all made in the U.S." The FTC, to prevent deception of customers and unfair competition, considers an unqualified claim of "American Made" to expressly claim exclusive manufacture in the U.S: "The FTC Act gives the Commission the power to bring law enforcement actions against false or misleading claims that a product is of U.S. origin." There are a number of alternatives to the demonym American as a citizen of the United States that do not simultaneously mean any inhabitant of the Americas. One uncommon alternative is Usonian, which usually describes a certain style of residential architecture designed by Frank Lloyd Wright. Other alternatives have also surfaced, but most have fallen into disuse and obscurity. Merriam-Webster's Dictionary of English Usage says: The list contains (in approximate historical order from 1789 to 1939) such terms as Columbian, Columbard, Fredonian, Frede, Unisian, United Statesian, Colonican, Appalacian, Usian, Washingtonian, Usonian, Uessian, U-S-ian, Uesican, United Stater. Nevertheless, no alternative to American is common.
A '''clause''' is a part of a [[sentence]]. Each clause is made up of a [[Subject (grammar)|subject]] (who or what the sentence is about) and a [[Predicate (grammar)|predicate]] (what happens in a sentence). Each predicate has only one main [[verb]]. ''I love you'' is a sentence which has only one clause. ''I love you and I will always love you'' is a sentence which has two clauses. The two clauses are ''I love you'' and ''I will always love you''. These clauses are joined together by the word ''and'', which is a [[conjunction]]. Clauses may be independent or dependent. == In use == Two clauses can be joined with a [[pronoun]]. For example: ''I live in London, which is in England''. Here, ''I live in London'' is the first clause, and ''which is in England'' is the second clause. The word ''which'' is a pronoun which takes the place of ''London''. It joins the two clauses. A sentence can contain many clauses. But sentences with fewer clauses are easier to understand. == Dependent and independent clauses == A [[simple sentence]] may also be called an [[independent clause]]. It may be a part of a [[Sentence|compound]] or [[complex sentence]], but it can also stand on its own as a simple sentence (or independent clause).<ref>{{cite web |url=http://www2.ivcc.edu/rambo/eng1001/sentences.htm |title=Sentences: Simple, Compound, and Complex |author=Randy Rambo |date=2012 |website= |publisher=English Composition 1 |accessdate=27 December 2015 |archive-date=26 December 2015 |archive-url=https://web.archive.org/web/20151226102435/http://www2.ivcc.edu/rambo/eng1001/sentences.htm |url-status=dead }}</ref> A '''subordinate clause''' also called a '''dependent clause''' is one which cannot stand by itself.<ref name=GM>{{cite web |url=http://www.grammar-monster.com/glossary/subordinate_clause.htm |title=What Is a Subordinate Clause? (with Examples) |author= |website= |publisher=Grammar Monster |accessdate=27 December 2015}}</ref> This is because it does not express a complete thought.<ref name=GM/> It contains both a [[subject (grammar)|subject]] and a [[verb]].<ref>{{cite web |url=http://www.chompchomp.com/terms/subordinateclause.htm |title=The Subordinate Clause |author=Robin L. Simmons |date=1997 |website= |publisher=Grammar Bytes |accessdate=27 December 2015}}</ref> A subordinate clause always depends on a main clause.<ref>{{cite web |url=http://www.oxforddictionaries.com/words/clauses |title=Clauses |author= |website= |publisher=Oxford Dictionaries |accessdate=27 December 2015 |archive-date=3 January 2016 |archive-url=https://web.archive.org/web/20160103104731/http://www.oxforddictionaries.com/words/clauses |url-status=dead }}</ref> The main clause is almost always an independent clause, therefore the main clause by itself makes sense and can stand on its own. However, the subordinate clause does not. For [[example]], ''I love you'' makes perfect sense left on its own. However, ''and always will'', does not. The only time a sentence can be made up of only dependent clauses is when they are joined by [[Conjunction|correlative conjunctions]]: conjunction pairs like "either/or", "neither/nor", and "not only/but also". == References == {{reflist}} == Other websites == * [http://grammar.ccc.commnet.edu/grammar/clauses.htm Clauses: the Essential Building-Blocks] {{Webarchive|url=https://web.archive.org/web/20151221024130/http://grammar.ccc.commnet.edu/GRAMMAR/clauses.htm |date=2015-12-21 }} * [https://owl.english.purdue.edu/owl/resource/598/01/ Purdue Owl: Identifying Independent and Dependent Clauses] [[Category:Grammar]]
In language, a clause is a constituent that comprises a semantic predicand (expressed or not) and a semantic predicate. A typical clause consists of a subject and a syntactic predicate, the latter typically a verb phrase composed of a verb with any objects and other modifiers. However, the subject is sometimes unvoiced if it is retrievable from context, especially in null-subject language but also in other languages, including English instances of the imperative mood. A complete simple sentence contains a single clause with a finite verb. Complex sentences contain at least one clause subordinated (dependent) to an independent clause (one that could stand alone as a simple sentence), which may be co-ordinated with other independents with or without dependents. Some dependent clauses are non-finite. A primary division for the discussion of clauses is the distinction between independent clauses and dependent clauses. An independent clause can stand alone, i.e. it can constitute a complete sentence by itself. A dependent clause, by contrast, is reliant on the presence of an independent clause. A second major distinction concerns the difference between finite and non-finite clauses. A finite clause contains a structurally central finite verb, whereas the structurally central word of a non-finite clause is often a non-finite verb. Traditional grammar focuses on finite clauses, the awareness of non-finite clauses having arisen much later in connection with the modern study of syntax. The discussion here also focuses on finite clauses, although some aspects of non-finite clauses are considered further below. Clauses can be classified according to a distinctive trait that is a prominent characteristic of their syntactic form. The position of the finite verb is one major trait used for classification, and the appearance of a specific type of focusing word (e.g. wh-word) is another. These two criteria overlap to an extent, which means that often no single aspect of syntactic form is always decisive in determining how the clause functions. There are, however, strong tendencies. Standard SV-clauses (subject-verb) are the norm in English. They are usually declarative (as opposed to exclamative, imperative, or interrogative); they express information in a neutral manner, e.g. Declarative clauses like these are by far the most frequently occurring type of clause in any language. They can be viewed as basic, other clause types being derived from them. Standard SV-clauses can also be interrogative or exclamative, however, given the appropriate intonation contour and/or the appearance of a question word, e.g. Examples like these demonstrate that how a clause functions cannot be known based entirely on a single distinctive syntactic criterion. SV-clauses are usually declarative, but intonation and/or the appearance of a question word can render them interrogative or exclamative. Verb first clauses in English usually play one of three roles: 1. They express a yes/no-question via subject–auxiliary inversion, 2. they express a condition as an embedded clause, or 3. they express a command via imperative mood, e.g. Most verb first clauses are independent clauses. Verb first conditional clauses, however, must be classified as embedded clauses because they cannot stand alone. In English, Wh-clauses contain a wh-word. Wh-words often serve to help express a constituent question. They are also prevalent, though, as relative pronouns, in which case they serve to introduce a relative clause and are not part of a question. The wh-word focuses a particular constituent, and most of the time, it appears in clause-initial position. The following examples illustrate standard interrogative wh-clauses. The b-sentences are direct questions (independent clauses), and the c-sentences contain the corresponding indirect questions (embedded clauses): One important aspect of matrix wh-clauses is that subject-auxiliary inversion is obligatory when something other than the subject is focused. When it is the subject (or something embedded in the subject) that is focused, however, subject-auxiliary inversion does not occur. Another important aspect of wh-clauses concerns the absence of subject-auxiliary inversion in embedded clauses, as illustrated in the c-examples just produced. Subject-auxiliary inversion is obligatory in matrix clauses when something other than the subject is focused, but it never occurs in embedded clauses regardless of the constituent that is focused. A systematic distinction in word order emerges across matrix wh-clauses, which can have VS order, and embedded wh-clauses, which always maintain SV order, e.g. Relative clauses are a mixed group. In English they can be standard SV-clauses if they are introduced by that or lack a relative pronoun entirely, or they can be wh-clauses if they are introduced by a wh-word that serves as a relative pronoun. Embedded clauses can be categorized according to their syntactic function in terms of predicate-argument structures. They can function as arguments, as adjuncts, or as predicative expressions. That is, embedded clauses can be an argument of a predicate, an adjunct on a predicate, or (part of) the predicate itself. The predicate in question is usually the predicate of an independent clause, but embedding of predicates is also frequent. A clause that functions as the argument of a given predicate is known as an argument clause. Argument clauses can appear as subjects, as objects, and as obliques. They can also modify a noun predicate, in which case they are known as content clauses. The following examples illustrate argument clauses that provide the content of a noun. Such argument clauses are content clauses: The content clauses like these in the a-sentences are arguments. Relative clauses introduced by the relative pronoun that as in the b-clauses here have an outward appearance that is closely similar to that of content clauses. The relative clauses are adjuncts, however, not arguments. Adjunct clauses are embedded clauses that modify an entire predicate-argument structure. All clause types (SV-, verb first, wh-) can function as adjuncts, although the stereotypical adjunct clause is SV and introduced by a subordinator (i.e. subordinate conjunction, e.g. after, because, before, now, etc.), e.g. These adjunct clauses modify the entire matrix clause. Thus before you did in the first example modifies the matrix clause Fred arrived. Adjunct clauses can also modify a nominal predicate. The typical instance of this type of adjunct is a relative clause, e.g. An embedded clause can also function as a predicative expression. That is, it can form (part of) the predicate of a greater clause. These predicative clauses are functioning just like other predicative expressions, e.g. predicative adjectives (That was good) and predicative nominals (That was the truth). They form the matrix predicate together with the copula. Some of the distinctions presented above are represented in syntax trees. These trees make the difference between main and subordinate clauses very clear, and they also illustrate well the difference between argument and adjunct clauses. The following dependency grammar trees show that embedded clauses are dependent on an element in the independent clause, often on a verb: The independent clause comprises the entire trees in both instances, whereas the embedded clauses constitute arguments of the respective independent clauses: the embedded wh-clause what we want is the object argument of the predicate know; the embedded clause that he is gaining is the subject argument of the predicate is motivating. Both of these argument clauses are dependent on the verb of the matrix clause. The following trees identify adjunct clauses using an arrow dependency edge: These two embedded clauses are adjunct clauses because they provide circumstantial information that modifies a superordinate expression. The first is a dependent of the main verb of the matrix clause and the second is a dependent of the object noun. The arrow dependency edges identify them as adjuncts. The arrow points away from the adjunct towards it governor to indicate that semantic selection is running counter to the direction of the syntactic dependency; the adjunct is selecting its governor. The next four trees illustrate the distinction mentioned above between matrix wh-clauses and embedded wh-clauses The embedded wh-clause is an object argument each time. The position of the wh-word across the matrix clauses (a-trees) and the embedded clauses (b-trees) captures the difference in word order. Matrix wh-clauses have V2 word order, whereas embedded wh-clauses have (what amounts to) V3 word order. In the matrix clauses, the wh-word is a dependent of the finite verb, whereas it is the head over the finite verb in the embedded wh-clauses. There has been confusion about the distinction between clauses and phrases. This confusion is due in part to how these concepts are employed in the phrase structure grammars of the Chomskyan tradition. In the 1970s, Chomskyan grammars began labeling many clauses as CPs (i.e. complementizer phrases) or as IPs (i.e. inflection phrases), and then later as TPs (i.e. tense phrases), etc. The choice of labels was influenced by the theory-internal desire to use the labels consistently. The X-bar schema acknowledged at least three projection levels for every lexical head: a minimal projection (e.g. N, V, P, etc.), an intermediate projection (e.g. N', V', P', etc.), and a phrase level projection (e.g. NP, VP, PP, etc.). Extending this convention to the clausal categories occurred in the interest of the consistent use of labels. This use of labels should not, however, be confused with the actual status of the syntactic units to which the labels are attached. A more traditional understanding of clauses and phrases maintains that phrases are not clauses, and clauses are not phrases. There is a progression in the size and status of syntactic units: words < phrases < clauses. The characteristic trait of clauses, i.e. the presence of a subject and a (finite) verb, is absent from phrases. Clauses can be, however, embedded inside phrases. The central word of a non-finite clause is usually a non-finite verb (as opposed to a finite verb). There are various types of non-finite clauses that can be acknowledged based in part on the type of non-finite verb at hand. Gerunds are widely acknowledged to constitute non-finite clauses, and some modern grammars also judge many to-infinitives to be the structural locus of non-finite clauses. Finally, some modern grammars also acknowledge so-called small clauses, which often lack a verb altogether. It should be apparent that non-finite clauses are (by and large) embedded clauses. The underlined words in the following examples are considered non-finite clauses, e.g. Each of the gerunds in the a-sentences (stopping, attempting, and cheating) constitutes a non-finite clause. The subject-predicate relationship that has long been taken as the defining trait of clauses is fully present in the a-sentences. The fact that the b-sentences are also acceptable illustrates the enigmatic behavior of gerunds. They seem to straddle two syntactic categories: they can function as non-finite verbs or as nouns. When they function as nouns as in the b-sentences, it is debatable whether they constitute clauses, since nouns are not generally taken to be constitutive of clauses. Some modern theories of syntax take many to-infinitives to be constitutive of non-finite clauses. This stance is supported by the clear predicate status of many to-infinitives. It is challenged, however, by the fact that to-infinitives do not take an overt subject, e.g. The to-infinitives to consider and to explain clearly qualify as predicates (because they can be negated). They do not, however, take overt subjects. The subjects she and he are dependents of the matrix verbs refuses and attempted, respectively, not of the to-infinitives. Data like these are often addressed in terms of control. The matrix predicates refuses and attempted are control verbs; they control the embedded predicates consider and explain, which means they determine which of their arguments serves as the subject argument of the embedded predicate. Some theories of syntax posit the null subject PRO (i.e. pronoun) to help address the facts of control constructions, e.g. With the presence of PRO as a null subject, to-infinitives can be construed as complete clauses, since both subject and predicate are present. PRO-theory is particular to one tradition in the study of syntax and grammar (Government and Binding Theory, Minimalist Program). Other theories of syntax and grammar (e.g. Head-Driven Phrase Structure Grammar, Construction Grammar, dependency grammar) reject the presence of null elements such as PRO, which means they are likely to reject the stance that to-infinitives constitute clauses. Another type of construction that some schools of syntax and grammar view as non-finite clauses is the so-called small clause. A typical small clause consists of a noun phrase and a predicative expression, e.g. The subject-predicate relationship is clearly present in the underlined strings. The expression on the right is a predication over the noun phrase immediately to its left. While the subject-predicate relationship is indisputably present, the underlined strings do not behave as single constituents, a fact that undermines their status as clauses. Hence one can debate whether the underlined strings in these examples should qualify as clauses. The layered structures of the chomskyan tradition are again likely to view the underlined strings as clauses, whereas the schools of syntax that posit flatter structures are likely to reject clause status for them.
[[File:Bruce McCandless II during EVA in 1984.jpg|thumb|250px|Astronaut [[Bruce McCandless II]] using a [[space suit]] outside the [[United States]]' {{OV|99}} in 1984.]] An '''astronaut''' or '''cosmonaut''' is a person who goes into [[outer space]]. The [[Soviet Union]] and countries that it controlled used the word cosmonaut. Western countries including the [[United States]] said astronaut. Astronauts are also called "taikonauts" in [[China]] or "spationaute" in [[France]]. The first person to go into space was a [[Russia]]n from the [[Soviet Union]]. His name was [[Yuri Gagarin]]. This happened on [[Vostok 1|April 12, 1961]]. The first and second people to walk on the [[Moon]] were the [[United States|American]]s [[Neil Armstrong]] and [[Buzz Aldrin]]. This happened on [[Apollo 11|July 20, 1969]]. No astronauts have gone to the moon since 1972. No people have visited any other [[planet]]s yet. Astronauts used to go into space using many different ways, but now they can only go on the American [[SpaceX Dragon 2|Dragon 2]], the American [[Boeing Starliner|Starliner]], the Russian [[Soyuz]], or the Chinese Shenzhou. Several countries have worked together to build an [[International Space Station]] where people stay and work in space for long periods of time. A few countries and companies are trying to make more ways to get people into space. The United States is building a very big rocket called the [[Space Launch System]]. A United States company called SpaceX is also building a very large rocket called "[[SpaceX Starship|Starship]]" that is fully reusable and could take people to [[Mars]] and [[Solar System|further.]] Some American companies, for example [[Boeing]], [[Lockheed Martin]] and [[SpaceX]], are being paid by the United States to make ways for people to go to space. == Related pages == * [[Spaceflight]] {{science-stub}} [[Category:Astronauts| ]]
An astronaut (from the Ancient Greek ἄστρον (astron), meaning 'star', and ναύτης (nautes), meaning 'sailor') is a person trained, equipped, and deployed by a human spaceflight program to serve as a commander or crew member aboard a spacecraft. Although generally reserved for professional space travelers, the term is sometimes applied to anyone who travels into space, including scientists, politicians, journalists, and tourists. "Astronaut" technically applies to all human space travelers regardless of nationality. However, astronauts fielded by Russia or the Soviet Union are typically known instead as cosmonauts (from the Russian "kosmos" (космос), meaning "space", also borrowed from Greek κόσμος). Comparatively recent developments in crewed spaceflight made by China have led to the rise of the term taikonaut (from the Mandarin "tàikōng" (太空), meaning "space"), although its use is somewhat informal and its origin is unclear. In China, the People's Liberation Army Astronaut Corps astronauts and their foreign counterparts are all officially called hángtiānyuán (航天员, meaning "heaven navigator" or literally "heaven-sailing staff"). Since 1961, 600 astronauts have flown in space. Until 2002, astronauts were sponsored and trained exclusively by governments, either by the military or by civilian space agencies. With the suborbital flight of the privately funded SpaceShipOne in 2004, a new category of astronaut was created: the commercial astronaut. The criteria for what constitutes human spaceflight vary, with some focus on the point where the atmosphere becomes so thin that centrifugal force, rather than aerodynamic force, carries a significant portion of the weight of the flight object. The Fédération Aéronautique Internationale (FAI) Sporting Code for astronautics recognizes only flights that exceed the Kármán line, at an altitude of 100 kilometers (62 mi). In the United States, professional, military, and commercial astronauts who travel above an altitude of 80 kilometres (50 mi) are awarded astronaut wings. As of 17 November 2016, 552 people from 36 countries have reached 100 km (62 mi) or more in altitude, of whom 549 reached low Earth orbit or beyond. Of these, 24 people have traveled beyond low Earth orbit, either to lunar orbit, the lunar surface, or, in one case, a loop around the Moon. Three of the 24—Jim Lovell, John Young and Eugene Cernan—did so twice. As of 17 November 2016, under the U.S. definition, 558 people qualify as having reached space, above 50 miles (80 km) altitude. Of eight X-15 pilots who exceeded 50 miles (80 km) in altitude, only one, Joseph A. Walker, exceeded 100 kilometers (about 62.1 miles) and he did it two times, becoming the first person in space twice. Space travelers have spent over 41,790 man-days (114.5 man-years) in space, including over 100 astronaut-days of spacewalks. As of 2016, the man with the longest cumulative time in space is Gennady Padalka, who has spent 879 days in space. Peggy A. Whitson holds the record for the most time in space by a woman, at 675 days. In 1959, when both the United States and Soviet Union were planning, but had yet to launch humans into space, NASA Administrator T. Keith Glennan and his Deputy Administrator, Hugh Dryden, discussed whether spacecraft crew members should be called astronauts or cosmonauts. Dryden preferred "cosmonaut", on the grounds that flights would occur in and to the broader cosmos, while the "astro" prefix suggested flight specifically to the stars. Most NASA Space Task Group members preferred "astronaut", which survived by common usage as the preferred American term. When the Soviet Union launched the first man into space, Yuri Gagarin in 1961, they chose a term which anglicizes to "cosmonaut". A professional space traveler is called an astronaut. The first known use of the term "astronaut" in the modern sense was by Neil R. Jones in his 1930 short story "The Death's Head Meteor". The word itself had been known earlier; for example, in Percy Greg's 1880 book Across the Zodiac, "astronaut" referred to a spacecraft. In Les Navigateurs de l'infini (1925) by J.-H. Rosny aîné, the word astronautique (astronautics) was used. The word may have been inspired by "aeronaut", an older term for an air traveler first applied in 1784 to balloonists. An early use of "astronaut" in a non-fiction publication is Eric Frank Russell's poem "The Astronaut", appearing in the November 1934 Bulletin of the British Interplanetary Society. The first known formal use of the term astronautics in the scientific community was the establishment of the annual International Astronautical Congress in 1950, and the subsequent founding of the International Astronautical Federation the following year. NASA applies the term astronaut to any crew member aboard NASA spacecraft bound for Earth orbit or beyond. NASA also uses the term as a title for those selected to join its Astronaut Corps. The European Space Agency similarly uses the term astronaut for members of its Astronaut Corps. By convention, an astronaut employed by the Russian Federal Space Agency (or its predecessor, the Soviet space program) is called a cosmonaut in English texts. The word is an Anglicization of kosmonavt (Russian: космонавт Russian pronunciation: [kəsmɐˈnaft]). Other countries of the former Eastern Bloc use variations of the Russian kosmonavt, such as the Polish: kosmonauta (although Poles also used astronauta, and the two words are considered synonyms). Coinage of the term космонавт has been credited to Soviet aeronautics (or "cosmonautics") pioneer Mikhail Tikhonravov (1900–1974). The first cosmonaut was Soviet Air Force pilot Yuri Gagarin, also the first person in space. He was part of the first six Soviet citizens, with German Titov, Yevgeny Khrunov, Andriyan Nikolayev, Pavel Popovich, and Grigoriy Nelyubov, who were given the title of pilot-cosmonaut in January 1961. Valentina Tereshkova was the first female cosmonaut and the first and youngest woman to have flown in space with a solo mission on the Vostok 6 in 1963. On 14 March 1995, Norman Thagard became the first American to ride to space on board a Russian launch vehicle, and thus became the first "American cosmonaut". In Chinese, the term Yǔ háng yuán (宇航员, "cosmos navigating personnel") is used for astronauts and cosmonauts in general, while hángtiān yuán (航天员, "navigating celestial-heaven personnel") is used for Chinese astronauts. Here, hángtiān (航天, literally "heaven-navigating", or spaceflight) is strictly defined as the navigation of outer space within the local star system, i.e. Solar System. The phrase tàikōng rén (太空人, "spaceman") is often used in Hong Kong and Taiwan. The term taikonaut is used by some English-language news media organizations for professional space travelers from China. The word has featured in the Longman and Oxford English dictionaries, and the term became more common in 2003 when China sent its first astronaut Yang Liwei into space aboard the Shenzhou 5 spacecraft. This is the term used by Xinhua News Agency in the English version of the Chinese People's Daily since the advent of the Chinese space program. The origin of the term is unclear; as early as May 1998, Chiew Lee Yih (趙裡昱) from Malaysia, used it in newsgroups. For its 2022 Astronaut Group, the European Space Agency envisioned recruiting an astronaut with a physical disability, a category they called "parastronauts", with the intention but not guarantee of spaceflight. The categories of disability considered for the program were individuals with lower limb deficiency (either through amputation or congenital), leg length difference, or a short stature (less than 130 centimetres or 4 feet 3 inches). On 23 November 2022, John McFall was selected to be the first ESA parastronaut. With the rise of space tourism, NASA and the Russian Federal Space Agency agreed to use the term "spaceflight participant" to distinguish those space travelers from professional astronauts on missions coordinated by those two agencies. While no nation other than Russia (and previously the Soviet Union), the United States, and China have launched a crewed spacecraft, several other nations have sent people into space in cooperation with one of these countries, e.g. the Soviet-led Interkosmos program. Inspired partly by these missions, other synonyms for astronaut have entered occasional English usage. For example, the term spationaut (French: spationaute) is sometimes used to describe French space travelers, from the Latin word spatium for "space"; the Malay term angkasawan (deriving from angkasa meaning 'space') was used to describe participants in the Angkasawan program (note its similarity with the Indonesian term antariksawan). Plans of the Indian Space Research Organisation to launch its crewed Gaganyaan spacecraft have spurred at times public discussion if another term than astronaut should be used for the crew members, suggesting vyomanaut (from the Sanskrit word vyoman meaning 'sky' or 'space') or gagannaut (from the Sanskrit word gagan for 'sky'). In Finland, the NASA astronaut Timothy Kopra, a Finnish American, has sometimes been referred to as sisunautti, from the Finnish word sisu. Across Germanic languages, the word for "astronaut" typically translates to "space traveler", as it does with German's Raumfahrer, Dutch's ruimtevaarder, Swedish's rymdfarare, and Norwegian's romfarer. As of 2021 in the United States, astronaut status is conferred on a person depending on the authorizing agency: On July 20, 2021, the FAA issued an order redefining the eligibility criteria to be an astronaut in response to the private suborbital spaceflights of Jeff Bezos and Richard Branson. The new criteria states that one must have "[d]emonstrated activities during flight that were essential to public safety, or contributed to human space flight safety" in order to qualify as an astronaut. This new definition excludes Bezos and Branson. The first human in space was Soviet Yuri Gagarin, who was launched on 12 April 1961, aboard Vostok 1 and orbited around the Earth for 108 minutes. The first woman in space was Soviet Valentina Tereshkova, who launched on 16 June 1963, aboard Vostok 6 and orbited Earth for almost three days. Alan Shepard became the first American and second person in space on 5 May 1961, on a 15-minute sub-orbital flight aboard Freedom 7. The first American to orbit the Earth was John Glenn, aboard Friendship 7 on 20 February 1962. The first American woman in space was Sally Ride, during Space Shuttle Challenger's mission STS-7, on 18 June 1983. In 1992, Mae Jemison became the first African American woman to travel in space aboard STS-47. Cosmonaut Alexei Leonov was the first person to conduct an extravehicular activity (EVA), (commonly called a "spacewalk"), on 18 March 1965, on the Soviet Union's Voskhod 2 mission. This was followed two and a half months later by astronaut Ed White who made the first American EVA on NASA's Gemini 4 mission. The first crewed mission to orbit the Moon, Apollo 8, included American William Anders who was born in Hong Kong, making him the first Asian-born astronaut in 1968. The Soviet Union, through its Intercosmos program, allowed people from other "socialist" (i.e. Warsaw Pact and other Soviet-allied) countries to fly on its missions, with the notable exceptions of France and Austria participating in Soyuz TM-7 and Soyuz TM-13, respectively. An example is Czechoslovak Vladimír Remek, the first cosmonaut from a country other than the Soviet Union or the United States, who flew to space in 1978 on a Soyuz-U rocket. Rakesh Sharma became the first Indian citizen to travel to space. He was launched aboard Soyuz T-11, on 2 April 1984. On 23 July 1980, Pham Tuan of Vietnam became the first Asian in space when he flew aboard Soyuz 37. Also in 1980, Cuban Arnaldo Tamayo Méndez became the first person of Hispanic and black African descent to fly in space, and in 1983, Guion Bluford became the first African American to fly into space. In April 1985, Taylor Wang became the first ethnic Chinese person in space. The first person born in Africa to fly in space was Patrick Baudry (France), in 1985. In 1985, Saudi Arabian Prince Sultan Bin Salman Bin AbdulAziz Al-Saud became the first Arab Muslim astronaut in space. In 1988, Abdul Ahad Mohmand became the first Afghan to reach space, spending nine days aboard the Mir space station. With the increase of seats on the Space Shuttle, the U.S. began taking international astronauts. In 1983, Ulf Merbold of West Germany became the first non-US citizen to fly in a US spacecraft. In 1984, Marc Garneau became the first of eight Canadian astronauts to fly in space (through 2010). In 1985, Rodolfo Neri Vela became the first Mexican-born person in space. In 1991, Helen Sharman became the first Briton to fly in space. In 2002, Mark Shuttleworth became the first citizen of an African country to fly in space, as a paying spaceflight participant. In 2003, Ilan Ramon became the first Israeli to fly in space, although he died during a re-entry accident. On 15 October 2003, Yang Liwei became China's first astronaut on the Shenzhou 5 spacecraft. On 30 May 2020, Doug Hurley and Bob Behnken became the first astronauts to launch on a private crewed spacecraft, Crew Dragon. The youngest person to reach space is Oliver Daemen, who was 18 years and 11 months old when he made a suborbital spaceflight on Blue Origin NS-16. Daemen, who was a commercial passenger aboard the New Shepard, broke the record of Soviet cosmonaut Gherman Titov, who was 25 years old when he flew Vostok 2. Titov remains the youngest human to reach orbit; he rounded the planet 17 times. Titov was also the first person to suffer space sickness and the first person to sleep in space, twice. The oldest person to reach space is William Shatner, who was 90 years old when he made a suborbital spaceflight on Blue Origin NS-18. The oldest person to reach orbit is John Glenn, one of the Mercury 7, who was 77 when he flew on STS-95. The longest time spent in space was by Russian Valeri Polyakov, who spent 438 days there. As of 2006, the most spaceflights by an individual astronaut is seven, a record held by both Jerry L. Ross and Franklin Chang-Diaz. The farthest distance from Earth an astronaut has traveled was 401,056 km (249,205 mi), when Jim Lovell, Jack Swigert, and Fred Haise went around the Moon during the Apollo 13 emergency. The first civilian in space was Valentina Tereshkova aboard Vostok 6 (she also became the first woman in space on that mission). Tereshkova was only honorarily inducted into the USSR's Air Force, which did not accept female pilots at that time. A month later, Joseph Albert Walker became the first American civilian in space when his X-15 Flight 90 crossed the 100 kilometers (54 nautical miles) line, qualifying him by the international definition of spaceflight. Walker had joined the US Army Air Force but was not a member during his flight. The first people in space who had never been a member of any country's armed forces were both Konstantin Feoktistov and Boris Yegorov aboard Voskhod 1. The first non-governmental space traveler was Byron K. Lichtenberg, a researcher from the Massachusetts Institute of Technology who flew on STS-9 in 1983. In December 1990, Toyohiro Akiyama became the first paying space traveler and the first journalist in space for Tokyo Broadcasting System, a visit to Mir as part of an estimated $12 million (USD) deal with a Japanese TV station, although at the time, the term used to refer to Akiyama was "Research Cosmonaut". Akiyama suffered severe space sickness during his mission, which affected his productivity. The first self-funded space tourist was Dennis Tito on board the Russian spacecraft Soyuz TM-3 on 28 April 2001. The first person to fly on an entirely privately funded mission was Mike Melvill, piloting SpaceShipOne flight 15P on a suborbital journey, although he was a test pilot employed by Scaled Composites and not an actual paying space tourist. Jared Isaacman was the first person to self-fund a mission to orbit, commanding Inspiration4 in 2021. Nine others have paid Space Adventures to fly to the International Space Station: The first NASA astronauts were selected for training in 1959. Early in the space program, military jet test piloting and engineering training were often cited as prerequisites for selection as an astronaut at NASA, although neither John Glenn nor Scott Carpenter (of the Mercury Seven) had any university degree, in engineering or any other discipline at the time of their selection. Selection was initially limited to military pilots. The earliest astronauts for both the US and the USSR tended to be jet fighter pilots, and were often test pilots. Once selected, NASA astronauts go through twenty months of training in a variety of areas, including training for extravehicular activity in a facility such as NASA's Neutral Buoyancy Laboratory. Astronauts-in-training (astronaut candidates) may also experience short periods of weightlessness (microgravity) in an aircraft called the "Vomit Comet," the nickname given to a pair of modified KC-135s (retired in 2000 and 2004, respectively, and replaced in 2005 with a C-9) which perform parabolic flights. Astronauts are also required to accumulate a number of flight hours in high-performance jet aircraft. This is mostly done in T-38 jet aircraft out of Ellington Field, due to its proximity to the Johnson Space Center. Ellington Field is also where the Shuttle Training Aircraft is maintained and developed, although most flights of the aircraft are conducted from Edwards Air Force Base. Astronauts in training must learn how to control and fly the Space Shuttle; further, it is vital that they are familiar with the International Space Station so they know what they must do when they get there. The master's degree requirement can also be met by: Mission Specialist Educators, or "Educator Astronauts", were first selected in 2004; as of 2007, there are three NASA Educator astronauts: Joseph M. Acaba, Richard R. Arnold, and Dorothy Metcalf-Lindenburger. Barbara Morgan, selected as back-up teacher to Christa McAuliffe in 1985, is considered to be the first Educator astronaut by the media, but she trained as a mission specialist. The Educator Astronaut program is a successor to the Teacher in Space program from the 1980s. Astronauts are susceptible to a variety of health risks including decompression sickness, barotrauma, immunodeficiencies, loss of bone and muscle, loss of eyesight, orthostatic intolerance, sleep disturbances, and radiation injury. A variety of large scale medical studies are being conducted in space via the National Space Biomedical Research Institute (NSBRI) to address these issues. Prominent among these is the Advanced Diagnostic Ultrasound in Microgravity Study in which astronauts (including former ISS commanders Leroy Chiao and Gennady Padalka) perform ultrasound scans under the guidance of remote experts to diagnose and potentially treat hundreds of medical conditions in space. This study's techniques are now being applied to cover professional and Olympic sports injuries as well as ultrasound performed by non-expert operators in medical and high school students. It is anticipated that remote guided ultrasound will have application on Earth in emergency and rural care situations, where access to a trained physician is often rare. A 2006 Space Shuttle experiment found that Salmonella typhimurium, a bacterium that can cause food poisoning, became more virulent when cultivated in space. More recently, in 2017, bacteria were found to be more resistant to antibiotics and to thrive in the near-weightlessness of space. Microorganisms have been observed to survive the vacuum of outer space. On 31 December 2012, a NASA-supported study reported that human spaceflight may harm the brain and accelerate the onset of Alzheimer's disease. In October 2015, the NASA Office of Inspector General issued a health hazards report related to space exploration, including a human mission to Mars. Over the last decade, flight surgeons and scientists at NASA have seen a pattern of vision problems in astronauts on long-duration space missions. The syndrome, known as visual impairment intracranial pressure (VIIP), has been reported in nearly two-thirds of space explorers after long periods spent aboard the International Space Station (ISS). On 2 November 2017, scientists reported that significant changes in the position and structure of the brain have been found in astronauts who have taken trips in space, based on MRI studies. Astronauts who took longer space trips were associated with greater brain changes. Being in space can be physiologically deconditioning on the body. It can affect the otolith organs and adaptive capabilities of the central nervous system. Zero gravity and cosmic rays can cause many implications for astronauts. In October 2018, NASA-funded researchers found that lengthy journeys into outer space, including travel to the planet Mars, may substantially damage the gastrointestinal tissues of astronauts. The studies support earlier work that found such journeys could significantly damage the brains of astronauts, and age them prematurely. Researchers in 2018 reported, after detecting the presence on the International Space Station (ISS) of five Enterobacter bugandensis bacterial strains, none pathogenic to humans, that microorganisms on ISS should be carefully monitored to continue assuring a medically healthy environment for astronauts. A study by Russian scientists published in April 2019 stated that astronauts facing space radiation could face temporary hindrance of their memory centers. While this does not affect their intellectual capabilities, it temporarily hinders formation of new cells in brain's memory centers. The study conducted by Moscow Institute of Physics and Technology (MIPT) concluded this after they observed that mice exposed to neutron and gamma radiation did not impact the rodents' intellectual capabilities. A 2020 study conducted on the brains of eight male Russian cosmonauts after they returned from long stays aboard the International Space Station showed that long-duration spaceflight causes many physiological adaptions, including macro- and microstructural changes. While scientists still know little about the effects of spaceflight on brain structure, this study showed that space travel can lead to new motor skills (dexterity), but also slightly weaker vision, both of which could possibly be long lasting. It was the first study to provide clear evidence of sensorimotor neuroplasticity, which is the brain's ability to change through growth and reorganization. An astronaut on the International Space Station requires about 830 g (29 oz) mass of food per meal each day (inclusive of about 120 g or 4.2 oz packaging mass per meal). Space Shuttle astronauts worked with nutritionists to select menus that appealed to their individual tastes. Five months before flight, menus were selected and analyzed for nutritional content by the shuttle dietician. Foods are tested to see how they will react in a reduced gravity environment. Caloric requirements are determined using a basal energy expenditure (BEE) formula. On Earth, the average American uses about 35 US gallons (130 L) of water every day. On board the ISS astronauts limit water use to only about three US gallons (11 L) per day. In Russia, cosmonauts are awarded Pilot-Cosmonaut of the Russian Federation upon completion of their missions, often accompanied with the award of Hero of the Russian Federation. This follows the practice established in the USSR where cosmonauts were usually awarded the title Hero of the Soviet Union. At NASA, those who complete astronaut candidate training receive a silver lapel pin. Once they have flown in space, they receive a gold pin. U.S. astronauts who also have active-duty military status receive a special qualification badge, known as the Astronaut Badge, after participation on a spaceflight. The United States Air Force also presents an Astronaut Badge to its pilots who exceed 50 miles (80 km) in altitude. As of 2020, eighteen astronauts (fourteen men and four women) have died during four space flights. By nationality, thirteen were American, four were Russian (Soviet Union), and one was Israeli. As of 2020, eleven people (all men) have died training for spaceflight: eight Americans and three Russians. Six of these were in crashes of training jet aircraft, one drowned during water recovery training, and four were due to fires in pure oxygen environments. Astronaut David Scott left a memorial consisting of a statuette titled Fallen Astronaut on the surface of the Moon during his 1971 Apollo 15 mission, along with a list of the names of eight of the astronauts and six cosmonauts known at the time to have died in service. The Space Mirror Memorial, which stands on the grounds of the Kennedy Space Center Visitor Complex, is maintained by the Astronauts Memorial Foundation and commemorates the lives of the men and women who have died during spaceflight and during training in the space programs of the United States. In addition to twenty NASA career astronauts, the memorial includes the names of an X-15 test pilot, a U.S. Air Force officer who died while training for a then-classified military space program, and a civilian spaceflight participant.
A '''hacker''' is someone who tries to get into another person's computer using [[computer software]]. The computer software could be [[Trojan horse (computing)|Trojan horse]] programs, [[computer virus]]es, and worms. It used to mean someone who likes to do new things with [[computer]]s. The correct name for a person doing these illegal things was ''cracker'', but many news stories used the word ''hacker'' even though it was in error.<ref>Manthan M Desai, ''Hacking For Beginners: a beginners guide to learn ethical hacking'', hackingtech.co.tv Revised Android Edition (Ht Hackingtech, 2012), p. 13</ref> Another meaning of ''hacker'' is someone who can change or program something, like how people can change or program a [[Furby]] to say something insulting or [[humor]]ous. A person doing something illegal with a [[computer]] belonging to someone else without asking for permission from the owner is also called a hacker. In the recent years computers were becoming increasingly more complex. This also means that people with bad intentions have more opportunities to use them for criminal purposes. Computer engineers are not always able to find vulnerable places in the computer code. Hackers find such places, and use them. There are also hackers who use the same skills for good purposes. Such hackers are called ''white hat'' hackers. They help computer engineers to find vulnerable places and correct problems. A lot of skills are often required to be a hacker. A hacker should be able to find problems or solutions that most other people cannot find. == References == {{reflist}} {{Tech-stub}} [[Category:Hackers| ]]
A hacker is a person skilled in information technology who achieves goals by non-standard means. Though the term hacker has become associated in popular culture with a security hacker – someone with knowledge of bugs or exploits to break into computer systems and access data which would otherwise be inaccessible to them – hacking can also be utilized by legitimate figures in legal situations. For example, law enforcement agencies sometimes use hacking techniques to collect evidence on criminals and other malicious actors. This could include using anonymity tools (such as a VPN or the dark web) to mask their identities online and pose as criminals. Likewise, covert world agencies can employ hacking techniques in the legal conduct of their work. Hacking and cyber-attacks are used extra-legally and illegally by law enforcement and security agencies (conducting warrantless activities), and employed by state actors as a weapon of legal and illegal warfare. Reflecting the two types of hackers, there are two definitions of the word "hacker": Mainstream usage of "hacker" mostly refers to computer criminals, due to the mass media usage of the word since the 1990s. This includes what hacker slang calls "script kiddies", people breaking into computers using programs written by others, with very little knowledge about the way they work. This usage has become so predominant that the general public is largely unaware that different meanings exist. Though the self-designation of hobbyists as hackers is generally acknowledged and accepted by computer security hackers, people from the programming subculture consider the computer intrusion related usage incorrect, and emphasize the difference between the two by calling security breakers "crackers" (analogous to a safecracker). The controversy is usually based on the assertion that the term originally meant someone messing about with something in a positive sense, that is, using playful cleverness to achieve a goal. But then, it is supposed, the meaning of the term shifted over the decades and came to refer to computer criminals. As the security-related usage has spread more widely, the original meaning has become less known. In popular usage and in the media, "computer intruders" or "computer criminals" is the exclusive meaning of the word. In computer enthusiast and hacker culture, the primary meaning is a complimentary description for a particularly brilliant programmer or technical expert. A large segment of the technical community insist the latter is the correct usage, as in the Jargon File definition. The mainstream media's current usage of the term may be traced back to the early 1980s. When the term, previously used only among computer enthusiasts, was introduced to wider society by the mainstream media in 1983, even those in the computer community referred to computer intrusion as "hacking", although not as the exclusive definition of the word. In reaction to the increasing media use of the term exclusively with the criminal connotation, the computer community began to differentiate their terminology. Alternative terms such as "cracker" were coined in an effort to maintain the distinction between "hackers" within the legitimate programmer community and those performing computer break-ins. Further terms such as "black hat", "white hat" and "gray hat" developed when laws against breaking into computers came into effect, to distinguish criminal activities from those activities which were legal. Network news' use of the term consistently pertains primarily to criminal activities, despite attempts by the technical community to preserve and distinguish the original meaning. Today, the mainstream media and general public continue to describe computer criminals, with all levels of technical sophistication, as "hackers" and do not generally make use of the word in any of its non-criminal connotations. Members of the media sometimes seem unaware of the distinction, grouping legitimate "hackers" such as Linus Torvalds and Steve Wozniak along with criminal "crackers". As a result, the definition is still the subject of heated controversy. The wider dominance of the pejorative connotation is resented by many who object to the term being taken from their cultural jargon and used negatively, including those who have historically preferred to self-identify as hackers. Many advocate using the more recent and nuanced alternate terms when describing criminals and others who negatively take advantage of security flaws in software and hardware. Others prefer to follow common popular usage, arguing that the positive form is confusing and unlikely to become widespread in the general public. A minority still use the term in both senses despite the controversy, leaving context to clarify (or leave ambiguous) which meaning is intended. However, because the positive definition of hacker was widely used as the predominant form for many years before the negative definition was popularized, "hacker" can therefore be seen as a shibboleth, identifying those who use the technically-oriented sense (as opposed to the exclusively intrusion-oriented sense) as members of the computing community. On the other hand, due to the variety of industries software designers may find themselves in, many prefer not to be referred to as hackers because the word holds a negative denotation in many of those industries. A possible middle ground position has been suggested, based on the observation that "hacking" describes a collection of skills and tools which are used by hackers of both descriptions for differing reasons. The analogy is made to locksmithing, specifically picking locks, which is a skill which can be used for good or evil. The primary weakness of this analogy is the inclusion of script kiddies in the popular usage of "hacker", despite their lack of an underlying skill and knowledge base. Sometimes, "hacker" is simply used synonymously with "geek": "A true hacker is not a group person. He's a person who loves to stay up all night, he and the machine in a love-hate relationship... They're kids who tended to be brilliant but not very interested in conventional goals It's a term of derision and also the ultimate compliment." Fred Shapiro thinks that "the common theory that 'hacker' originally was a benign term and the malicious connotations of the word were a later perversion is untrue." He found that the malicious connotations were already present at MIT in 1963 (quoting The Tech, an MIT student newspaper), and at that time referred to unauthorized users of the telephone network, that is, the phreaker movement that developed into the computer security hacker subculture of today. Hacker culture is an idea derived from a community of enthusiast computer programmers and systems designers in the 1960s around the Massachusetts Institute of Technology's (MIT's) Tech Model Railroad Club (TMRC) and the MIT Artificial Intelligence Laboratory. The concept expanded to the hobbyist home computing community, focusing on hardware in the late 1970s (e.g. the Homebrew Computer Club) and on software (video games, software cracking, the demoscene) in the 1980s/1990s. Later, this would go on to encompass many new definitions such as art, and life hacking. Security hackers are people involved with circumvention of computer security. Among security hackers, there are several types, including: White hats are hackers who work to keep data safe from other hackers by finding system vulnerabilities that can be mitigated. White hats are usually employed by the target system's owner and are typically paid (sometimes quite well) for their work. Their work is not illegal because it is done with the system owner's consent. Black hats or crackers are hackers with malicious intentions. They often steal, exploit, and sell data, and are usually motivated by personal gain. Their work is usually illegal. A cracker is like a black hat hacker, but is specifically someone who is very skilled and tries via hacking to make profits or to benefit, not just to vandalize. Crackers find exploits for system vulnerabilities and often use them to their advantage by either selling the fix to the system owner or selling the exploit to other black hat hackers, who in turn use it to steal information or gain royalties. A grey hat is a computer hacker or computer security expert who may sometimes violate laws or typical ethical standards, but does not have the malicious intent typical of a black hat hacker. Civic hackers use their security and/or programming acumens to create solutions, often public and open-sourced, addressing challenges relevant to neighborhoods, cities, states or countries and the infrastructure within them. Municipalities and major government agencies such as NASA have been known to host hackathons or promote a specific date as a "National Day of Civic Hacking" to encourage participation from civic hackers. Civic hackers, though often operating autonomously and independently, may work alongside or in coordination with certain aspects of government or local infrastructure such as trains and buses. For example, in 2008, Philadelphia-based civic hacker William Entriken developed a web application that displayed a comparison of the actual arrival times of local SEPTA trains to their scheduled times after being reportedly frustrated by the discrepancy. Four primary motives have been proposed as possibilities for why hackers attempt to break into computers and networks. First, there is a criminal financial gain to be had when hacking systems with the specific purpose of stealing credit card numbers or manipulating banking systems. Second, many hackers thrive off of increasing their reputation within the hacker subculture and will leave their handles on websites they defaced or leave some other evidence as proof that they were involved in a specific hack. Third, corporate espionage allows companies to acquire information on products or services that can be stolen or used as leverage within the marketplace. Lastly, state-sponsored attacks provide nation states with both wartime and intelligence collection options conducted on, in, or through cyberspace. The main basic difference between programmer subculture and computer security hacker is their mostly separate historical origin and development. However, the Jargon File reports that considerable overlap existed for the early phreaking at the beginning of the 1970s. An article from MIT's student paper The Tech used the term hacker in this context already in 1963 in its pejorative meaning for someone messing with the phone system. The overlap quickly started to break when people joined in the activity who did it in a less responsible way. This was the case after the publication of an article exposing the activities of Draper and Engressia. According to Raymond, hackers from the programmer subculture usually work openly and use their real name, while computer security hackers prefer secretive groups and identity-concealing aliases. Also, their activities in practice are largely distinct. The former focus on creating new and improving existing infrastructure (especially the software environment they work with), while the latter primarily and strongly emphasize the general act of circumvention of security measures, with the effective use of the knowledge (which can be to report and help fixing the security bugs, or exploitation reasons) being only rather secondary. The most visible difference in these views was in the design of the MIT hackers' Incompatible Timesharing System, which deliberately did not have any security measures. There are some subtle overlaps, however, since basic knowledge about computer security is also common within the programmer subculture of hackers. For example, Ken Thompson noted during his 1983 Turing Award lecture that it is possible to add code to the UNIX "login" command that would accept either the intended encrypted password or a particular known password, allowing a backdoor into the system with the latter password. He named his invention the "Trojan horse". Furthermore, Thompson argued, the C compiler itself could be modified to automatically generate the rogue code, to make detecting the modification even harder. Because the compiler is itself a program generated from a compiler, the Trojan horse could also be automatically installed in a new compiler program, without any detectable modification to the source of the new compiler. However, Thompson disassociated himself strictly from the computer security hackers: "I would like to criticize the press in its handling of the 'hackers,' the 414 gang, the Dalton gang, etc. The acts performed by these kids are vandalism at best and probably trespass and theft at worst. ... I have watched kids testifying before Congress. It is clear that they are completely unaware of the seriousness of their acts." The programmer subculture of hackers sees secondary circumvention of security mechanisms as legitimate if it is done to get practical barriers out of the way for doing actual work. In special forms, that can even be an expression of playful cleverness. However, the systematic and primary engagement in such activities is not one of the actual interests of the programmer subculture of hackers and it does not have significance in its actual activities, either. A further difference is that, historically, members of the programmer subculture of hackers were working at academic institutions and used the computing environment there. In contrast, the prototypical computer security hacker had access exclusively to a home computer and a modem. However, since the mid-1990s, with home computers that could run Unix-like operating systems and with inexpensive internet home access being available for the first time, many people from outside of the academic world started to take part in the programmer subculture of hacking. Since the mid-1980s, there are some overlaps in ideas and members with the computer security hacking community. The most prominent case is Robert T. Morris, who was a user of MIT-AI, yet wrote the Morris worm. The Jargon File hence calls him "a true hacker who blundered". Nevertheless, members of the programmer subculture have a tendency to look down on and disassociate from these overlaps. They commonly refer disparagingly to people in the computer security subculture as crackers and refuse to accept any definition of hacker that encompasses such activities. The computer security hacking subculture, on the other hand, tends not to distinguish between the two subcultures as harshly, acknowledging that they have much in common including many members, political and social goals, and a love of learning about technology. They restrict the use of the term cracker to their categories of script kiddies and black hat hackers instead. All three subcultures have relations to hardware modifications. In the early days of network hacking, phreaks were building blue boxes and various variants. The programmer subculture of hackers has stories about several hardware hacks in its folklore, such as a mysterious "magic" switch attached to a PDP-10 computer in MIT's AI lab that, when switched off, crashed the computer. The early hobbyist hackers built their home computers themselves from construction kits. However, all these activities have died out during the 1980s when the phone network switched to digitally controlled switchboards, causing network hacking to shift to dialing remote computers with modems when pre-assembled inexpensive home computers were available and when academic institutions started to give individual mass-produced workstation computers to scientists instead of using a central timesharing system. The only kind of widespread hardware modification nowadays is case modding. An encounter of the programmer and the computer security hacker subculture occurred at the end of the 1980s, when a group of computer security hackers, sympathizing with the Chaos Computer Club (which disclaimed any knowledge in these activities), broke into computers of American military organizations and academic institutions. They sold data from these machines to the Soviet secret service, one of them in order to fund his drug addiction. The case was solved when Clifford Stoll, a scientist working as a system administrator, found ways to log the attacks and to trace them back (with the help of many others). 23, a German film adaption with fictional elements, shows the events from the attackers' perspective. Stoll described the case in his book The Cuckoo's Egg and in the TV documentary The KGB, the Computer, and Me from the other perspective. According to Eric S. Raymond, it "nicely illustrates the difference between 'hacker' and 'cracker'. Stoll's portrait of himself, his lady Martha, and his friends at Berkeley and on the Internet paints a marvelously vivid picture of how hackers and the people around them like to live and how they think."
[[File:WTUL Microphone.jpg|right|220px]] In [[communications]], such as [[radio]] and [[television]], '''broadcasting''' means sending information such as television shows or music [[electronics|electronically]] to a large audience. The information is sent through the air in [[radio waves]], through a [[coaxial cable|wire]], or by a [[communications satellite]], and then the television viewers or radio listeners pick up the signal using their television sets and radio receivers. [[Gugliemo Marconi]] invented wireless [[telegraph]]y, in December 1901 he transmitted first radio signals across the [[Atlantic Ocean]]. This was point to point. Experiments with voice broadcasting began a few years later, and it grew rapidly in the 1920s. ==Types of broadcasters== Different broadcasters use different radio waves and different modulating methods. [[amplitude modulation]] on medium waves was the first to be much used. Some [[television network]]s are said to be broadcasting, even if they are only or mainly on [[cable TV]] rather than by radio waves. === Public broadcasters === Many countries have [[Public broadcasting]], using money from the government to broadcast television shows and radio programs. Examples include the [[BBC]] in [[Britain]], [[NHK]] in [[Japan]], and the [[CBC]] in [[Canada]]. In the [[United States]], most public broadcast radio and television stations are run by educational groups (such as colleges or universities) or by a state's educational department. Public broadcasters, by law, cannot accept or display commercial [[advertising]]. However, businesses can make [[donation]]s to a public broadcaster. Most are part of the [[Public Broadcasting System]]. [[PBS]] is different than the other public broadcasters such as [[BBC]], [[NHK]] and [[Canadian Broadcasting Corporation|CBC]], because the PBS gets a lot of its funding (money) from donations by viewers and listeners. Public broadcasters make programs that the private companies are not interested in making, such as educational children's shows, [[documentary|documentaries]], and public affairs shows about current issues. === Private broadcasters === As well, there are private broadcasting companies. These are companies that broadcast television and radio programs. To make money, private broadcasting companies sell [[advertising|advertisements]] called commercials. ===Community broadcasters=== A third type of broadcaster is community broadcasters. There are community television stations and community radio stations. Community television stations are often provided on cable networks. Community television stations usually have shows about local issues and community events. Some community television stations film and broadcast community cultural activities, such as musical performances or town hall meetings. Community radio stations play music and have public affairs shows about community issues. Community radio stations are usually small organizations that are run by [[Volunteering|volunteers]]. Community radio stations often get their funding (money) from local governments, local universities, and from donations by listeners. Some community radio stations also have poetry readings by local poets, or performances by local musicians or singers. ==Other meanings== '''Broadcasting''' can also mean sending a [[:wikt:message|message]] to many users on a [[computer]] [[network]] at exactly the same time, or sending a message from one computer to many other computers, giving information about itself, such as its name and location. Sending information to a small selected group is called ''narrowcasting''. {{commonscat}} [[Category:Broadcasting| ]] [[el:Ραδιοφωνία]]
Broadcasting is the distribution of audio or video content to a dispersed audience via any electronic mass communications medium, but typically one using the electromagnetic spectrum (radio waves), in a one-to-many model. Broadcasting began with AM radio, which came into popular use around 1920 with the spread of vacuum tube radio transmitters and receivers. Before this, most implementations of electronic communication (early radio, telephone, and telegraph) were one-to-one, with the message intended for a single recipient. The term broadcasting evolved from its use as the agricultural method of sowing seeds in a field by casting them broadly about. It was later adopted for describing the widespread distribution of information by printed materials or by telegraph. Examples applying it to "one-to-many" radio transmissions of an individual station to multiple listeners appeared as early as 1898. Over the air broadcasting is usually associated with radio and television, though more recently, both radio and television transmissions have begun to be distributed by cable (cable television). The receiving parties may include the general public or a relatively small subset; the point is that anyone with the appropriate receiving technology and equipment (e.g., a radio or television set) can receive the signal. The field of broadcasting includes both government-managed services such as public radio, community radio and public television, and private commercial radio and commercial television. The U.S. Code of Federal Regulations, title 47, part 97 defines broadcasting as "transmissions intended for reception by the general public, either direct or relayed". Private or two-way telecommunications transmissions do not qualify under this definition. For example, amateur ("ham") and citizens band (CB) radio operators are not allowed to broadcast. As defined, transmitting and broadcasting are not the same. Transmission of radio and television programs from a radio or television station to home receivers by radio waves is referred to as over the air (OTA) or terrestrial broadcasting and in most countries requires a broadcasting license. Transmissions using a wire or cable, like cable television (which also retransmits OTA stations with their consent), are also considered broadcasts but do not necessarily require a license (though in some countries, a license is required). In the 2000s, transmissions of television and radio programs via streaming digital technology have increasingly been referred to as broadcasting as well. The earliest broadcasting consisted of sending telegraph signals over the airwaves, using Morse code, a system developed in the 1830s by Samuel Morse, physicist Joseph Henry and Alfred Vail. They developed an electrical telegraph system which sent pulses of electric current along wires which controlled an electromagnet that was located at the receiving end of the telegraph system. A code was needed to transmit natural language using only these pulses, and the silence between them. Morse therefore developed the forerunner to modern International Morse code. This was particularly important for ship-to-ship and ship-to-shore communication, but it became increasingly important for business and general news reporting, and as an arena for personal communication by radio amateurs. In 1894, Italian inventor Guglielmo Marconi began developing a wireless communication using the then-newly discovered phenomenon of radio waves, showing by 1901 that they could be transmitted across the Atlantic Ocean. This was the start of wireless telegraphy by radio. Audio radio broadcasting began experimentally in the first decade of the 20th century. On 17 December 1902, a transmission from the Marconi station in Glace Bay, Nova Scotia, Canada, became the world's first radio message to cross the Atlantic from North America. In 1904, a commercial service was established to transmit nightly news summaries to subscribing ships, which incorporated them into their onboard newspapers. World War I accelerated the development of radio for military communications. After the war, commercial radio AM broadcasting began in the 1920s and became an important mass medium for entertainment and news. World War II again accelerated the development of radio for the wartime purposes of aircraft and land communication, radio navigation, and radar. Development of stereo FM broadcasting of radio began in the 1930s in the United States and the 1970s in the United Kingdom, displacing AM as the dominant commercial standard. On 25 March 1925, John Logie Baird demonstrated the transmission of moving pictures at the London department store Selfridges. Baird's device relied upon the Nipkow disk and thus became known as the mechanical television. It formed the basis of experimental broadcasts done by the British Broadcasting Corporation beginning on 30 September 1929. However, for most of the 20th century, televisions depended on the cathode ray tube invented by Karl Braun. The first version of such a television to show promise was produced by Philo Farnsworth and demonstrated to his family on 7 September 1927. After World War II, interrupted experiments resumed and television became an important home entertainment broadcast medium, using VHF and UHF spectrum. Satellite broadcasting was initiated in the 1960s and moved into general industry usage in the 1970s, with DBS (Direct Broadcast Satellites) emerging in the 1980s. Originally all broadcasting was composed of analog signals using analog transmission techniques but in the 2000s, broadcasters switched to digital signals using digital transmission. An analog signal is any continuous signal representing some other quantity, i.e., analogous to another quantity. For example, in an analog audio signal, the instantaneous signal voltage varies continuously with the pressure of the sound waves. In contrast, a digital signal represents the original time-varying quantity as a sampled sequence of quantized values which imposes some bandwidth and dynamic range constraints on the representation. In general usage, broadcasting most frequently refers to the transmission of information and entertainment programming from various sources to the general public. The world's technological capacity to receive information through one-way broadcast networks more than quadrupled during the two decades from 1986 to 2007, from 432 exabytes of (optimally compressed) information, to 1.9 zettabytes. This is the information equivalent of 55 newspapers per person per day in 1986, and 175 newspapers per person per day by 2007. In a broadcast system, the central high-powered broadcast tower transmits a high-frequency electromagnetic wave to numerous receivers. The high-frequency wave sent by the tower is modulated with a signal containing visual or audio information. The receiver is then tuned so as to pick up the high-frequency wave and a demodulator is used to retrieve the signal containing the visual or audio information. The broadcast signal can be either analog (signal is varied continuously with respect to the information) or digital (information is encoded as a set of discrete values). Historically, there have been several methods used for broadcasting electronic media audio and video to the general public: There are several means of providing financial support for continuous broadcasting: Broadcasters may rely on a combination of these business models. For example, in the United States, National Public Radio (NPR) and the Public Broadcasting Service (PBS, television) supplement public membership subscriptions and grants with funding from the Corporation for Public Broadcasting (CPB), which is allocated bi-annually by Congress. US public broadcasting corporate and charitable grants are generally given in consideration of underwriting spots which differ from commercial advertisements in that they are governed by specific FCC restrictions, which prohibit the advocacy of a product or a "call to action". The first regular television broadcasts started in 1937. Broadcasts can be classified as recorded or live. The former allows correcting errors, and removing superfluous or undesired material, rearranging it, applying slow-motion and repetitions, and other techniques to enhance the program. However, some live events like sports television can include some of the aspects including slow-motion clips of important goals/hits, etc., in between the live television telecast. American radio-network broadcasters habitually forbade prerecorded broadcasts in the 1930s and 1940s, requiring radio programs played for the Eastern and Central time zones to be repeated three hours later for the Pacific time zone (See: Effects of time on North American broadcasting). This restriction was dropped for special occasions, as in the case of the German dirigible airship Hindenburg disaster at Lakehurst, New Jersey, in 1937. During World War II, prerecorded broadcasts from war correspondents were allowed on U.S. radio. In addition, American radio programs were recorded for playback by Armed Forces Radio radio stations around the world. A disadvantage of recording first is that the public may learn the outcome of an event before the recording is broadcast, which may be a spoiler. Prerecording may be used to prevent announcers from deviating from an officially approved script during a live radio broadcast, as occurred with propaganda broadcasts from Germany in the 1940s and with Radio Moscow in the 1980s. Many events are advertised as being live, although they are often recorded live (sometimes called "live-to-tape"). This is particularly true of performances of musical artists on radio when they visit for an in-studio concert performance. Similar situations have occurred in television production ("The Cosby Show is recorded in front of a live television studio audience") and news broadcasting. A broadcast may be distributed through several physical means. If coming directly from the radio studio at a single station or television station, it is sent through the studio/transmitter link to the transmitter and hence from the television antenna located on the radio masts and towers out to the world. Programming may also come through a communications satellite, played either live or recorded for later transmission. Networks of stations may simulcast the same programming at the same time, originally via microwave link, now usually by satellite. Distribution to stations or networks may also be through physical media, such as magnetic tape, compact disc (CD), DVD, and sometimes other formats. Usually these are included in another broadcast, such as when electronic news gathering (ENG) returns a story to the station for inclusion on a news programme. The final leg of broadcast distribution is how the signal gets to the listener or viewer. It may come over the air as with a radio station or television station to an antenna and radio receiver, or may come through cable television or cable radio (or wireless cable) via the station or directly from a network. The Internet may also bring either internet radio or streaming media television to the recipient, especially with multicasting allowing the signal and bandwidth to be shared. The term broadcast network is often used to distinguish networks that broadcast over-the-air television signals that can be received using a tuner inside a television set with a television antenna from so-called networks that are broadcast only via cable television (cablecast) or satellite television that uses a dish antenna. The term broadcast television can refer to the television programs of such networks. The sequencing of content in a broadcast is called a schedule. As with all technological endeavors, a number of technical terms and slang have developed. A list of these terms can be found at List of broadcasting terms. Television and radio programs are distributed through radio broadcasting or cable, often both simultaneously. By coding signals and having a cable converter box with decoding equipment in homes, the latter also enables subscription-based channels, pay-tv and pay-per-view services. In his essay, John Durham Peters wrote that communication is a tool used for dissemination. Peters stated, "Dissemination is a lens—sometimes a usefully distorting one—that helps us tackle basic issues such as interaction, presence, and space and time ... on the agenda of any future communication theory in general". Dissemination focuses on the message being relayed from one main source to one large audience without the exchange of dialogue in between. It is possible for the message to be changed or corrupted by government officials once the main source releases it. There is no way to predetermine how the larger population or audience will absorb the message. They can choose to listen, analyze, or ignore it. Dissemination in communication is widely used in the world of broadcasting. Broadcasting focuses on getting a message out and it is up to the general public to do what they wish with it. Peters also states that broadcasting is used to address an open-ended destination. There are many forms of broadcasting, but they all aim to distribute a signal that will reach the target audience. Broadcasters typically arrange audiences into entire assemblies. In terms of media broadcasting, a radio show can gather a large number of followers who tune in every day to specifically listen to that specific disc jockey. The disc jockey follows the script for his or her radio show and just talks into the microphone. He or she does not expect immediate feedback from any listeners. The message is broadcast across airwaves throughout the community, but there the listeners cannot always respond immediately, especially since many radio shows are recorded prior to the actual air time. Broadcast engineering is the field of electrical engineering, and now to some extent computer engineering and information technology, which deals with radio and television broadcasting. Audio engineering and RF engineering are also essential parts of broadcast engineering, being their own subsets of electrical engineering. Broadcast engineering involves both the studio and transmitter aspects (the entire airchain), as well as remote broadcasts. Every station has a broadcast engineer, though one may now serve an entire station group in a city. In small media markets the engineer may work on a contract basis for one or more stations as needed.
[[File:New Harmony, Indiana, por F. Bates.jpg|thumb|New Harmony, a model community presented by Robert Owen, 1838]] '''Socialism''' is an economic system in which [[Industry|industries]] are owned by [[Worker|workers]] rather than by private [[Business|businesses]]. It is different from [[capitalism]], where private actors, like [[business]] owners and [[Shareholder|shareholders]], own the means of production. The state can also act as a capitalist by owning the means of production and by directing the economy. Socialists believe that capitalists owning the means of production is a form of [[exploitation]] because it lets them own a majority of the labor that workers produce, even though the owners haven't done any work. By allowing capitalists to own a worker's labor, they can extract surplus value and generate [[profit]]. Socialists propose that workers themselves get to own and manage their labor. Socialists believe that sharing ownership of the means of production equally among society would increase people's [[quality of life]]. Socialists want to give people free access to basic life necessities like food, [[housing]], and [[healthcare]]. Some socialists also believe employment should be guaranteed as a [[Human rights|human right]]. Socialists want to prevent problems that they believe come from unchecked capitalism, like poor treatment of workers and [[inequality]]. There are varying views among socialists as to how exactly the issues should be prevented. There are varying views among socialists as to how different issues should be approached. Most socialists would argue that capitalism, in its purest form, can inflict great harm. There are lots of kinds of socialism. One big thing that splits socialists is how to make socialism happen. [[Reformist socialism|Reformists]] think people can get socialism by voting in [[Election|elections]] and changing the government bit by bit. [[Revolution|Revolutionaries]] believe that private owners are too powerful and will use their power to stop any [[Reform|reforms]], so the only way to get socialism is to take their ownership by force. Many groups have criticized socialism for a lot of different reasons. For example, some people think it goes against basic [[human nature]], which they see as naturally [[Greed|greedy]] and [[Competition|competitive]]. Some are concerned about the amount of [[authoritarianism]] needed to implement socialist policies. They point to things like the lack of [[Freedom of speech|free speech]] in past or present socialist nations. And some think socialism will stifle [[innovation]] because people won't be able to profit from ownership of their [[Invention|inventions]]. Socialists such as the anarchist philosopher [[Peter Kropotkin]] responded to this criticism by saying that human nature is defined by co-operation and mutual aid. Also, in his writings, he advocated for anarchist communism, which addresses the concerns of potential socialist authoritarianism. == Forms of socialism == There are many kinds of socialism. In all types, at least in principle, ''the state or workers own the means of production''.<ref>Lamb, Peter & J.C. Docherty. 2006. ''Historical dictionary of socialism''. Lanham, Maryland, UK; Oxford, England, UK: Scarecrow Press. p. 1.</ref> Means of production is originally a communist term which describes the tools that are needed to make goods. In a capitalist economy, these tools are normally owned by one person, and can range from a machine in a [[factory]] to an entire corporation. The one person hires workers for a [[wage]], and keeps any profits made. The profit is not split up amongst the workers. This can create inequality as the owner of the mean of production may be able to become very [[Wealth|wealthy]]. The major differences between the different varieties are the role of the [[free market]] (market planning), how the means of production are controlled, the role of management of workers, and the government's role in the economy. === Collectivization === One kind of socialism is "[[collectivization]]." In this system, money and goods are shared more equally among the people. In theory, this system results in the gap between classes getting smaller. The state helps the nation's poorest people. The richest are subjected to higher [[Tax|taxes]] and [[Economics|economic]] restrictions. === Communism as a goal === Socialists with more radical views believe that socialism will evolve into what they see as a more advanced system: [[communism]], with no [[state]], [[money]], [[hierarchy]] or social classes whatsoever. In [[Marxism|Marxist]] theory, socialism is a temporary social state between [[capitalism]] and [[communism]]. Some socialists have no intention of transitioning to communism. Many label these economic theories into one as "communism" when they mean the [[Marxism|Marxist]] and [[Leninism|Leninist]] ideas and beliefs of [[Russia]]'s [[Bolshevik]] party. [[Karl Marx|Marx]] believed that capitalism followed the economic and political system of [[feudalism]]. He also believed that capitalism would unfairly treat many people. He thought that those people would eventually [[revolt]] and switch to socialism. He thought that socialism could be another bridge on a path to communism. However, many people incorrectly use the term "Communist" to refer to a socialist state as a pejorative insult. Others call this 'State Socialism,' to distinguish it from the communist goal that does not need a state or any form of government. To non-communists, the word 'socialism' is now used mostly for attempts to come close to this goal in a capitalist state. === Democratic socialism === [[Democratic socialism]] is the belief that socialism can be achieved through reform of what Marxists call Bourgeois Democracy<ref>{{Cite web|title=On Authority|url=https://www.marxists.org/archive/marx/works/1872/10/authority.htm|access-date=2023-02-15|website=www.marxists.org}}</ref>―as opposed to achieving socialism by a [[revolution]]. The goals of these reforms are aimed to achieve a more just society, so that ordinary people can participate in the many decisions that affect their lives, which should require voting. === Social democracy === [[Social democracy]] is a type of socialism that attempts to mix parts of socialism with [[capitalism]]. In this system, despite there still being [[private property]], the government generates tax revenue. Tax comes from the wealthiest in the society and [[Corporation|corporations]]. It goes to the poor, or even everyone in the society in the form of social programs. These programs range from [[Single-payer health care|single-payer healthcare]], to other welfare programs such as expanding [[Supplemental Nutrition Assistance Program|SNAP]] benefits. The intentions of social democracy and socialism can be similar or shared. Social democracy keeps the capitalist system intact, and tries to reform it. Achieving socialism would mean completely getting rid of the capitalist system. Social democracy is often confused with democratic socialism due to the similar names and having the same short-term goals. The biggest difference is social democrats want to stop reforming capitalism when they think their reforms are good enough, but democratic socialists will not stop until capitalism is gone. Some examples of social democracies are the [[Scandinavia|Scandinavian]] countries. In social democracies, some services and industries are given [[subsidy]] money to help them run partly controlled by the [[government]]. [[Education]], [[National Health Service|health care]], [[House|housing]], [[Public utility|utility]] companies and [[public transportation]] are some industries that might be owned/supported by the government in a social democracy. For the most part, people working in these industries are paid by the government, with money paid by the people as taxes. A strong [[Welfare]] system is key to social democracy. ===Utopian socialism=== [[Utopian socialism]] is the first form of modern socialism. Unlike other forms of socialism, utopian socialism doesn't include belief in the necessity of [[class struggle]]. ===Anarcho-socialism=== Anarcho-socialism, also called [[Anarchism|anarchist]] socialism, [[libertarian socialism]] or free socialism is the idea of a socialist society where the state does not exist. Anarcho-socialists believe that both the state and the [[Bourgeoisie|bourgeois]] class (the class that owns the means of production in capitalism) oppresses the workers and a stateless mutual aid society should be established through a revolution. Mutual aid is the belief that people in a community (for example city districts or towns and villages) should help each other when necessary. Anarcho-socialism is very similar to anarcho-communism, and many anarcho-socialists are anarcho-communists. Other anarcho-socialists include mutualists (of [[Pierre-Joseph Proudhon]]'s philosophy) === Other === Many people and countries see socialism differently. The [[Socialist International]] is an organization dedicated to the cause of promoting socialist ideals, and has ties with many socialist parties, especially [[:Category:Social democratic parties|Social Democratic]] parties. == History == [[File:Sesel map of socialist states.PNG|thumb|Map of Socialist and Communist Countries during the Cold War in 1985]] The followers of a [[Wales|Welshman]], [[Robert Owen]], began calling themselves socialists in 1841.<ref name=gale>Gale (2001). "Socialism" . World of Sociology. Retrieved 15 June 2011.</ref> Owen is seen as a founder of the [[Co-op]]erative Movement in Britain. He said that workers should own the companies they worked for. The workers would then share the [[profit]]s among themselves. He set up a new model factory in [[New Lanark]], Scotland.<ref name=wt>{{cite encyclopedia |url= http://www.credoreference.com/entry/sharpewt/socialism |title=Socialism |first= |last= |encyclopedia=Encyclopedia of World Trade From Ancient Times to the Present |year=2005 |accessdate=15 June 2011}}</ref> [[Karl Marx]] is the most well-known creator of the theory of socialism, and of communism. He wrote a book about capitalism, socialism, and communism, called "[[Das Kapital|A critique of the social economy]]". [[Friedrich Engels]] co-wrote the book, and paid for much of Marx's work and research. Many socialist [[political party|political parties]] were formed during the 19th century and early part of the 20th century. During the [[decolonization]] movement in the 20th century, many [[armies]] fighting for [[independence]] planned to establish socialist countries. ==Related pages== * [[Capitalism]] * [[Communism]] * [[Social democracy]] ==References== {{reflist}} {{Political spectrum}} {{Political ideologies}} {{Philosophy topics}} {{Social and political philosophy}} [[Category:Forms of government]] [[Category:Socialism| ]]
Socialism is a political philosophy and movement encompassing a range of economic and social systems characterised by social ownership of the means of production, as opposed to private ownership. It describes the economic, political, and social theories and movements associated with the implementation of such systems. Social ownership can take various forms including: public, community, collective, cooperative, or employee. No single definition encapsulates the many types of socialism, but social ownership is the common element. Traditionally, socialism is on the left-wing of the political spectrum. Types of socialism vary based on the role of markets and planning in resource allocation, the structure of management in organizations, and different approaches from below or from above. Some socialists favour a party, state, or technocratic-driven approach, while others disagree on whether government is the correct vehicle for change. Socialist systems divide into non-market and market forms. Non-market socialism substitutes factor markets with integrated economic planning and engineering, or technical criteria based on calculation performed in-kind, thereby producing a different economic mechanism that functions according to different economic laws and dynamics than those of capitalism. A non-market socialist system seeks to eliminate the perceived inefficiencies, irrationalities, unpredictability, and crises that socialists traditionally associate with capital accumulation and the profit system. Market socialism retains the use of monetary prices, factor markets and in some cases the profit motive, with respect to the operation of socially owned enterprises and the allocation of capital goods between them. Profits generated by these firms would be controlled directly by the workforce of each firm or accrue to society at large in the form of a social dividend. Socialist parties and ideas remain a political force with varying degrees of power and influence, heading national governments in several countries. Socialist politics have been internationalist and nationalist; organised through political parties and opposed to party politics; at times overlapping with trade unions and other times independent and critical of them, and present in industrialised and developing nations. Social democracy originated within the socialist movement, supporting economic and social interventions to promote social justice. While retaining socialism as a long-term goal, in the post-war period social democracy embraced a mixed economy based on Keynesianism within a predominantly developed capitalist market economy and liberal democratic polity that expands state intervention to include income redistribution, regulation, and a welfare state. Economic democracy proposes a sort of market socialism, with more democratic control of companies, investments and natural resources. The socialist political movement includes political philosophies that originated in the revolutionary movements of the mid-to-late 18th century and out of concern for the social problems that socialists associated with capitalism. By the late 19th century, after the work of Karl Marx and his collaborator Friedrich Engels, socialism had come to signify anti-capitalism and advocacy for a post-capitalist system based on some form of social ownership of the means of production. By the early 1920s, communism and social democracy had become the two dominant political tendencies within the international socialist movement, with socialism itself becoming the most influential secular movement of the 20th century. Many socialists also adopted the causes of other social movements, such as feminism, environmentalism, and progressivism. While the emergence of the Soviet Union as the world's first nominally socialist state led to socialism's widespread association with the Soviet economic model, several scholars posit that in practice, the model functioned as a form of state capitalism. Several academics, political commentators, and scholars have noted that some Western European countries have been governed by socialist parties or have mixed economies that are sometimes called "democratic socialist". Following the revolutions of 1989, many of these countries have moved away from socialism as a neoliberal consensus replaced the social democratic consensus in the advanced capitalist world, while many former socialist politicians and political parties embraced "Third Way" politics, remaining committed to equality and welfare, while abandoning public ownership and class-based politics. Socialism experienced a resurgence in popularity in the 2010s, most prominently in the form of democratic socialism. For Andrew Vincent, "[t]he word 'socialism' finds its root in the Latin sociare, which means to combine or to share. The related, more technical term in Roman and then medieval law was societas. This latter word could mean companionship and fellowship as well as the more legalistic idea of a consensual contract between freemen". Initial use of socialism was claimed by Pierre Leroux, who alleged he first used the term in the Parisian journal Le Globe in 1832. Leroux was a follower of Henri de Saint-Simon, one of the founders of what would later be labelled utopian socialism. Socialism contrasted with the liberal doctrine of individualism that emphasized the moral worth of the individual while stressing that people act or should act as if they are in isolation from one another. The original utopian socialists condemned this doctrine of individualism for failing to address social concerns during the Industrial Revolution, including poverty, oppression, and vast wealth inequality. They viewed their society as harming community life by basing society on competition. They presented socialism as an alternative to liberal individualism based on the shared ownership of resources. Saint-Simon proposed economic planning, scientific administration and the application of scientific understanding to the organisation of society. By contrast, Robert Owen proposed to organise production and ownership via cooperatives. Socialism is also attributed in France to Marie Roch Louis Reybaud while in Britain it is attributed to Owen, who became one of the fathers of the cooperative movement. The definition and usage of socialism settled by the 1860s, replacing associationist, co-operative, and mutualist that had been used as synonyms while communism fell out of use during this period. An early distinction between communism and socialism was that the latter aimed to only socialise production while the former aimed to socialise both production and consumption (in the form of free access to final goods). By 1888, Marxists employed socialism in place of communism as the latter had come to be considered an old-fashioned synonym for socialism. It was not until after the Bolshevik Revolution that socialism was appropriated by Vladimir Lenin to mean a stage between capitalism and communism. He used it to defend the Bolshevik program from Marxist criticism that Russia's productive forces were not sufficiently developed for communism. The distinction between communism and socialism became salient in 1918 after the Russian Social Democratic Labour Party renamed itself to the All-Russian Communist Party, interpreting communism specifically to mean socialists who supported the politics and theories of Bolshevism, Leninism and later that of Marxism–Leninism, although communist parties continued to describe themselves as socialists dedicated to socialism. According to The Oxford Handbook of Karl Marx, "Marx used many terms to refer to a post-capitalist society—positive humanism, socialism, communism, realm of free individuality, free association of producers, etc. He used these terms completely interchangeably. The notion that 'socialism' and 'communism' are distinct historical stages is alien to his work and only entered the lexicon of Marxism after his death". In Christian Europe, communists were believed to have adopted atheism. In Protestant England, communism was too close to the Roman Catholic communion rite, hence socialist was the preferred term. Engels wrote that in 1848, when The Communist Manifesto was published, socialism was respectable in Europe while communism was not. The Owenites in England and the Fourierists in France were considered respectable socialists while working-class movements that "proclaimed the necessity of total social change" denoted themselves communists. This branch of socialism produced the communist work of Étienne Cabet in France and Wilhelm Weitling in Germany. British moral philosopher John Stuart Mill discussed a form of economic socialism within a liberal context that would later be known as liberal socialism. In later editions of his Principles of Political Economy (1848), Mill posited that "as far as economic theory was concerned, there is nothing in principle in economic theory that precludes an economic order based on socialist policies" and promoted substituting capitalist businesses with worker cooperatives. While democrats looked to the Revolutions of 1848 as a democratic revolution which in the long run ensured liberty, equality, and fraternity, Marxists denounced it as a betrayal of working-class ideals by a bourgeoisie indifferent to the proletariat. Socialist models and ideas espousing common or public ownership have existed since antiquity. The economy of the 3rd century BCE Mauryan Empire of India, an absolute monarchy, has been described by some scholars as "a socialized monarchy" and "a sort of state socialism" due to "nationalisation of industries". Other scholars have suggested that elements of socialist thought were present in the politics of classical Greek philosophers Plato and Aristotle. Mazdak the Younger (died c. 524 or 528 CE), a Persian communal proto-socialist, instituted communal possessions and advocated the public good. Abu Dharr al-Ghifari, a Companion of Muhammad, is credited by multiple authors as a principal antecedent of Islamic socialism. The teachings of Jesus are frequently described as socialist, especially by Christian socialists. However, both socialists and non-socialists have disputed claims of Jesus being a socialist. In the Bible, Acts 4:32–35 records that the first Jewish Christian communities were organized along the principles of common ownership and redistribution of goods, although the pattern soon disappears from church history except within monasticism. Christian socialism was one of the founding threads of the British Labour Party and is claimed to begin with the uprising of Wat Tyler and John Ball in the 14th century CE. After the French Revolution, activists and theorists such as François-Noël Babeuf, Étienne-Gabriel Morelly, Philippe Buonarroti and Auguste Blanqui influenced the early French labour and socialist movements. In Britain, Thomas Paine proposed a detailed plan to tax property owners to pay for the needs of the poor in Agrarian Justice while Charles Hall wrote The Effects of Civilization on the People in European States, denouncing capitalism's effects on the poor of his time. This work influenced the utopian schemes of Thomas Spence. The first self-conscious socialist movements developed in the 1820s and 1830s. Groups such as the Fourierists, Owenites and Saint-Simonians provided a series of analyses and interpretations of society. Especially the Owenites overlapped with other working-class movements such as the Chartists in the United Kingdom. This was also the first time that the term socialism itself applies in a fashion recognisably similar to its modern meaning; the word was first used in 1827 in the London Cooperative Magazine in the UK and later in 1832 in the French periodical Le Globe. An earlier usage of the word socialism appears in the Italian language in 1803, but not with the modern meaning of the term. The Chartists gathered significant numbers around the People's Charter of 1838 which sought democratic reforms focused on the extension of suffrage to all male adults. Leaders in the movement called for a more equitable distribution of income and better living conditions for the working classes. The first trade unions and consumer cooperative societies followed the Chartist movement. Pierre-Joseph Proudhon proposed his philosophy of mutualism in which "everyone had an equal claim, either alone or as part of a small cooperative, to possess and use land and other resources as needed to make a living". Other currents inspired Christian socialism "often in Britain and then usually coming out of left liberal politics and a romantic anti-industrialism" which produced theorists such as Edward Bellamy, Charles Kingsley and Frederick Denison Maurice. The first advocates of socialism favoured social levelling in order to create a meritocratic or technocratic society based on individual talent. Henri de Saint-Simon was fascinated by the potential of science and technology and advocated a socialist society that would eliminate the disorderly aspects of capitalism based on equal opportunities. He sought a society in which each person was ranked according to his or her capacities and rewarded according to his or her work. His key focus was on administrative efficiency and industrialism and a belief that science was essential to progress. This was accompanied by a desire for a rationally organised economy based on planning and geared towards large-scale scientific and material progress. West European social critics, including Louis Blanc, Charles Fourier, Charles Hall, Robert Owen, Pierre-Joseph Proudhon and Saint-Simon were the first modern socialists who criticised the poverty and inequality of the Industrial Revolution. They advocated reform, Owen advocating the transformation of society to small communities without private property. Owen's contribution to modern socialism was his claim that individual actions and characteristics were largely determined by their social environment. On the other hand, Fourier advocated Phalanstères (communities that respected individual desires, including sexual preferences), affinities and creativity and saw that work has to be made enjoyable for people. Owen and Fourier's ideas were practiced in intentional communities around Europe and North America in the mid-19th century. The Paris Commune was a government that ruled Paris from 18 March (formally, from 28 March) to 28 May 1871. The Commune was the result of an uprising in Paris after France was defeated in the Franco-Prussian War. The Commune elections were held on 26 March. They elected a Commune council of 92 members, one member for each 20,000 residents. Because the Commune was able to meet on fewer than 60 days in total, only a few decrees were actually implemented. These included the separation of church and state; the remission of rents owed for the period of the siege (during which payment had been suspended); the abolition of night work in the hundreds of Paris bakeries; the granting of pensions to the unmarried companions and children of National Guards killed on active service; and the free return of all workmen's tools and household items valued up to 20 francs that had been pledged during the siege. In 1864, the First International was founded in London. It united diverse revolutionary currents, including socialists such as the French followers of Proudhon, Blanquists, Philadelphes, English trade unionists and social democrats. In 1865 and 1866, it held a preliminary conference and had its first congress in Geneva, respectively. Due to their wide variety of philosophies, conflict immediately erupted. The first objections to Marx came from the mutualists who opposed state socialism. Shortly after Mikhail Bakunin and his followers joined in 1868, the First International became polarised into camps headed by Marx and Bakunin. The clearest differences between the groups emerged over their proposed strategies for achieving their visions. The First International became the first major international forum for the promulgation of socialist ideas. Bakunin's followers were called collectivists and sought to collectivise ownership of the means of production while retaining payment proportional to the amount and kind of labour of each individual. Like Proudhonists, they asserted the right of each individual to the product of his labour and to be remunerated for his particular contribution to production. By contrast, anarcho-communists sought collective ownership of both the means and the products of labour. As Errico Malatesta put it, "instead of running the risk of making a confusion in trying to distinguish what you and I each do, let us all work and put everything in common. In this way each will give to society all that his strength permits until enough is produced for every one; and each will take all that he needs, limiting his needs only in those things of which there is not yet plenty for every one". Anarcho-communism as a coherent economic-political philosophy was first formulated in the Italian section of the First International by Malatesta, Carlo Cafiero, Emilio Covelli, Andrea Costa and other ex-Mazzinian republicans. Out of respect for Bakunin, they did not make their differences with collectivist anarchism explicit until after his death. Syndicalism emerged in France inspired in part by Proudhon and later by Pelloutier and Georges Sorel. It developed at the end of the 19th century out of the French trade-union movement (syndicat is the French word for trade union). It was a significant force in Italy and Spain in the early 20th century until it was crushed by the fascist regimes in those countries. In the United States, syndicalism appeared in the guise of the Industrial Workers of the World, or "Wobblies", founded in 1905. Syndicalism is an economic system that organises industries into confederations (syndicates) and the economy is managed by negotiation between specialists and worker representatives of each field, comprising multiple non-competitive categorised units. Syndicalism is a form of communism and economic corporatism, but also refers to the political movement and tactics used to bring about this type of system. An influential anarchist movement based on syndicalist ideas is anarcho-syndicalism. The International Workers Association is an international anarcho-syndicalist federation of various labour unions. The Fabian Society is a British socialist organisation established to advance socialism via gradualist and reformist means. The society laid many foundations of the Labour Party and subsequently affected the policies of states emerging from the decolonisation of the British Empire, most notably India and Singapore. Originally, the Fabian Society was committed to the establishment of a socialist economy, alongside a commitment to British imperialism as a progressive and modernising force. Later, the society functioned primarily as a think tank and is one of fifteen socialist societies affiliated with the Labour Party. Similar societies exist in Australia (the Australian Fabian Society), in Canada (the Douglas-Coldwell Foundation and the now disbanded League for Social Reconstruction) and in New Zealand. Guild socialism is a political movement advocating workers' control of industry through the medium of trade-related guilds "in an implied contractual relationship with the public". It originated in the United Kingdom and was at its most influential in the first quarter of the 20th century. Inspired by medieval guilds, theorists such as Samuel George Hobson and G. D. H. Cole advocated the public ownership of industries and their workforces' organisation into guilds, each of which under the democratic control of its trade union. Guild socialists were less inclined than Fabians to invest power in a state. At some point, like the American Knights of Labor, guild socialism wanted to abolish the wage system. As the ideas of Marx and Engels gained acceptance, particularly in central Europe, socialists sought to unite in an international organisation. In 1889 (the centennial of the French Revolution), the Second International was founded, with 384 delegates from twenty countries representing about 300 labour and socialist organisations. Engels was elected honorary president at the third congress in 1893. Anarchists were banned, mainly due to pressure from Marxists. It has been argued that at some point the Second International turned "into a battleground over the issue of libertarian versus authoritarian socialism. Not only did they effectively present themselves as champions of minority rights; they also provoked the German Marxists into demonstrating a dictatorial intolerance which was a factor in preventing the British labour movement from following the Marxist direction indicated by such leaders as H. M. Hyndman". Reformism arose as an alternative to revolution. Eduard Bernstein was a leading social democrat in Germany who proposed the concept of evolutionary socialism. Revolutionary socialists quickly targeted reformism: Rosa Luxemburg condemned Bernstein's Evolutionary Socialism in her 1900 essay Social Reform or Revolution? Revolutionary socialism encompasses multiple social and political movements that may define "revolution" differently. The Social Democratic Party of Germany (SPD) became the largest and most powerful socialist party in Europe, despite working illegally until the anti-socialist laws were dropped in 1890. In the 1893 elections, it gained 1,787,000 votes, a quarter of the total votes cast, according to Engels. In 1895, the year of his death, Engels emphasised The Communist Manifesto's emphasis on winning, as a first step, the "battle of democracy". In South America, the Socialist Party of Argentina was established in the 1890s led by Juan B. Justo and Nicolás Repetto, among others. It was the first mass party in the country and in Latin America. The party affiliated itself with the Second International. For four months in 1904, Australian Labor Party leader Chris Watson was the Prime Minister of the country. Watson thus became the head of the world's first socialist or social democratic parliamentary government. Australian historian Geoffrey Blainey argues that the Labor Party was not socialist at all in the 1890s, and that socialist and collectivist elements only made their way in the party's platform in the early 20th century. In 1909, the first Kibbutz was established in Palestine by Russian Jewish Immigrants. The Kibbutz Movement expanded through the 20th century following a doctrine of Zionist socialism. The British Labour Party first won seats in the House of Commons in 1902. By 1917, the patriotism of World War I changed into political radicalism in Australia, most of Europe and the United States. Other socialist parties from around the world who were beginning to gain importance in their national politics in the early 20th century included the Italian Socialist Party, the French Section of the Workers' International, the Spanish Socialist Workers' Party, the Swedish Social Democratic Party, the Russian Social Democratic Labour Party and the Socialist Party in Argentina, the Socialist Workers' Party in Chile and the Socialist Party of America in the United States. In February 1917, a revolution occurred in Russia. Workers, soldiers and peasants established soviets (councils), the monarchy fell and a provisional government convened pending the election of a constituent assembly. In April of that year, Vladimir Lenin, leader of the Bolshevik faction of socialists in Russia and known for his profound and controversial expansions of Marxism, was allowed to cross Germany to return from exile in Switzerland. Lenin had published essays on his analysis of imperialism, the monopoly and globalisation phase of capitalism, as well as analyses on social conditions. He observed that as capitalism had further developed in Europe and America, the workers remained unable to gain class consciousness so long as they were too busy working to pay their expenses. He therefore proposed that the social revolution would require the leadership of a vanguard party of class-conscious revolutionaries from the educated and politically active part of the population. Upon arriving in Petrograd, Lenin declared that the revolution in Russia had only begun, and that the next step was for the workers' soviets to take full authority. He issued a thesis outlining the Bolshevik programme, including rejection of any legitimacy in the provisional government and advocacy for state power to be administered through the soviets. The Bolsheviks became the most influential force. On 7 November, the capitol of the provisional government was stormed by Bolshevik Red Guards in what later was officially known in the Soviet Union as the Great October Socialist Revolution. The provisional government ended and the Russian Socialist Federative Soviet Republic—the world's first constitutionally socialist state—was established. On 25 January 1918, Lenin declared "Long live the world socialist revolution!" at the Petrograd Soviet and proposed an immediate armistice on all fronts and transferred the land of the landed proprietors, the crown and the monasteries to the peasant committees without compensation. The day after assuming executive power on 25 January, Lenin wrote Draft Regulations on Workers' Control, which granted workers control of businesses with more than five workers and office employees and access to all books, documents and stocks and whose decisions were to be "binding upon the owners of the enterprises". Governing through the elected soviets and in alliance with the peasant-based Left Socialist-Revolutionaries, the Bolshevik government began nationalising banks and industry; and disavowed the national debts of the deposed Romanov royal régime. It sued for peace, withdrawing from World War I and convoked a Constituent Assembly in which the peasant Socialist-Revolutionary Party (SR) won a majority. Lenin's government also instituted a number of progressive measures such as universal education, healthcare and equal rights for women. The Constituent Assembly elected SR leader Victor Chernov President of a Russian republic, but rejected the Bolshevik proposal that it endorse the Soviet decrees on land, peace and workers' control and acknowledge the power of the Soviets of Workers', Soldiers' and Peasants' Deputies. The next day, the Bolsheviks declared that the assembly was elected on outdated party lists and the All-Russian Central Executive Committee of the Soviets dissolved it. In March 1919, world communist parties formed Comintern (also known as the Third International) at a meeting in Moscow. In the interwar period, Soviet Union experienced two major famines. The First famine occurred in 1921–1922 with death estimates varying between 1 and 10 million dead. It was caused by a combination of factors – severe drought and failed harvests, continuous war since 1914, forced collectivisation of farms and requisition of grain and seed from peasants (preventing the sowing of crops) by the Soviet authorities, and an economic blockade of the Soviet Union by the Allies. The experience with the famine led Lenin to replace war communism with the New Economic Policy (NEP) in 1921 to alleviate the extreme shortages. Under the NEP, private ownership was allowed for small and medium-sized enterprises. While large industry remained state controlled. A second major famine occurred in 1930–1933, resulting in millions of deaths. The Soviet economy was the modern world's first centrally planned economy. It adopted state ownership of industry managed through Gosplan (the State Planning Commission), Gosbank (the State Bank) and the Gossnab (State Commission for Materials and Equipment Supply). Economic planning was conducted through serial Five-Year Plans. The emphasis was on development of heavy industry at expense of agriculture. Rapid industrialization served two purposes: to bring largely agrarian societies into the modern age, and to establish a politically loyal working class. Modernization brought about a general increase in the standard of living in the 1950s and 60's. The Bolshevik Russian Revolution of January 1918 launched Communist parties in many countries and a wave of revolutions until the mid-1920s. Few communists doubted that the Russian experience depended on successful, working-class socialist revolutions in developed capitalist countries. In 1919, Lenin and Leon Trotsky organised the world's Communist parties into an international association of workers—the Communist International (Comintern), also called the Third International. The Russian Revolution influenced uprisings in other countries. The German Revolution of 1918–1919 replaced Germany's imperial government with a republic. The revolution lasted from November 1918 until the establishment of the Weimar Republic in August 1919. It included an episode known as the Bavarian Soviet Republic and the Spartacist uprising. A short lived Hungarian Soviet Republic was set up in Hungary March 21 to August 1, 1919. It was led by Béla Kun. It instituted a Red Terror. After the regime was put down, an even more brutal White Terror followed. Kun managed to escape to the Soviet Union, where he co-led murder of tens of thousands of White Russians. He was killed in the 1930 Soviet purges. In Italy, the events known as the Biennio Rosso were characterised by mass strikes, worker demonstrations and self-management experiments through land and factory occupations. In Turin and Milan, workers' councils were formed and many factory occupations took place led by anarcho-syndicalists organised around the Unione Sindacale Italiana. There was a short-lived Persian Socialist Soviet Republic in 1920–21. Patagonia Rebelde was a syndicalist-led revolution in Argentina lasting for a year and a half from in 1920–21. The anarchist-led Guangzhou City Commune in China lasted six years from 1921. In 1924, the Mongolian People's Republic was established and was ruled by the Mongolian People's Party. The Shinmin Prefecture in Manchuria lasted two years from 1929. Many of these revolutions initiated societies and economic models that have been described as socialist. In 1922, the fourth congress of the Communist International took up the policy of the united front. It urged communists to work with rank-and-file social democrats while remaining critical of their leaders. They criticised those leaders for betraying the working class by supporting the capitalists' war efforts. The social democrats pointed to the dislocation caused by revolution and later the growing authoritarianism of the communist parties. The Labour Party rejected the Communist Party of Great Britain's application to affiliate to them in 1920. On seeing the Soviet State's growing coercive power in 1923, a dying Lenin said Russia had reverted to "a bourgeois tsarist machine ... barely varnished with socialism". After Lenin's death in January 1924, the Communist Party of the Soviet Union—then increasingly under the control of Joseph Stalin—rejected the theory that socialism could not be built solely in the Soviet Union in favour of the concept of socialism in one country. Stalin developed a bureaucratic and totalitarian government, which was condemned by democratic socialists and anarchists for undermining the Revolution's ideals. The Russian Revolution and its aftermath motivated national Communist parties elsewhere that gained political and social influence, in France, the United States, Italy, China, Mexico, the Brazil, Chile and Indonesia. Left-wing groups which did not agree to the centralisation and abandonment of the soviets by the Bolshevik Party (see anti-Stalinist left) led left-wing uprisings against the Bolsheviks. Such groups included Socialist Revolutionaries, Left Socialist Revolutionaries, Mensheviks and anarchists. Within this left-wing discontent, the most large-scale events were the Kronstadt rebellion and the Makhnovist movement. The International Socialist Commission (ISC, also known as Berne International) was formed in February 1919 at a meeting in Bern by parties that wanted to resurrect the Second International. Centrist socialist parties which did not want to be a part of the resurrected Second International (ISC) or Comintern formed the International Working Union of Socialist Parties (IWUSP, also known as Vienna International, Vienna Union, or Two-and-a-Half International) on 27 February 1921 at a conference in Vienna. The ISC and the IWUSP joined to form the Labour and Socialist International (LSI) in May 1923 at a meeting in Hamburg. The 1920s and 1930s were marked by an increasing divergence between democratic and reformists socialists (mainly affiliated with the Labour and Socialist International) and revolutionary socialists (mainly affiliated with the Communist International), but also by tension within the Communist movement between the dominant Stalinists and dissidents such as Trotsky's followers in the Left Opposition. Trotsky's Fourth International was established in France in 1938 when Trotskyists argued that the Comintern or Third International had become irretrievably "lost to Stalinism" and thus incapable of leading the working class to power. In the Spanish Civil War (1936–1939), socialists (including the democratic socialist Spanish Socialist Workers' Party and the Marxist Workers' Party of Marxist Unification) participated on the Republican side, loyal to the left-leaning Popular Front government of the Second Spanish Republic, in alliance with anarchists of the communist and syndicalist variety and supported by the socialist Workers' General Union. The Spanish Revolution of 1936 was a workers' social revolution during the war, that is often seen as a model of socialism from below. An anarchist-inspired movement of peasants and workers, supported by armed militias, took control of Barcelona and of large areas of rural Spain where they collectivised the land. The Spanish Revolution was a workers' social revolution that began with the Spanish Civil War in 1936 and resulted in the widespread implementation of anarchist and more broadly libertarian socialist organisational principles in some areas for two to three years, primarily Catalonia, Aragon, Andalusia and parts of Levante. Much of Spain's economy came under worker control. In anarchist strongholds like Catalonia the figure was as high as 75%, but lower in areas with heavy Communist Party influence, which actively resisted attempts at collectivisation. Factories were run through worker committees, agrarian areas became collectivised and run as libertarian communes. Anarchist historian Sam Dolgoff estimated that about eight million people participated directly or indirectly in the Spanish Revolution. The rise of Nazism and the start of World War II led to the dissolution of the LSI in 1940. After the War, the Socialist International was formed in Frankfurt in July 1951 as its successor. After World War II, social democratic governments introduced social reform and wealth redistribution via welfare and taxation. Social democratic parties dominated post-war politics in countries such as France, Italy, Czechoslovakia, Belgium and Norway. At one point, France claimed to be the world's most state-controlled capitalist country. It nationalised public utilities including Charbonnages de France (CDF), Électricité de France (EDF), Gaz de France (GDF), Air France, Banque de France and Régie Nationale des Usines Renault. In 1945, the British Labour Party led by Clement Attlee was elected based on a radical socialist programme. The Labour government nationalised industries including mines, gas, coal, electricity, rail, iron, steel and the Bank of England. British Petroleum was officially nationalised in 1951. Anthony Crosland said that in 1956 25% of British industry was nationalised and that public employees, including those in nationalised industries, constituted a similar proportion of the country's workers. The Labour Governments of 1964–1970 and 1974–1979 intervened further. It re-nationalised British Steel (1967) after the Conservatives had denationalised it and nationalised British Leyland (1976). The National Health Service provided taxpayer-funded health care to everyone, free at the point of service. Working-class housing was provided in council housing estates and university education became available via a school grant system. During most of the post-war era, Sweden was governed by the Swedish Social Democratic Party largely in cooperation with trade unions and industry. The party held power from 1936 to 1976, 1982 to 1991, 1994 to 2006 and 2014 to 2022, most often in minority governments. Party leader Tage Erlander led the government from 1946 to 1969, the longest uninterrupted parliamentary government. These governments substantially expanded the welfare state. Swedish Prime Minister Olof Palme identified as a "democratic socialist" and was described as a "revolutionary reformist". The Norwegian Labour Party was established in 1887 and was largely a trade union federation. The party did not proclaim a socialist agenda, elevating universal suffrage and dissolution of the union with Sweden as its top priorities. In 1899, the Norwegian Confederation of Trade Unions separated from the Labour Party. Around the time of the Russian Revolution, the Labour Party moved to the left and joined the Communist International from 1919 through 1923. Thereafter, the party still regarded itself as revolutionary, but the party's left-wing broke away and established the Communist Party of Norway while the Labour Party gradually adopted a reformist line around 1930. In 1935, Johan Nygaardsvold established a coalition that lasted until 1945. From 1946 to 1962, the Norwegian Labour Party held an absolute majority in the parliament led by Einar Gerhardsen, who remained Prime Minister for seventeen years. Although the party abandoned most of its pre-war socialist ideas, the welfare state was expanded under Gerhardsen to ensure the universal provision of basic human rights and stabilise the economy. In the 1945 Norwegian parliamentary election, the Communist Party took 12% of the votes, but it largely vanished during the Cold War. In the 1950s, popular socialism emerged in Nordic countries. It placed itself between communism and social democracy. In the early 1960s, the Socialist Left Party challenged the Labour Party from the left. Also in the 1960s, Gerhardsen established a planning agency and tried to establish a planned economy. In the 1970s, a more radical socialist party, the Worker's Communist Party (AKP), broke from the Socialist Left Party and had notable influence in student associations and some trade unions. The AKP identified with Communist China and Albania rather than the Soviet Union. In countries such as Sweden, the Rehn–Meidner model allowed capitalists owning productive and efficient firms to retain profits at the expense of the firms' workers, exacerbating inequality and causing workers to agitate for a share of the profits in the 1970s. At that time, women working in the state sector began to demand better wages. Rudolf Meidner established a study committee that came up with a 1976 proposal to transfer excess profits into worker-controlled investment funds, with the intention that firms would create jobs and pay higher wages rather than reward company owners and managers. Capitalists immediately labeled this proposal as socialism and launched an unprecedented opposition—including calling off the class compromise established in the 1938 Saltsjöbaden Agreement. Social democratic parties are some of the oldest such parties and operate in all Nordic countries. Countries or political systems that have long been dominated by social democratic parties are often labelled social democratic. Those countries fit the social democratic type of "high socialism" which is described as favouring "a high level of decommodification and a low degree of stratification". The Nordic model is a form of economic-political system common to the Nordic countries (Denmark, Finland, Iceland, Norway and Sweden). It has three main ingredients, namely peaceful, institutionalised negotiation between employers and trade unions; active, predictable and measured macroeconomic policy; and universal welfare and free education. The welfare system is governmental in Norway and Sweden whereas trade unions play a greater role in Denmark, Finland and Iceland. The Nordic universal welfare model is often labelled social democratic and contrasted with the selective continental model and the residual Anglo-American model. Major reforms in the Nordic countries are the results of consensus and compromise across the political spectrum. Key reforms were implemented under social democratic cabinets in Denmark, Norway and Sweden while centre-right parties dominated during the implementation of the model in Finland and Iceland. Since World War II, Nordic countries have largely maintained a social democratic mixed economy, characterised by labour force participation, gender equality, egalitarian and universal benefits, redistribution of wealth and expansionary fiscal policy. In 2015, then-Prime Minister of Denmark Lars Løkke Rasmussen denied that Denmark is socialist, saying "I know that some people in the US associate the Nordic model with some sort of socialism. Therefore I would like to make one thing clear. Denmark is far from a socialist planned economy. Denmark is a market economy". In Norway, the first mandatory social insurances were introduced by conservative cabinets in 1895 (Francis Hagerups's cabinet) and 1911 (Konow's Cabinet). During the 1930s, the Labour Party adopted the conservatives' welfare state project. After World War II, all political parties agreed that the welfare state should be expanded. Universal social security (Folketrygden) was introduced by the conservative Borten's Cabinet. Norway's economy is open to the international or European market for most products and services, joining the European Union's internal market in 1994 through European Economic Area. Some of the mixed economy institutions from the post-war period were relaxed by the conservative cabinet of the 1980s and the finance market was deregulated. Within the Varieties of Capitalism-framework, Finland, Norway and Sweden are identified as coordinated market economies. The Soviet era saw competition between the Soviet-led Eastern Bloc and the United States-led Western Bloc. The Soviet system was seen as a rival of and a threat to Western capitalism for most of the 20th century. The Eastern Bloc was the group of Communist states of Central and Eastern Europe, including the Soviet Union and the countries of the Warsaw Pact, including Poland, the German Democratic Republic, the Hungary, Bulgaria, Czechoslovakia, Romania, Albania, and initially Yugoslavia. In the Informbiro period from 1948, Yugoslavia under Josip Broz Tito pursued a different, more decentralised form of state socialism than the rest of the Eastern Bloc, known as Socialist self-management. The Hungarian Revolution of 1956, a spontaneous nationwide revolt against the Communist government brutally suppressed by Soviet forces, and USSR leader Nikita Khrushchev's denunciation of the excesses of Stalin's regime during the Twentieth Communist Party Congress the same year produced disunity within Western European Communist parties, leading to the emergence of the New Left (see below). Over a decade later, Czechoslovakia under Alexander Dubček also attempted to pursue a more democratic model of state socialism, under the name "Socialism with a human face", during the Prague Spring; this was also brutally suppressed by the Soviet Union. In the post-war years, socialism became increasingly influential in many then-developing countries. Embracing Third World socialism, countries in Africa, Asia and Latin America often nationalised industries. During India's freedom movement and fight for independence, many figures in the left-wing faction of the Indian National Congress organised themselves as the Congress Socialist Party. Their politics and those of the early and intermediate periods of Jayaprakash Narayan's career combined a commitment to the socialist transformation of society with a principled opposition to the one-party authoritarianism they perceived in the Stalinist model. The Chinese Communist Revolution was the second stage in the Chinese Civil War, which ended with the establishment of the People's Republic of China led by the Chinese Communist Party. The then-Chinese Kuomintang Party in the 1920s incorporated Chinese socialism as part of its ideology. Between 1958 and 1962 during the Great Leap Forward in the People's Republic of China, some 30 million people starved to death and at least 45 million died overall. The emergence of this new political entity in the frame of the Cold War was complex and painful. Several tentative efforts were made to organise newly independent states to establish a common front to limit the United States' and the Soviet Union's influence on them. This led to the Sino-Soviet split. The Non-Aligned Movement gathered around the figures of Jawaharlal Nehru of India, Sukarno of Indonesia, Josip Broz Tito of Yugoslavia and Gamal Abdel Nasser of Egypt. After the 1954 Geneva Conference which ended the French war in Vietnam, the 1955 Bandung Conference gathered Nasser, Nehru, Tito, Sukarno and Chinese Premier Zhou Enlai. As many African countries gained independence during the 1960s, some of them rejected capitalism in favour of African socialism as defined by Julius Nyerere of Tanzania, Léopold Senghor of Senegal, Kwame Nkrumah of Ghana and Sékou Touré of Guinea. The Cuban Revolution (1953–1959) was an armed revolt conducted by Fidel Castro's 26th of July Movement and its allies against the government of Fulgencio Batista. Castro's government eventually adopted socialism and the communist ideology, becoming the Communist Party of Cuba in October 1965. In Indonesia in the mid-1960s, a coup attempt blamed on the Communist Party of Indonesia (PKI) was countered by an anti-communist purge led by Suharto, which mainly targeted the growing influence of the PKI and other leftist groups, with significant support from the United States, which culminated in the overthrow of Sukarno. These events resulted not only in the total destruction of the PKI but also the political left in Indonesia, and paved the way for a major shift in the balance of power in Southeast Asia towards the West, a significant turning point in the global Cold War. The New Left was a term used mainly in the United Kingdom and United States in reference to activists, educators and others in the 1960s and 1970s who sought to implement a broad range of reforms on issues such as gay rights, abortion, gender roles and drugs in contrast to earlier leftist or Marxist movements that had taken a more vanguardist approach to social justice and focused mostly on labour unionisation and questions of social class. The New Left rejected involvement with the labour movement and Marxism's historical theory of class struggle. In the United States, the New Left was associated with the Hippie movement and anti-war college campus protest movements as well as the black liberation movements such as the Black Panther Party. While initially formed in opposition to the "Old Left" Democratic Party, groups composing the New Left gradually became central players in the Democratic coalition. The protests of 1968 represented a worldwide escalation of social conflicts, predominantly characterised by popular rebellions against military, capitalist and bureaucratic elites who responded with an escalation of political repression. These protests marked a turning point for the civil rights movement in the United States which produced revolutionary movements like the Black Panther Party. The prominent civil rights leader Martin Luther King Jr. organised the "Poor People's Campaign" to address issues of economic justice, while personally showing sympathy with democratic socialism. In reaction to the Tet Offensive, protests also sparked a broad movement in opposition to the Vietnam War all over the United States and even into London, Paris, Berlin and Rome. In 1968, the International of Anarchist Federations was founded during a conference held in Carrara by the three existing European federations of France, the Italian and the Iberian Anarchist Federation as well as the Bulgarian federation in French exile. Mass socialist movements grew not only in the United States, but also in most European countries. In many other capitalist countries, struggles against dictatorships, state repression and colonisation were also marked by protests in 1968, such as the Tlatelolco massacre in Mexico City and the escalation of guerrilla warfare against the military dictatorship in Brazil. Countries governed by Communist parties saw protests against bureaucratic and military elites too. In Eastern Europe, widespread protests escalated particularly in the Prague Spring in Czechoslovakia. In response, Soviet Union occupied Czechoslovakia. The occupation was denounced by the Italian and French Communist parties and the Communist Party of Finland, but defended by the Portuguese Communist Party secretary-general Álvaro Cunhal the Communist Party of Luxembourg and conservative factions of the Communist Party of Greece. In the Chinese Cultural Revolution, a socio-political youth movement mobilised against "bourgeois" elements which were seen to be infiltrating the government and society at large, aiming to restore capitalism. This movement motivated Maoism-inspired movements around the world in the context of the Sino-Soviet split. In the 1960s, a socialist tendency within the Latin American Catholic church appeared and was known as liberation theology It motivated the Colombian priest Camilo Torres Restrepo to enter the ELN guerrilla. In Chile, Salvador Allende, a physician and candidate for the Socialist Party of Chile, was elected president in 1970. In 1973, his government was ousted by the United States-backed military dictatorship of Augusto Pinochet, which lasted until the late 1980s. In Jamaica, the democratic socialist Michael Manley served as the fourth Prime Minister of Jamaica from 1972 to 1980 and from 1989 to 1992. According to opinion polls, he remains one of Jamaica's most popular Prime Ministers since independence. The Nicaraguan Revolution encompassed the rising opposition to the Somoza dictatorship in the 1960s and 1970s, the campaign led by the Sandinista National Liberation Front (FSLN) to violently oust the dictatorship in 1978–1979, the subsequent efforts of the FSLN to govern Nicaragua from 1979 until 1990 and the socialist measures which included wide-scale agrarian reform and educational programs. The People's Revolutionary Government was proclaimed on 13 March 1979 in Grenada which was overthrown by armed forces of the United States in 1983. The Salvadoran Civil War (1979–1992) was a conflict between the military-led government of El Salvador and the Farabundo Martí National Liberation Front (FMLN), a coalition or umbrella organisation of five socialist guerrilla groups. A coup on 15 October 1979 led to the killings of anti-coup protesters by the government as well as anti-disorder protesters by the guerrillas and is widely seen as the tipping point towards the civil war. In 1976, the word socialist was added to the Preamble of the Indian Constitution by the 42nd amendment Act during the Emergency, implying both social and economic equality. In 1982, the newly elected French socialist government of François Mitterrand nationalised parts of a few key industries, including banks and insurance companies. Eurocommunism was a trend in the 1970s and 1980s in various Western European Communist parties to develop a theory and practice of social transformation that was more relevant for a Western European country and less aligned to the influence or control of the Communist Party of the Soviet Union. Outside Western Europe, it is sometimes called neocommunism. Some Communist parties with strong popular support, notably the Italian Communist Party (PCI) and the Communist Party of Spain (PCE). adopted Eurocommunism most enthusiastically and the Communist Party of Finland was dominated by Eurocommunists. The French Communist Party (PCF) and many smaller parties strongly opposed Eurocommunism and stayed aligned with the Communist Party of the Soviet Union until the end of the Soviet Union. Also emerging from the Communist movement but moving in a more left-wing direction, in Italy Autonomia Operaia was particularly active from 1976 to 1978; it took an important role in the autonomist movement in the 1970s, alongside earlier organisations such as Potere Operaio (created after May 1968) and Lotta Continua, promoting a radical form of socialism based on working class self-activity rather than vanguard parties and state planning. Until its 1976 Geneva Congress, the Socialist International (SI) had few members outside Europe and no formal involvement with Latin America. In the late 1970s and in the 1980s, the SI had extensive contacts and discussion with the two powers of the Cold War, the United States and the Soviet Union, about east–west relations and arms control, and admitted as member parties the Nicaraguan FSLN, the left-wing Puerto Rican Independence Party, as well as former Communist parties such as the Democratic Party of the Left of Italy and the Front for the Liberation of Mozambique (FRELIMO). The SI aided social democratic parties in re-establishing themselves when dictatorship gave way to democracy in Portugal (1974) and Spain (1975). After Mao Zedong's death in 1976 and the arrest of the faction known as the Gang of Four, who were blamed for the excesses of the Cultural Revolution, Deng Xiaoping took power and led the People's Republic of China to significant economic reforms. The Chinese Communist Party (CCP) loosened governmental control over citizens' personal lives and the communes were disbanded in favour of private land leases, thus China's transition from a planned economy to a mixed economy named as "socialism with Chinese characteristics" which maintained state ownership rights over land, state or cooperative ownership of much of the heavy industrial and manufacturing sectors and state influence in the banking and financial sectors. China adopted its current constitution on 4 December 1982. Chinese Communist Party General Secretary Jiang Zemin, Premiers Li Peng and Zhu Rongji led the nation in the 1990s. Under their administration, China sustained an average annual gross domestic product growth rate of 11.2%. At the Sixth National Congress of the Communist Party of Vietnam in December 1986, reformist politicians replaced the "old guard" government with new leadership. The reformers were led by 71-year-old Nguyen Van Linh, who became the party's new general secretary. Linh and the reformers implemented a series of free market reforms—known as Đổi Mới ("Renovation")—which carefully managed the transition from a planned economy to a "socialist-oriented market economy". The Soviet Union experienced continued increases in mortality rate (particularly among men) as far back as 1965. Mikhail Gorbachev wished to move the Soviet Union towards of Nordic-style social democracy, calling it "a socialist beacon for all mankind". Prior to its dissolution in 1991, the economy of the Soviet Union was by some measures the second largest in the world after the United States. However, the economy was also beset by economic stagnation, an inflationary spiral, shortages of consumer goods, and fiscal mismanagement. With the collapse of the Soviet Union, the economic integration of the Soviet republics was dissolved and overall industrial activity and economic productivity declined substantially. A lasting legacy of Communism in Soviet Union remains in the physical infrastructure created during decades of combined industrial production practices, and widespread environmental destruction. The transition to capitalist market economies in the former Soviet Union and Eastern Bloc was accompanied by Washington Consensus-inspired "shock therapy", advocated by Western institutions and economists with the intent to replace state socialism with capitalism and integrate these countries into the capitalist western world. Following a transition to free-market capitalism, there was initially a steep fall in the standard of living. Post-Communist Russia experienced rising economic inequality and poverty a surge in excess mortality amongst men, and a decline in life expectancy, which was accompanied by the entrenchment of a newly established business oligarchy. By contrast, the Central European states of the former Eastern Bloc–Poland, Hungary, the Czech Republic and Slovakia–showed healthy increases in life expectancy from the 1990s onward, compared to nearly thirty years of stagnation under socialism. Bulgaria and Romania followed this trend after the introduction of more serious economic reforms in the late 1990s. The right-libertarian think tank Cato Institute has stated that the analyses done of post-communist countries in the 1990s were "premature" and "that early and rapid reformers by far outperformed gradual reformers" on GDP per capita, the United Nations Human Development Index, political freedom, and developed better institutions. The institute also stated that the process of privatization in Russia was "deeply flawed" due to Russia's reforms being "far less rapid" than those of Central Europe and the Baltic states. The average post-Communist country had returned to 1989 levels of per-capita GDP by 2005, and as of 2015, some countries were still behind that. Several scholars state that the negative economic developments in post-Communist countries after the fall of Communism led to increased nationalist sentiment and nostalgia for the Communist era. In 2011, The Guardian published an analysis of the former Soviet countries twenty years after the fall of the USSR. They found that "GDP fell as much as 50 percent in the 1990s in some republics... as capital flight, industrial collapse, hyperinflation and tax avoidance took their toll," but that there was a rebound in the 2000s, and by 2010 "some economies were five times as big as they were in 1991." Life expectancy has grown since 1991 in some of the countries, but fallen in others; likewise, some held free and fair elections, while others remained authoritarian. By 2019, the majority of people in most Eastern European countries approved of the shift to multiparty democracy and a market economy, with approval being highest among residents of Poland and residents in the territory of what was once East Germany, and disapproval being the highest among residents of Russia and Ukraine. In addition, 61 per cent said that standards of living were now higher than they had been under Communism, while only 31 per cent said that they were worse, with the remaining 8 per cent saying that they did not know or that standards of living had not changed. Many social democratic parties, particularly after the Cold War, adopted neoliberal market policies including privatisation, deregulation and financialisation. They abandoned their pursuit of moderate socialism in favour of economic liberalism. By the 1980s, with the rise of conservative neoliberal politicians such as Ronald Reagan in the United States, Margaret Thatcher in Britain, Brian Mulroney in Canada and Augusto Pinochet in Chile, the Western welfare state was dismantled from within, but state support for the corporate sector was maintained. In the United Kingdom, Labour Party leader Neil Kinnock expelled some Trotskyist members and refused to support the 1984–1985 miners' strike over pit closures. In 1989, the 18th Congress of the SI adopted a new Declaration of Principles, stating: "Democratic socialism is an international movement for freedom, social justice, and solidarity. Its goal is to achieve a peaceful world where these basic values can be enhanced and where each individual can live a meaningful life with the full development of his or her personality and talents, and with the guarantee of human and civil rights in a democratic framework of society." In the 1990s, the British Labour Party under Tony Blair enacted policies based on the free-market economy to deliver public services via the private finance initiative. Influential in these policies was the idea of a Third Way between Old Left state socialism and New Right market capitalism, and a re-evaluation of welfare state policies. In 1995, the Labour Party re-defined its stance on socialism by re-wording Clause IV of its constitution, defining socialism in ethical terms and removing all references to public, direct worker or municipal ownership of the means of production. The Labour Party stated: "The Labour Party is a democratic socialist party. It believes that, by the strength of our common endeavour we achieve more than we achieve alone, so as to create, for each of us, the means to realise our true potential, and, for all of us, a community in which power, wealth, and opportunity are in the hands of the many, not the few." Left-wing critics of the Third Way argued that it reduced equality to an equal opportunity to compete in an economy in which the rich were growing richer and the poor were becoming more disadvantaged, which the leftists argue is not socialist. Starting in the late 20th century, the development of a post-industrial economy in which information and knowledge matter more than material production and labor raised doubts about the continued relevance of socialism, since socialism emerged in response to industrialization under capitalism. Several scholars argued that socialism was dead in the immediate aftermath of the end of the Cold War. German sociologist and liberal politician Ralf Dahrendorf declared that "socialism is dead, and none of its variants can be revived for a world awakening from the double nightmare of Stalinism and Brezhnevism." Andre Gorz, a left-wing philosopher, also declared that "As a system, socialism is dead. As a movement and an organized political force, it is on its last legs. All the goals it once proclaimed are out of date." American economist Robert Heilbroner wrote that "Less than seventy-five years after it officially began, the contest between capitalism and socialism is over: capitalism has won." However there was also a counter-argument put forward by the socialist political scholars Antonio Negri and Felix Guattari, who argued that "whether perestroika succeeds in the present form or in a second wave that will inevitably follow, whether the Russian empire endures or not – these are all problems that concern only the Soviets," arguing that the role of socialism in global politics was not tied to the fate of the Soviet Union. In 1990, the São Paulo Forum was launched by the Workers' Party (Brazil), linking left-wing socialist parties in Latin America. Its members were associated with the Pink tide of left-wing governments on the continent in the early 21st century. Member parties ruling countries included the Front for Victory in Argentina, the PAIS Alliance in Ecuador, Farabundo Martí National Liberation Front in El Salvador, Peru Wins in Peru, and the United Socialist Party of Venezuela, whose leader Hugo Chávez initiated what he called "Socialism of the 21st century". Many mainstream democratic socialist and social democratic parties continued to drift right-wards. On the right of the socialist movement, the Progressive Alliance was founded in 2013 by current or former members of the Socialist International. The organisation states the aim of becoming the global network of "the progressive, democratic, social-democratic, socialist and labour movement". Mainstream social democratic and socialist parties are also networked in Europe in the Party of European Socialists formed in 1992. Many of these parties lost large parts of their electoral base in the early 21st century. This phenomenon is known as Pasokification from the Greek party PASOK, which saw a declining share of the vote in national elections—from 43.9% in 2009 to 13.2% in May 2012, to 12.3% in June 2012 and 4.7% in 2015—due to its poor handling of the Greek government-debt crisis and implementation of harsh austerity measures. In Europe, the share of votes for such socialist parties was at its 70-year lowest in 2015. For example, the Socialist Party, after winning the 2012 French presidential election, rapidly lost its vote share, the Social Democratic Party of Germany's fortunes declined rapidly from 2005 to 2019, and outside Europe the Israeli Labor Party fell from being the dominant force in Israeli politics to 4.43% of the vote in the April 2019 Israeli legislative election, and the Peruvian Aprista Party went from ruling party in 2011 to a minor party. The decline of these mainstream parties opened space for more radical and populist left parties in some countries, such as Spain's Podemos, Greece's Syriza (in government, 2015–19), Germany's Die Linke, and France's La France Insoumise. In other countries, left-wing revivals have taken place within mainstream democratic socialist and centrist parties, as with Jeremy Corbyn in the United Kingdom and Bernie Sanders in the United States. Few of these radical left parties have won national government in Europe, while some more mainstream socialist parties have managed to, such as Portugal's Socialist Party. Bhaskar Sunkara, the founding editor of the American socialist magazine Jacobin, argued that the appeal of socialism persists due to the inequality and "tremendous suffering" under current global capitalism, the use of wage labor "which rests on the exploitation and domination of humans by other humans," and ecological crises, such as climate change. In contrast, Mark J. Perry of the conservative American Enterprise Institute (AEI) argued that despite socialism's resurgence, it is still "a flawed system based on completely faulty principles that aren't consistent with human behavior and can't nurture the human spirit.", adding that "While it promised prosperity, equality, and security, it delivered poverty, misery, and tyranny." Some in the scientific community have suggested that a contemporary radical response to social and ecological problems could be seen in the emergence of movements associated with degrowth, eco-socialism and eco-anarchism. Early socialist thought took influences from a diverse range of philosophies such as civic republicanism, Enlightenment rationalism, romanticism, forms of materialism, Christianity (both Catholic and Protestant), natural law and natural rights theory, utilitarianism and liberal political economy. Another philosophical basis for a great deal of early socialism was the emergence of positivism during the European Enlightenment. Positivism held that both the natural and social worlds could be understood through scientific knowledge and be analysed using scientific methods. The fundamental objective of socialism is to attain an advanced level of material production and therefore greater productivity, efficiency and rationality as compared to capitalism and all previous systems, under the view that an expansion of human productive capability is the basis for the extension of freedom and equality in society. Many forms of socialist theory hold that human behaviour is largely shaped by the social environment. In particular, socialism holds that social mores, values, cultural traits and economic practices are social creations and not the result of an immutable natural law. The object of their critique is thus not human avarice or human consciousness, but the material conditions and man-made social systems (i.e. the economic structure of society) which give rise to observed social problems and inefficiencies. Bertrand Russell, often considered to be the father of analytic philosophy, identified as a socialist. Russell opposed the class struggle aspects of Marxism, viewing socialism solely as an adjustment of economic relations to accommodate modern machine production to benefit all of humanity through the progressive reduction of necessary work time. Socialists view creativity as an essential aspect of human nature and define freedom as a state of being where individuals are able to express their creativity unhindered by constraints of both material scarcity and coercive social institutions. The socialist concept of individuality is intertwined with the concept of individual creative expression. Karl Marx believed that expansion of the productive forces and technology was the basis for the expansion of human freedom and that socialism, being a system that is consistent with modern developments in technology, would enable the flourishing of "free individualities" through the progressive reduction of necessary labour time. The reduction of necessary labour time to a minimum would grant individuals the opportunity to pursue the development of their true individuality and creativity. Socialists argue that the accumulation of capital generates waste through externalities that require costly corrective regulatory measures. They also point out that this process generates wasteful industries and practices that exist only to generate sufficient demand for products such as high-pressure advertisement to be sold at a profit, thereby creating rather than satisfying economic demand. Socialists argue that capitalism consists of irrational activity, such as the purchasing of commodities only to sell at a later time when their price appreciates, rather than for consumption, even if the commodity cannot be sold at a profit to individuals in need and therefore a crucial criticism often made by socialists is that "making money", or accumulation of capital, does not correspond to the satisfaction of demand (the production of use-values). The fundamental criterion for economic activity in capitalism is the accumulation of capital for reinvestment in production, but this spurs the development of new, non-productive industries that do not produce use-value and only exist to keep the accumulation process afloat (otherwise the system goes into crisis), such as the spread of the financial industry, contributing to the formation of economic bubbles. Such accumulation and reinvestment, when it demands a constant rate of profit, causes problems if the earnings in the rest of society do not increase in proportion. Socialists view private property relations as limiting the potential of productive forces in the economy. According to socialists, private property becomes obsolete when it concentrates into centralised, socialised institutions based on private appropriation of revenue—but based on cooperative work and internal planning in allocation of inputs—until the role of the capitalist becomes redundant. With no need for capital accumulation and a class of owners, private property in the means of production is perceived as being an outdated form of economic organisation that should be replaced by a free association of individuals based on public or common ownership of these socialised assets. Private ownership imposes constraints on planning, leading to uncoordinated economic decisions that result in business fluctuations, unemployment and a tremendous waste of material resources during crisis of overproduction. Excessive disparities in income distribution lead to social instability and require costly corrective measures in the form of redistributive taxation, which incurs heavy administrative costs while weakening the incentive to work, inviting dishonesty and increasing the likelihood of tax evasion while (the corrective measures) reduce the overall efficiency of the market economy. These corrective policies limit the incentive system of the market by providing things such as minimum wages, unemployment insurance, taxing profits and reducing the reserve army of labour, resulting in reduced incentives for capitalists to invest in more production. In essence, social welfare policies cripple capitalism and its incentive system and are thus unsustainable in the long run. Marxists argue that the establishment of a socialist mode of production is the only way to overcome these deficiencies. Socialists and specifically Marxian socialists argue that the inherent conflict of interests between the working class and capital prevent optimal use of available human resources and leads to contradictory interest groups (labour and business) striving to influence the state to intervene in the economy in their favour at the expense of overall economic efficiency. Early socialists (utopian socialists and Ricardian socialists) criticised capitalism for concentrating power and wealth within a small segment of society. In addition, they complained that capitalism does not use available technology and resources to their maximum potential in the interests of the public. At a certain stage of development, the material productive forces of society come into conflict with the existing relations of production or—this merely expresses the same thing in legal terms—with the property relations within the framework of which they have operated hitherto. Then begins an era of social revolution. The changes in the economic foundation lead sooner or later to the transformation of the whole immense superstructure. —Karl Marx, Critique of the Gotha Program Karl Marx and Friedrich Engels argued that socialism would emerge from historical necessity as capitalism rendered itself obsolete and unsustainable from increasing internal contradictions emerging from the development of the productive forces and technology. It was these advances in the productive forces combined with the old social relations of production of capitalism that would generate contradictions, leading to working-class consciousness. Marx and Engels held the view that the consciousness of those who earn a wage or salary (the working class in the broadest Marxist sense) would be moulded by their conditions of wage slavery, leading to a tendency to seek their freedom or emancipation by overthrowing ownership of the means of production by capitalists and consequently, overthrowing the state that upheld this economic order. For Marx and Engels, conditions determine consciousness and ending the role of the capitalist class leads eventually to a classless society in which the state would wither away. Marx and Engels used the terms socialism and communism interchangeably, but many later Marxists defined socialism as a specific historical phase that would displace capitalism and precede communism. The major characteristics of socialism (particularly as conceived by Marx and Engels after the Paris Commune of 1871) are that the proletariat would control the means of production through a workers' state erected by the workers in their interests. For orthodox Marxists, socialism is the lower stage of communism based on the principle of "from each according to his ability, to each according to his contribution" while upper stage communism is based on the principle of "from each according to his ability, to each according to his need", the upper stage becoming possible only after the socialist stage further develops economic efficiency and the automation of production has led to a superabundance of goods and services. Marx argued that the material productive forces (in industry and commerce) brought into existence by capitalism predicated a cooperative society since production had become a mass social, collective activity of the working class to create commodities but with private ownership (the relations of production or property relations). This conflict between collective effort in large factories and private ownership would bring about a conscious desire in the working class to establish collective ownership commensurate with the collective efforts their daily experience. Socialists have taken different perspectives on the state and the role it should play in revolutionary struggles, in constructing socialism and within an established socialist economy. In the 19th century, the philosophy of state socialism was first explicitly expounded by the German political philosopher Ferdinand Lassalle. In contrast to Karl Marx's perspective of the state, Lassalle rejected the concept of the state as a class-based power structure whose main function was to preserve existing class structures. Lassalle also rejected the Marxist view that the state was destined to "wither away". Lassalle considered the state to be an entity independent of class allegiances and an instrument of justice that would therefore be essential for achieving socialism. Preceding the Bolshevik-led revolution in Russia, many socialists including reformists, orthodox Marxist currents such as council communism, anarchists and libertarian socialists criticised the idea of using the state to conduct central planning and own the means of production as a way to establish socialism. Following the victory of Leninism in Russia, the idea of "state socialism" spread rapidly throughout the socialist movement and eventually state socialism came to be identified with the Soviet economic model. Joseph Schumpeter rejected the association of socialism and social ownership with state ownership over the means of production because the state as it exists in its current form is a product of capitalist society and cannot be transplanted to a different institutional framework. Schumpeter argued that there would be different institutions within socialism than those that exist within modern capitalism, just as feudalism had its own distinct and unique institutional forms. The state, along with concepts like property and taxation, were concepts exclusive to commercial society (capitalism) and attempting to place them within the context of a future socialist society would amount to a distortion of these concepts by using them out of context. Utopian socialism is a term used to define the first currents of modern socialist thought as exemplified by the work of Henri de Saint-Simon, Charles Fourier and Robert Owen which inspired Karl Marx and other early socialists. Visions of imaginary ideal societies, which competed with revolutionary social democratic movements, were viewed as not being grounded in the material conditions of society and as reactionary. Although it is technically possible for any set of ideas or any person living at any time in history to be a utopian socialist, the term is most often applied to those socialists who lived in the first quarter of the 19th century who were ascribed the label "utopian" by later socialists as a negative term to imply naivete and dismiss their ideas as fanciful or unrealistic. Religious sects whose members live communally such as the Hutterites are not usually called "utopian socialists", although their way of living is a prime example. They have been categorised as religious socialists by some. Similarly, modern intentional communities based on socialist ideas could also be categorised as "utopian socialist". For Marxists, the development of capitalism in Western Europe provided a material basis for the possibility of bringing about socialism because according to The Communist Manifesto "[w]hat the bourgeoisie produces above all is its own grave diggers", namely the working class, which must become conscious of the historical objectives set it by society. Revolutionary socialists believe that a social revolution is necessary to effect structural changes to the socioeconomic structure of society. Among revolutionary socialists there are differences in strategy, theory and the definition of revolution. Orthodox Marxists and left communists take an impossibilist stance, believing that revolution should be spontaneous as a result of contradictions in society due to technological changes in the productive forces. Lenin theorised that under capitalism the workers cannot achieve class consciousness beyond organising into trade unions and making demands of the capitalists. Therefore, Leninists argue that it is historically necessary for a vanguard of class conscious revolutionaries to take a central role in coordinating the social revolution to overthrow the capitalist state and eventually the institution of the state altogether. Revolution is not necessarily defined by revolutionary socialists as violent insurrection, but as a complete dismantling and rapid transformation of all areas of class society led by the majority of the masses: the working class. Reformism is generally associated with social democracy and gradualist democratic socialism. Reformism is the belief that socialists should stand in parliamentary elections within capitalist society and if elected use the machinery of government to pass political and social reforms for the purposes of ameliorating the instabilities and inequities of capitalism. Within socialism, reformism is used in two different ways. One has no intention of bringing about socialism or fundamental economic change to society and is used to oppose such structural changes. The other is based on the assumption that while reforms are not socialist in themselves, they can help rally supporters to the cause of revolution by popularizing the cause of socialism to the working class. The debate on the ability for social democratic reformism to lead to a socialist transformation of society is over a century old. Reformism is criticized for being paradoxical as it seeks to overcome the existing economic system of capitalism while trying to improve the conditions of capitalism, thereby making it appear more tolerable to society. According to Rosa Luxemburg, capitalism is not overthrown, "but is on the contrary strengthened by the development of social reforms". In a similar vein, Stan Parker of the Socialist Party of Great Britain argues that reforms are a diversion of energy for socialists and are limited because they must adhere to the logic of capitalism. French social theorist Andre Gorz criticized reformism by advocating a third alternative to reformism and social revolution that he called "non-reformist reforms", specifically focused on structural changes to capitalism as opposed to reforms to improve living conditions within capitalism or to prop it up through economic interventions. The economic anarchy of capitalist society as it exists today is, in my opinion, the real source of the evil. ... I am convinced there is only one way to eliminate these grave evils, namely through the establishment of a socialist economy, accompanied by an educational system which would be oriented toward social goals. In such an economy, the means of production are owned by society itself and are utilised in a planned fashion. A planned economy, which adjusts production to the needs of the community, would distribute the work to be done among all those able to work and would guarantee a livelihood to every man, woman, and child. The education of the individual, in addition to promoting his own innate abilities, would attempt to develop in him a sense of responsibility for his fellow men in place of the glorification of power and success in our present society. —Albert Einstein, "Why Socialism?", 1949 Socialist economics starts from the premise that "individuals do not live or work in isolation but live in cooperation with one another. Furthermore, everything that people produce is in some sense a social product, and everyone who contributes to the production of a good is entitled to a share in it. Society as whole, therefore, should own or at least control property for the benefit of all its members". The original conception of socialism was an economic system whereby production was organised in a way to directly produce goods and services for their utility (or use-value in classical and Marxian economics), with the direct allocation of resources in terms of physical units as opposed to financial calculation and the economic laws of capitalism (see law of value), often entailing the end of capitalistic economic categories such as rent, interest, profit and money. In a fully developed socialist economy, production and balancing factor inputs with outputs becomes a technical process to be undertaken by engineers. Market socialism refers to an array of different economic theories and systems that use the market mechanism to organise production and to allocate factor inputs among socially owned enterprises, with the economic surplus (profits) accruing to society in a social dividend as opposed to private capital owners. Variations of market socialism include libertarian proposals such as mutualism, based on classical economics, and neoclassical economic models such as the Lange Model. Some economists, such as Joseph Stiglitz, Mancur Olson, and others not specifically advancing anti-socialists positions have shown that prevailing economic models upon which such democratic or market socialism models might be based have logical flaws or unworkable presuppositions. These criticisms have been incorporated into the models of market socialism developed by John Roemer and Nicholas Vrousalis. The ownership of the means of production can be based on direct ownership by the users of the productive property through worker cooperative; or commonly owned by all of society with management and control delegated to those who operate/use the means of production; or public ownership by a state apparatus. Public ownership may refer to the creation of state-owned enterprises, nationalisation, municipalisation or autonomous collective institutions. Some socialists feel that in a socialist economy, at least the "commanding heights" of the economy must be publicly owned. Economic liberals and right libertarians view private ownership of the means of production and the market exchange as natural entities or moral rights which are central to their conceptions of freedom and liberty and view the economic dynamics of capitalism as immutable and absolute, therefore they perceive public ownership of the means of production, cooperatives and economic planning as infringements upon liberty. Management and control over the activities of enterprises are based on self-management and self-governance, with equal power-relations in the workplace to maximise occupational autonomy. A socialist form of organisation would eliminate controlling hierarchies so that only a hierarchy based on technical knowledge in the workplace remains. Every member would have decision-making power in the firm and would be able to participate in establishing its overall policy objectives. The policies/goals would be carried out by the technical specialists that form the coordinating hierarchy of the firm, who would establish plans or directives for the work community to accomplish these goals: The role and use of money in a hypothetical socialist economy is a contested issue. Nineteenth century socialists including Karl Marx, Robert Owen, Pierre-Joseph Proudhon and John Stuart Mill advocated various forms of labour vouchers or labour credits, which like money would be used to acquire articles of consumption, but unlike money they are unable to become capital and would not be used to allocate resources within the production process. Bolshevik revolutionary Leon Trotsky argued that money could not be arbitrarily abolished following a socialist revolution. Money had to exhaust its "historic mission", meaning it would have to be used until its function became redundant, eventually being transformed into bookkeeping receipts for statisticians and only in the more distant future would money not be required for even that role. A planned economy is a type of economy consisting of a mixture of public ownership of the means of production and the coordination of production and distribution through economic planning. A planned economy can be either decentralised or centralised. Enrico Barone provided a comprehensive theoretical framework for a planned socialist economy. In his model, assuming perfect computation techniques, simultaneous equations relating inputs and outputs to ratios of equivalence would provide appropriate valuations to balance supply and demand. The most prominent example of a planned economy was the economic system of the Soviet Union and as such the centralised-planned economic model is usually associated with the communist states of the 20th century, where it was combined with a single-party political system. In a centrally planned economy, decisions regarding the quantity of goods and services to be produced are planned in advance by a planning agency (see also the analysis of Soviet-type economic planning). The economic systems of the Soviet Union and the Eastern Bloc are further classified as "command economies", which are defined as systems where economic coordination is undertaken by commands, directives and production targets. Studies by economists of various political persuasions on the actual functioning of the Soviet economy indicate that it was not actually a planned economy. Instead of conscious planning, the Soviet economy was based on a process whereby the plan was modified by localised agents and the original plans went largely unfulfilled. Planning agencies, ministries and enterprises all adapted and bargained with each other during the formulation of the plan as opposed to following a plan passed down from a higher authority, leading some economists to suggest that planning did not actually take place within the Soviet economy and that a better description would be an "administered" or "managed" economy. Although central planning was largely supported by Marxist–Leninists, some factions within the Soviet Union before the rise of Stalinism held positions contrary to central planning. Leon Trotsky rejected central planning in favour of decentralised planning. He argued that central planners, regardless of their intellectual capacity, would be unable to coordinate effectively all economic activity within an economy because they operated without the input and tacit knowledge embodied by the participation of the millions of people in the economy. As a result, central planners would be unable to respond to local economic conditions. State socialism is unfeasible in this view because information cannot be aggregated by a central body and effectively used to formulate a plan for an entire economy, because doing so would result in distorted or absent price signals. Socialism, you see, is a bird with two wings. The definition is 'social ownership and democratic control of the instruments and means of production.' —Upton Sinclair A self-managed, decentralised economy is based on autonomous self-regulating economic units and a decentralised mechanism of resource allocation and decision-making. This model has found support in notable classical and neoclassical economists including Alfred Marshall, John Stuart Mill and Jaroslav Vanek. There are numerous variations of self-management, including labour-managed firms and worker-managed firms. The goals of self-management are to eliminate exploitation and reduce alienation. Guild socialism is a political movement advocating workers' control of industry through the medium of trade-related guilds "in an implied contractual relationship with the public". It originated in the United Kingdom and was at its most influential in the first quarter of the 20th century. It was strongly associated with G. D. H. Cole and influenced by the ideas of William Morris. One such system is the cooperative economy, a largely free market economy in which workers manage the firms and democratically determine remuneration levels and labour divisions. Productive resources would be legally owned by the cooperative and rented to the workers, who would enjoy usufruct rights. Another form of decentralised planning is the use of cybernetics, or the use of computers to manage the allocation of economic inputs. The socialist-run government of Salvador Allende in Chile experimented with Project Cybersyn, a real-time information bridge between the government, state enterprises and consumers. Another, more recent variant is participatory economics, wherein the economy is planned by decentralised councils of workers and consumers. Workers would be remunerated solely according to effort and sacrifice, so that those engaged in dangerous, uncomfortable and strenuous work would receive the highest incomes and could thereby work less. A contemporary model for a self-managed, non-market socialism is Pat Devine's model of negotiated coordination. Negotiated coordination is based upon social ownership by those affected by the use of the assets involved, with decisions made by those at the most localised level of production. Michel Bauwens identifies the emergence of the open software movement and peer-to-peer production as a new alternative mode of production to the capitalist economy and centrally planned economy that is based on collaborative self-management, common ownership of resources and the production of use-values through the free cooperation of producers who have access to distributed capital. Anarcho-communism is a theory of anarchism which advocates the abolition of the state, private property and capitalism in favour of common ownership of the means of production. Anarcho-syndicalism was practised in Catalonia and other places in the Spanish Revolution during the Spanish Civil War. Sam Dolgoff estimated that about eight million people participated directly or at least indirectly in the Spanish Revolution. The economy of the former Socialist Federal Republic of Yugoslavia established a system based on market-based allocation, social ownership of the means of production and self-management within firms. This system substituted Yugoslavia's Soviet-type central planning with a decentralised, self-managed system after reforms in 1953. The Marxian economist Richard D. Wolff argues that "re-organising production so that workers become collectively self-directed at their work-sites" not only moves society beyond both capitalism and state socialism of the last century, but would also mark another milestone in human history, similar to earlier transitions out of slavery and feudalism. As an example, Wolff claims that Mondragon is "a stunningly successful alternative to the capitalist organisation of production". State socialism can be used to classify any variety of socialist philosophies that advocates the ownership of the means of production by the state apparatus, either as a transitional stage between capitalism and socialism, or as an end-goal in itself. Typically, it refers to a form of technocratic management, whereby technical specialists administer or manage economic enterprises on behalf of society and the public interest instead of workers' councils or workplace democracy. A state-directed economy may refer to a type of mixed economy consisting of public ownership over large industries, as promoted by various Social democratic political parties during the 20th century. This ideology influenced the policies of the British Labour Party during Clement Attlee's administration. In the biography of the 1945 United Kingdom Labour Party Prime Minister Clement Attlee, Francis Beckett states: "[T]he government ... wanted what would become known as a mixed economy." Nationalisation in the United Kingdom was achieved through compulsory purchase of the industry (i.e. with compensation). British Aerospace was a combination of major aircraft companies British Aircraft Corporation, Hawker Siddeley and others. British Shipbuilders was a combination of the major shipbuilding companies including Cammell Laird, Govan Shipbuilders, Swan Hunter and Yarrow Shipbuilders, whereas the nationalisation of the coal mines in 1947 created a coal board charged with running the coal industry commercially so as to be able to meet the interest payable on the bonds which the former mine owners' shares had been converted into. Market socialism consists of publicly owned or cooperatively owned enterprises operating in a market economy. It is a system that uses the market and monetary prices for the allocation and accounting of the means of production, thereby retaining the process of capital accumulation. The profit generated would be used to directly remunerate employees, collectively sustain the enterprise or finance public institutions. In state-oriented forms of market socialism, in which state enterprises attempt to maximise profit, the profits can be used to fund government programs and services through a social dividend, eliminating or greatly diminishing the need for various forms of taxation that exist in capitalist systems. Neoclassical economist Léon Walras believed that a socialist economy based on state ownership of land and natural resources would provide a means of public finance to make income taxes unnecessary. Yugoslavia implemented a market socialist economy based on cooperatives and worker self-management. Some of the economic reforms introduced during the Prague Spring by Alexander Dubček, the leader of Czechoslovakia, included elements of market socialism. Mutualism is an economic theory and anarchist school of thought that advocates a society where each person might possess a means of production, either individually or collectively, with trade representing equivalent amounts of labour in the free market. Integral to the scheme was the establishment of a mutual-credit bank that would lend to producers at a minimal interest rate, just high enough to cover administration. Mutualism is based on a labour theory of value that holds that when labour or its product is sold, in exchange it ought to receive goods or services embodying "the amount of labour necessary to produce an article of exactly similar and equal utility". The current economic system in China is formally referred to as a socialist market economy with Chinese characteristics. It combines a large state sector that comprises the commanding heights of the economy, which are guaranteed their public ownership status by law, with a private sector mainly engaged in commodity production and light industry responsible from anywhere between 33% to over 70% of GDP generated in 2005. Although there has been a rapid expansion of private-sector activity since the 1980s, privatisation of state assets was virtually halted and were partially reversed in 2005. The current Chinese economy consists of 150 corporatised state-owned enterprises that report directly to China's central government. By 2008, these state-owned corporations had become increasingly dynamic and generated large increases in revenue for the state, resulting in a state-sector led recovery during the 2009 financial crises while accounting for most of China's economic growth. The Chinese economic model is widely cited as a contemporary form of state capitalism, the major difference between Western capitalism and the Chinese model being the degree of state-ownership of shares in publicly listed corporations. The Socialist Republic of Vietnam has adopted a similar model after the Doi Moi economic renovation but slightly differs from the Chinese model in that the Vietnamese government retains firm control over the state sector and strategic industries, but allows for private-sector activity in commodity production. While major socialist political movements include anarchism, communism, the labour movement, Marxism, social democracy, and syndicalism, independent socialist theorists, utopian socialist authors, and academic supporters of socialism may not be represented in these movements. Some political groups have called themselves socialist while holding views that some consider antithetical to socialism. Socialist has been used by members of the political right as an epithet, including against individuals who do not consider themselves to be socialists and against policies that are not considered socialist by their proponents. While there are many variations of socialism, and there is no single definition encapsulating all of socialism, there have been common elements identified by scholars. In his Dictionary of Socialism (1924), Angelo S. Rappoport analysed forty definitions of socialism to conclude that common elements of socialism include general criticism of the social effects of private ownership and control of capital—as being the cause of poverty, low wages, unemployment, economic and social inequality and a lack of economic security; a general view that the solution to these problems is a form of collective control over the means of production, distribution and exchange (the degree and means of control vary among socialist movements); an agreement that the outcome of this collective control should be a society based upon social justice, including social equality, economic protection of people and should provide a more satisfying life for most people. In The Concepts of Socialism (1975), Bhikhu Parekh identifies four core principles of socialism and particularly socialist society, namely sociality, social responsibility, cooperation and planning. In his study Ideologies and Political Theory (1996), Michael Freeden states that all socialists share five themes: the first is that socialism posits that society is more than a mere collection of individuals; second, that it considers human welfare a desirable objective; third, that it considers humans by nature to be active and productive; fourth, it holds the belief of human equality; and fifth, that history is progressive and will create positive change on the condition that humans work to achieve such change. Anarchism advocates stateless societies often defined as self-governed voluntary institutions, but that several authors have defined as more specific institutions based on non-hierarchical free associations. While anarchism holds the state to be undesirable, unnecessary or harmful, it is not the central aspect. Anarchism entails opposing authority or hierarchical organisation in the conduct of human relations, including the state system. Mutualists support market socialism, collectivist anarchists favour workers cooperatives and salaries based on the amount of time contributed to production, anarcho-communists advocate a direct transition from capitalism to libertarian communism and a gift economy and anarcho-syndicalists prefer workers' direct action and the general strike. The authoritarian–libertarian struggles and disputes within the socialist movement go back to the First International and the expulsion in 1872 of the anarchists, who went on to lead the Anti-authoritarian International and then founded their own libertarian international, the Anarchist St. Imier International. In 1888, the individualist anarchist Benjamin Tucker, who proclaimed himself to be an anarchistic socialist and libertarian socialist in opposition to the authoritarian state socialism and the compulsory communism, included the full text of a "Socialistic Letter" by Ernest Lesigne in his essay on "State Socialism and Anarchism". According to Lesigne, there are two types of socialism: "One is dictatorial, the other libertarian". Tucker's two socialisms were the authoritarian state socialism which he associated to the Marxist school and the libertarian anarchist socialism, or simply anarchism, that he advocated. Tucker noted that the fact that the authoritarian "State Socialism has overshadowed other forms of Socialism gives it no right to a monopoly of the Socialistic idea". According to Tucker, what those two schools of socialism had in common was the labor theory of value and the ends, by which anarchism pursued different means. According to anarchists such as the authors of An Anarchist FAQ, anarchism is one of the many traditions of socialism. For anarchists and other anti-authoritarian socialists, socialism "can only mean a classless and anti-authoritarian (i.e. libertarian) society in which people manage their own affairs, either as individuals or as part of a group (depending on the situation). In other words, it implies self-management in all aspects of life", including at the workplace. Michael Newman includes anarchism as one of many socialist traditions. Peter Marshall argues that "[i]n general anarchism is closer to socialism than liberalism. ... Anarchism finds itself largely in the socialist camp, but it also has outriders in liberalism. It cannot be reduced to socialism, and is best seen as a separate and distinctive doctrine." You can't talk about ending the slums without first saying profit must be taken out of slums. You're really tampering and getting on dangerous ground because you are messing with folk then. You are messing with captains of industry. Now this means that we are treading in difficult water, because it really means that we are saying that something is wrong with capitalism. There must be a better distribution of wealth, and maybe America must move toward a democratic socialism. —Martin Luther King Jr., 1966 Democratic socialism represents any socialist movement that seeks to establish an economy based on economic democracy by and for the working class. Democratic socialism is difficult to define and groups of scholars have radically different definitions for the term. Some definitions simply refer to all forms of socialism that follow an electoral, reformist or evolutionary path to socialism rather than a revolutionary one. According to Christopher Pierson, "[i]f the contrast which 1989 highlights is not that between socialism in the East and liberal democracy in the West, the latter must be recognised to have been shaped, reformed and compromised by a century of social democratic pressure". Pierson further claims that "social democratic and socialist parties within the constitutional arena in the West have almost always been involved in a politics of compromise with existing capitalist institutions (to whatever far distant prize its eyes might from time to time have been lifted)". For Pierson, "if advocates of the death of socialism accept that social democrats belong within the socialist camp, as I think they must, then the contrast between socialism (in all its variants) and liberal democracy must collapse. For actually existing liberal democracy is, in substantial part, a product of socialist (social democratic) forces". Social democracy is a socialist tradition of political thought. Many social democrats refer to themselves as socialists or democratic socialists and some such as Tony Blair employ these terms interchangeably. Others found "clear differences" between the three terms and prefer to describe their own political beliefs by using the term social democracy. The two main directions were to establish democratic socialism or to build first a welfare state within the capitalist system. The first variant advances democratic socialism through reformist and gradualist methods. In the second variant, social democracy is a policy regime involving a welfare state, collective bargaining schemes, support for publicly financed public services and a mixed economy. It is often used in this manner to refer to Western and Northern Europe during the later half of the 20th century. It was described by Jerry Mander as "hybrid economics", an active collaboration of capitalist and socialist visions. Some studies and surveys indicate that people tend to live happier and healthier lives in social democratic societies rather than neoliberal ones. Social democrats advocate a peaceful, evolutionary transition of the economy to socialism through progressive social reform. It asserts that the only acceptable constitutional form of government is representative democracy under the rule of law. It promotes extending democratic decision-making beyond political democracy to include economic democracy to guarantee employees and other economic stakeholders sufficient rights of co-determination. It supports a mixed economy that opposes inequality, poverty and oppression while rejecting both a totally unregulated market economy or a fully planned economy. Common social democratic policies include universal social rights and universally accessible public services such as education, health care, workers' compensation and other services, including child care and elder care. Social democracy supports the trade union labour movement and supports collective bargaining rights for workers. Most social democratic parties are affiliated with the Socialist International. Modern democratic socialism is a broad political movement that seeks to promote the ideals of socialism within the context of a democratic system. Some democratic socialists support social democracy as a temporary measure to reform the current system while others reject reformism in favour of more revolutionary methods. Modern social democracy emphasises a program of gradual legislative modification of capitalism to make it more equitable and humane while the theoretical end goal of building a socialist society is relegated to the indefinite future. According to Sheri Berman, Marxism is loosely held to be valuable for its emphasis on changing the world for a more just, better future. The two movements are widely similar both in terminology and in ideology, although there are a few key differences. The major difference between social democracy and democratic socialism is the object of their politics in that contemporary social democrats support a welfare state and unemployment insurance as well as other practical, progressive reforms of capitalism and are more concerned to administrate and humanise it. On the other hand, democratic socialists seek to replace capitalism with a socialist economic system, arguing that any attempt to humanise capitalism through regulations and welfare policies would distort the market and create economic contradictions. Ethical socialism appeals to socialism on ethical and moral grounds as opposed to economic, egoistic, and consumeristic grounds. It emphasizes the need for a morally conscious economy based upon the principles of altruism, cooperation, and social justice while opposing possessive individualism. Ethical socialism has been the official philosophy of mainstream socialist parties. Liberal socialism incorporates liberal principles to socialism. It has been compared to post-war social democracy for its support of a mixed economy that includes both public and private capital goods. While democratic socialism and social democracy are anti-capitalist positions insofar as criticism of capitalism is linked to the private ownership of the means of production, liberal socialism identifies artificial and legalistic monopolies to be the fault of capitalism and opposes an entirely unregulated market economy. It considers both liberty and social equality to be compatible and mutually dependent. Principles that can be described as ethical or liberal socialist have been based upon or developed by philosophers such as John Stuart Mill, Eduard Bernstein, John Dewey, Carlo Rosselli, Norberto Bobbio, and Chantal Mouffe. Other important liberal socialist figures include Guido Calogero, Piero Gobetti, Leonard Trelawny Hobhouse, John Maynard Keynes and R. H. Tawney. Liberal socialism has been particularly prominent in British and Italian politics. Blanquism is a conception of revolution named for Louis Auguste Blanqui. It holds that socialist revolution should be carried out by a relatively small group of highly organised and secretive conspirators. Upon seizing power, the revolutionaries introduce socialism. Rosa Luxemburg and Eduard Bernstein criticised Lenin, stating that his conception of revolution was elitist and Blanquist. Marxism–Leninism combines Marx's scientific socialist concepts and Lenin's anti-imperialism, democratic centralism and vanguardism. Hal Draper defined socialism from above as the philosophy which employs an elite administration to run the socialist state. The other side of socialism is a more democratic socialism from below. The idea of socialism from above is much more frequently discussed in elite circles than socialism from below—even if that is the Marxist ideal—because it is more practical. Draper viewed socialism from below as being the purer, more Marxist version of socialism. According to Draper, Karl Marx and Friedrich Engels were devoutly opposed to any socialist institution that was "conducive to superstitious authoritarianism". Draper makes the argument that this division echoes the division between "reformist or revolutionary, peaceful or violent, democratic or authoritarian, etc." and further identifies six major varieties of socialism from above, among them "Philanthropism", "Elitism", "Pannism", "Communism", "Permeationism" and "Socialism-from-Outside". According to Arthur Lipow, Marx and Engels were "the founders of modern revolutionary democratic socialism", described as a form of "socialism from below" that is "based on a mass working-class movement, fighting from below for the extension of democracy and human freedom". This type of socialism is contrasted to that of the "authoritarian, anti-democratic creed" and "the various totalitarian collectivist ideologies which claim the title of socialism" as well as "the many varieties of 'socialism from above' which have led in the twentieth century to movements and state forms in which a despotic 'new class' rules over a statified economy in the name of socialism", a division that "runs through the history of the socialist movement". Lipow identifies Bellamyism and Stalinism as two prominent authoritarian socialist currents within the history of the socialist movement. Libertarian socialism, sometimes called left-libertarianism, social anarchism and socialist libertarianism, is an anti-authoritarian, anti-statist and libertarian tradition within socialism that rejects centralised state ownership and control including criticism of wage labour relationships (wage slavery) as well as the state itself. It emphasises workers' self-management and decentralised structures of political organisation. Libertarian socialism asserts that a society based on freedom and equality can be achieved through abolishing authoritarian institutions that control production. Libertarian socialists generally prefer direct democracy and federal or confederal associations such as libertarian municipalism, citizens' assemblies, trade unions and workers' councils. Anarcho-syndicalist Gaston Leval explained: We therefore foresee a Society in which all activities will be coordinated, a structure that has, at the same time, sufficient flexibility to permit the greatest possible autonomy for social life, or for the life of each enterprise, and enough cohesiveness to prevent all disorder. ... In a well-organised society, all of these things must be systematically accomplished by means of parallel federations, vertically united at the highest levels, constituting one vast organism in which all economic functions will be performed in solidarity with all others and that will permanently preserve the necessary cohesion". All of this is typically done within a general call for libertarian and voluntary free associations through the identification, criticism and practical dismantling of illegitimate authority in all aspects of human life. As part of the larger socialist movement, it seeks to distinguish itself from Bolshevism, Leninism and Marxism–Leninism as well as social democracy. Past and present political philosophies and movements commonly described as libertarian socialist include anarchism (anarcho-communism, anarcho-syndicalism, collectivist anarchism, individualist anarchism and mutualism), autonomism, Communalism, participism, libertarian Marxism (council communism and Luxemburgism), revolutionary syndicalism and utopian socialism (Fourierism). Christian socialism is a broad concept involving an intertwining of Christian religion with socialism. Islamic socialism is a more spiritual form of socialism. Muslim socialists believe that the teachings of the Quran and Muhammad are not only compatible with, but actively promoting the principles of equality and public ownership, drawing inspiration from the early Medina welfare state he established. Muslim socialists are more conservative than their Western contemporaries and find their roots in anti-imperialism, anti-colonialism and sometimes, if in an Arab speaking country, Arab nationalism. Islamic socialists believe in deriving legitimacy from political mandate as opposed to religious texts. Socialist feminism is a branch of feminism that argues that liberation can only be achieved by working to end both economic and cultural sources of women's oppression. Marxist feminism's foundation was laid by Engels in The Origin of the Family, Private Property, and the State (1884). August Bebel's Woman under Socialism (1879), is the "single work dealing with sexuality most widely read by rank-and-file members of the Social Democratic Party of Germany (SPD)". In the late 19th and early 20th centuries, both Clara Zetkin and Eleanor Marx were against the demonisation of men and supported a proletariat revolution that would overcome as many male-female inequalities as possible. As their movement already had the most radical demands in women's equality, most Marxist leaders, including Clara Zetkin and Alexandra Kollontai, counterposed Marxism against liberal feminism rather than trying to combine them. Anarcha-feminism began with late 19th- and early 20th-century authors and theorists such as anarchist feminists Goldman and Voltairine de Cleyre In the Spanish Civil War, an anarcha-feminist group, Mujeres Libres ("Free Women") linked to the Federación Anarquista Ibérica, organised to defend both anarchist and feminist ideas. In 1972, the Chicago Women's Liberation Union published "Socialist Feminism: A Strategy for the Women's Movement", which is believed to be the first published use of the term "socialist feminism". Many socialists were early advocates of LGBT rights. For early socialist Charles Fourier, true freedom could only occur without suppressing passions, as the suppression of passions is not only destructive to the individual, but to society as a whole. Writing before the advent of the term "homosexuality", Fourier recognised that both men and women have a wide range of sexual needs and preferences which may change throughout their lives, including same-sex sexuality and androgénité. He argued that all sexual expressions should be enjoyed as long as people are not abused and that "affirming one's difference" can actually enhance social integration. In Oscar Wilde's The Soul of Man Under Socialism, he advocates an egalitarian society where wealth is shared by all, while warning of the dangers of social systems that crush individuality. Edward Carpenter actively campaigned for homosexual rights. His work The Intermediate Sex: A Study of Some Transitional Types of Men and Women was a 1908 book arguing for gay liberation. who was an influential personality in the foundation of the Fabian Society and the Labour Party. After the Russian Revolution under the leadership of Lenin and Trotsky, the Soviet Union abolished previous laws against homosexuality. Harry Hay was an early leader in the American LGBT rights movement as well as a member of the Communist Party USA. He is known for his involvement in the founding of gay organisations, including the Mattachine Society, the first sustained gay rights group in the United States which in its early days reflected a strong Marxist influence. The Encyclopedia of Homosexuality reports that "[a]s Marxists the founders of the group believed that the injustice and oppression which they suffered stemmed from relationships deeply embedded in the structure of American society". Emerging from events such as the May 1968 insurrection in France, the anti-Vietnam war movement in the US and the Stonewall riots of 1969, militant gay liberation organisations began to spring up around the world. Many sprang from left radicalism more than established homophile groups, although the Gay Liberation Front took an anti-capitalist stance and attacked the nuclear family and traditional gender roles. Eco-socialism is a political strain merging aspects of socialism, Marxism or libertarian socialism with green politics, ecology and alter-globalisation. Eco-socialists generally claim that the expansion of the capitalist system is the cause of social exclusion, poverty, war and environmental degradation through globalisation and imperialism under the supervision of repressive states and transnational structures. Contrary to the depiction of Karl Marx by some environmentalists, social ecologists and fellow socialists as a productivist who favoured the domination of nature, eco-socialists revisited Marx's writings and believe that he "was a main originator of the ecological world-view". Marx discussed a "metabolic rift" between man and nature, stating that "private ownership of the globe by single individuals will appear quite absurd as private ownership of one man by another" and his observation that a society must "hand it [the planet] down to succeeding generations in an improved condition". English socialist William Morris is credited with developing principles of what was later called eco-socialism. During the 1880s and 1890s, Morris promoted his ideas within the Social Democratic Federation and Socialist League. Green anarchism blends anarchism with environmental issues. An important early influence was Henry David Thoreau and his book Walden as well as Élisée Reclus. In the late 19th century, anarcho-naturism fused anarchism and naturist philosophies within individualist anarchist circles in France, Spain, Cuba and Portugal. Murray Bookchin's first book Our Synthetic Environment was followed by his essay "Ecology and Revolutionary Thought" which introduced ecology as a concept in radical politics. In the 1970s, Barry Commoner, claimed that capitalist technologies were chiefly responsible for environmental degradation as opposed to population pressures. In the 1990s socialist/feminists Mary Mellor and Ariel Salleh adopt an eco-socialist paradigm. An "environmentalism of the poor" combining ecological awareness and social justice has also become prominent. Pepper critiqued the current approach of many within green politics, particularly deep ecologists. Syndicalism operates through industrial trade unions. It rejects state socialism and the use of establishment politics. Syndicalists reject state power in favour of strategies such as the general strike. Syndicalists advocate a socialist economy based on federated unions or syndicates of workers who own and manage the means of production. Some Marxist currents advocate syndicalism, such as De Leonism. Anarcho-syndicalism views syndicalism as a method for workers in capitalist society to gain control of an economy. The Spanish Revolution was largely orchestrated by the anarcho-syndicalist trade union CNT. The International Workers' Association is an international federation of anarcho-syndicalist labour unions and initiatives. According to analytical Marxist sociologist Erik Olin Wright, "The Right condemned socialism as violating individual rights to private property and unleashing monstrous forms of state oppression", while "the Left saw it as opening up new vistas of social equality, genuine freedom and the development of human potentials." Because of socialism's many varieties, most critiques have focused on a specific approach. Proponents of one approach typically criticise others. Socialism has been criticised in terms of its models of economic organization as well as its political and social implications. Other critiques are directed at the socialist movement, parties, or existing states. Some forms of criticism occupy theoretical grounds, such as in the economic calculation problem presented by proponents of the Austrian School as part of the socialist calculation debate, while others support their criticism by examining historical attempts to establish socialist societies. The economic calculation problem concerns the feasibility and methods of resource allocation for a planned socialist system. Central planning is also criticized by elements of the radical left. Libertarian socialist economist Robin Hahnel notes that even if central planning overcame its inherent inhibitions of incentives and innovation, it would nevertheless be unable to maximize economic democracy and self-management, which he believes are concepts that are more intellectually coherent, consistent and just than mainstream notions of economic freedom. Economic liberals and right-libertarians argue that private ownership of the means of production and market exchange are natural entities or moral rights which are central to freedom and liberty and argue that the economic dynamics of capitalism are immutable and absolute. As such, they also argue that public ownership of the means of production and economic planning are infringements upon liberty. Critics of socialism have argued that in any society where everyone holds equal wealth, there can be no material incentive to work because one does not receive rewards for a work well done. They further argue that incentives increase productivity for all people and that the loss of those effects would lead to stagnation. Some critics of socialism argue that income sharing reduces individual incentives to work and therefore incomes should be individualized as much as possible. Some philosophers have also criticized the aims of socialism, arguing that equality erodes away at individual diversities and that the establishment of an equal society would have to entail strong coercion. Milton Friedman argued that the absence of private economic activity would enable political leaders to grant themselves coercive powers, powers that, under a capitalist system, would instead be granted by a capitalist class, which Friedman found preferable. Many commentators on the political right point to the mass killings under communist regimes, claiming them as an indictment of socialism. Opponents of this view, including supporters of socialism, state that these killings were aberrations caused by specific authoritarian regimes, and not caused by socialism itself, and draw comparisons to killings and excess deaths under colonialism and anti-communist authoritarian governments.
[[File:Lenin ataki.jpg|thumb|300px|A statue of Lenin, in Ataki.]] '''Leninism''' is a way of thinking about how the [[Communism|communist]] [[Political party|party]] should be organized. It says it should be a [[dictatorship of the proletariat]] (meaning the [[working class]] holds the power). Some [[Historian|historians]] say that Leninism is one of the first steps towards [[socialism]] (where the workers own the [[Factory|factories]] and other [[means of production]]).<ref name="Modern Thought Third Edition 1999 pp. 476">''The New Fontana Dictionary of Modern Thought'' Third Edition (1999) pp. 476–477.</ref> It is one part of [[Marxism–Leninism]], which emphasizes the transition from [[capitalism]] to socialism. ==Beliefs== [[Vladimir Lenin]] was a Russian [[Marxism|Marxist]]. He had a set of ideas based on [[Marxism]]. Lenin's development of Marxism has become known as '''Leninism'''. These ideas include: * [[Democratic Centralism]], also known as an idea of the vanguard party. Like other [[Communism|communists]], Lenin wanted to see a socialist [[revolution]] led by the [[working class]]. But he thought the [[worker]]s needed strong leadership in the form of a Revolutionary Party based on Democratic Centralism. Lenin wanted Communist [[political party|political parties]] in every country to lead the revolution. He thought the vanguard party would need to have strong discipline, or it would fail. * The idea that [[capitalism]] is the cause of [[imperialism]] ([[empire]]-building). He thought that imperialism was the "highest stage" of capitalism. * Accepting the idea that the oppressed [[Ethnic group|ethnic]] [[Minority group|minorities]] (smaller groups of people) should get to have [[nationalism]] and decide how they should be [[Government|governed]].<ref>Lenin, V.I. (1914) ''The Right of Nations to Self-Determination'', from ''Lenin's Collected Works'', Progress Publishers, 1972, Moscow, Volume 20, pp. 393-454. Available online at [http://marxists.org/archive/lenin/works/1914/self-det/index.htm Marxists.org]. Retrieved 17 July 2021</ref> * Teaching the proletariat (the working class) about [[politics]], especially Marxism.<ref>Central Committee, ''On Proletcult Organisations'', ''Pravda'' No. 270, 1 December 1920.</ref> == Related pages == * [[Vladimir Lenin]] * [[Marxism-Leninism]] * [[Communism]] == References == <references /> == Other websites == === Books by Vladimir Lenin === * [http://www.marxists.org/archive/lenin/works/1901/witbd/ ''What Is To Be Done?'']. * [http://www.marxists.org/archive/lenin/works/1916/imp-hsc/index.htm ''Imperialism: The Highest Stage of Capitalism'']. * [http://www.marxists.org/archive/lenin/works/1917/staterev/ch01.htm#s1 ''The State and Revolution'']. * [http://www.marxists.org/archive/lenin/index.htm "The Lenin Archive"]. * [http://www.marx2mao.com/Lenin/FCCI19.html "First Conference of the Communist International"]. === Other similar links === * [http://members.optushome.com.au/spainter/Liebman.html "Marcel Liebman on Lenin and democracy"] {{Webarchive|url=https://web.archive.org/web/20110927013619/http://members.optushome.com.au/spainter/Liebman.html |date=2011-09-27 }}. * [https://web.archive.org/web/20080416210712/http://www.zmag.org/chomsky/other/intellectuals-state.html "An excerpt on Leninism and State Capitalism from the work of Noam Chomsky"]. * [[Rosa Luxemburg]]. [http://www.marxists.org/archive/luxemburg/1904/questions-rsd/index.htm "Organizational Questions of the Russian Social Democracy"]. * [[Karl Korsch]]. [https://web.archive.org/web/20091028121450/http://geocities.com/~johngray/lenphl13.htm "Lenin's Philosophy"]. * [http://www.leninism.org/ "Cyber Leninism"]. * [http://leninist.biz/en/HTML "Leninist Ebooks"] {{Webarchive|url=https://web.archive.org/web/20150126064552/http://leninist.biz/en/HTML |date=2015-01-26 }}. * [[Anton Pannekoek]]. [http://www.marxists.org/archive/pannekoe/1938/lenin/index.htm "Lenin as a Philosopher"]. * [[Paul Mattick]]. [http://www.marxists.org/archive/mattick-paul/1935/lenin-legend.htm "The Lenin Legend"]. * [[Paul Craig Roberts]]. [https://web.archive.org/web/20110805112206/http://www.counterpunch.org/roberts10072009.html "Dead Labor: Marx and Lenin Reconsidered"]. [[Category:Communism]] [[Category:Marxism]] {{politics-stub}}
Leninism is a political ideology developed by Russian Marxist revolutionary Vladimir Lenin that proposes the establishment of the dictatorship of the proletariat led by a revolutionary vanguard party as the political prelude to the establishment of communism. Lenin's ideological contributions to the Marxist ideology relate to his theories on the party, imperialism, the state, and revolution. The function of the Leninist vanguard party is to provide the working classes with the political consciousness (education and organisation) and revolutionary leadership necessary to depose capitalism. Leninist revolutionary leadership is based upon The Communist Manifesto (1848), identifying the communist party as "the most advanced and resolute section of the working class parties of every country; that section which pushes forward all others." As the vanguard party, the Bolsheviks viewed history through the theoretical framework of dialectical materialism, which sanctioned political commitment to the successful overthrow of capitalism, and then to instituting socialism; and, as the revolutionary national government, to realise the socio-economic transition by all means. In the aftermath of the October Revolution in 1917, Leninism was the dominant version of Marxism in Russia. In establishing the socialist mode of production in Soviet Russia – with the 1917 Decree on Land, war communism (1918–1921) and the New Economic Policy (1921–1928) – the revolutionary régime suppressed most political opposition, including Marxists who opposed Lenin's actions, the anarchists and the Mensheviks, factions of the Socialist Revolutionary Party and the Left Socialist-Revolutionaries. The Russian Civil War (1917–1922), which included the fight against the White Army, Entente intervention, left-wing uprisings against the Bolsheviks and widespread peasant rebellions was an external and internal war which transformed Bolshevik Russia into the Russian Socialist Federative Soviet Republic (RSFSR), the core and largest republic that founded the Union of Soviet Socialist Republics (USSR). As revolutionary praxis, Leninism originally was neither a proper philosophy nor a discrete political theory. Leninism comprises politico-economic developments of orthodox Marxism and Lenin's interpretations of Marxism, which function as a pragmatic synthesis for practical application to the actual conditions (political, social, economic) of the post-emancipation agrarian society of Imperial Russia in the early 20th century. As a political-science term, Lenin's theory of proletarian revolution entered common usage at the fifth congress of the Communist International (1924), when Grigory Zinoviev applied the term Leninism to denote "vanguard-party revolution." Leninism was accepted as part of Russian Communist Party (b)'s vocabulary and doctrine around 1922, and in January 1923, despite objections from Lenin, it entered the public vocabulary. In the 19th century, Karl Marx and Friedrich Engels wrote the Manifesto of the Communist Party (1848), in which they called for the political unification of the European working classes in order to achieve communist revolution; and proposed that because the socio-economic organisation of communism was of a higher form than that of capitalism, a workers' revolution first would occur in the industrialised countries. In Germany, Marxist social democracy was the political perspective of the Social Democratic Party of Germany, inspiring Russian Marxists, such as Lenin. In the early 20th century, the socio-economic backwardness of Imperial Russia (1721–1917) — characterized by combined and uneven economic development — facilitated rapid and intensive industrialisation, which produced a united, working-class proletariat in a predominantly agrarian society. Moreover, because industrialisation was financed chiefly with foreign capital, Imperial Russia did not possess a revolutionary bourgeoisie with political and economic influence upon the workers and the peasants, as had been the case in the French Revolution (1789–1799) in the 18th century. Although Russia's political economy was agrarian and semi-feudal, the task of democratic revolution fell to the urban, industrial working class as the only social class capable of effecting land reform and democratisation, in view that the Russian bourgeoisie would suppress any revolution. In the April Theses (1917), the political strategy of the October Revolution (7–8 November 1917), Lenin proposed that the Russian revolution was not an isolated national event but a fundamentally international event—the first socialist revolution in the world. Lenin's practical application of Marxism and proletarian revolution to the social, political, and economic conditions of agrarian Russia motivated and impelled the "revolutionary nationalism of the poor" to depose the absolute monarchy of the three-hundred-year dynasty of the House of Romanov (1613–1917), as tsars of Russia. In Imperialism, the Highest Stage of Capitalism (1916), Lenin's economic analyses indicated that capitalism would transform into a global financial system, by which industrialised countries exported financial capital to their colonies and so realise the exploitation of the labour of the natives and the exploitation of the natural resources of their countries. Such superexploitation allows wealthy countries to maintain a domestic labour aristocracy with a slightly higher standard of living than most workers, ensuring peaceful labour–capital relations in the capitalist homeland. Therefore, a proletarian revolution of workers and peasants could not occur in capitalist countries whilst the imperialist global-finance system remained in place. The first proletarian revolution would have to occur in an underdeveloped country, such as Imperial Russia, the politically weakest country in the capitalist global-finance system in the early 20th century. In the United States of Europe Slogan (1915), Lenin wrote: Workers of the world, unite!—Uneven economic and political development is an absolute law of capitalism. Hence the victory of socialism is possible, first in several, or even in one capitalist country taken separately. The victorious proletariat of that country, having expropriated the capitalists and organised its own socialist production, would stand up against the rest of the world, the capitalist world. In "Left-Wing" Communism: An Infantile Disorder (1920), Lenin wrote: The more powerful enemy can be vanquished only by exerting the utmost effort, and by the most thorough, careful, attentive, skillful and obligatory use of any, even the smallest, rift between the enemies, any conflict of interests among the bourgeoisie of the various countries and among the various groups or types of bourgeoisie within the various countries, and also by taking advantage of any, even the smallest, opportunity of winning a mass ally, even though this ally is temporary, vacillating, unstable, unreliable and conditional. Those who do not understand this reveal a failure to understand even the smallest grain of Marxism, of modern scientific socialism in general. Those who have not proved in practice, over a fairly considerable period of time and in fairly varied political situations, their ability to apply this truth in practice have not yet learned to help the revolutionary class in its struggle to emancipate all toiling humanity from the exploiters. And this applies equally to the period before and after the proletariat has won political power. In Chapter II, "Proletarians and Communists", of The Communist Manifesto (1848), Marx and Engels present the communist party as the political vanguard solely qualified to lead the proletariat in revolution: The Communists, therefore, are, on the one hand, practically the most advanced and resolute section of the working-class parties of every country, that section which pushes forward all others; on the other hand, theoretically, they have over the great mass of the proletariat the advantage of clearly understanding the lines of march, the conditions, and the ultimate general results of the proletarian movement. The immediate aim of the Communists is the same as that of all other proletarian parties: Formation of the proletariat into a class, overthrow of the bourgeois supremacy, conquest of political power by the proletariat. The revolutionary purpose of the Leninist vanguard party is to establish the dictatorship of the proletariat with the working class's support. The communist party would lead the popular deposition of the Tsarist government and then transfer government power to the working class; that change of the ruling class—from the bourgeoisie to the proletariat—makes establishing socialism possible. In What Is To Be Done? (1902), Lenin said that a revolutionary vanguard party, recruited from the working class, should lead the political campaign because only in that way would the proletariat successfully realise their revolution; unlike the economic campaign of trade-union-struggle advocated by other socialist political parties and the anarcho-syndicalists. Like Marx, Lenin distinguished between the aspects of a revolution, the "economic campaign" (labour strikes for increased wages and work concessions) that featured diffused plural leadership; and the "political campaign" (socialist changes to society), which required the decisive, revolutionary leadership of the Bolshevik vanguard party. Based upon the First International (IWA, International Workingmen's Association, 1864–1876), Lenin organised the Bolsheviks as a democratically centralised vanguard party; wherein free political speech was recognised as legitimate until policy consensus; afterwards, every member of the party was expected to abide by the agreed policy. Democratic debate was Bolshevik practice, even after Lenin banned factions among the Party in 1921. Despite being a guiding political influence, Lenin did not exercise absolute power and continually debated to have his points of view accepted as a course of revolutionary action. In Freedom to Criticise and Unity of Action (1905), Lenin said: Of course, the application of this principle in practice will sometimes give rise to disputes and misunderstandings; but only on the basis of this principle can all disputes and all misunderstandings be settled honourably for the Party. ... The principle of democratic centralism and autonomy for local Party organisations implies universal and full freedom to criticise, so long as this does not disturb the unity of a definite action; it rules out all criticism which disrupts or makes difficult the unity of an action decided on by the Party. Before the October Revolution, despite supporting moderate political reform—including Bolsheviks elected to the Duma when opportune—Lenin said that capitalism could only be overthrown with proletarian revolution, not with gradual reforms—from within (Fabianism) and from without (social democracy)—which would fail because the bourgeoisie's control of the means of production determined the nature of political power in Russia. As epitomised in the slogan "For a Democratic Dictatorship of the Proletariat and Peasantry", a proletarian revolution in underdeveloped Russia required a united proletariat (peasants and industrial workers) to assume government power in the cities successfully. Moreover, owing to the middle-class aspirations of much of the peasantry, Leon Trotsky said that the proletarian leadership of the revolution would ensure truly socialist and democratic socio-economic change. In Bolshevik Russia, government by direct democracy was realised and effected by the soviets (elected councils of workers), which Lenin said was the "democratic dictatorship of the proletariat" postulated in orthodox Marxism. The soviets comprised representative committees from the factories and the trade unions but excluded the capitalist social class to establish a proletarian government by and for the working class and the peasants. Concerning the political disenfranchisement of the capitalist social class in Bolshevik Russia, Lenin said that "depriving the exploiters of the franchise is a purely Russian question, and not a question of the dictatorship of the proletariat, in general. ... In which countries ...democracy for the exploiters will be, in one or another form, restricted ...is a question of the specific national features of this or that capitalism." In chapter five of The State and Revolution (1917), Lenin describes the dictatorship of the proletariat as: the organisation of the vanguard of the oppressed as the ruling class for the purpose of crushing the oppressors. ... An immense expansion of democracy, which, for the first time, becomes democracy for the poor, democracy for the people, and not democracy for the rich ... and suppression by force, i.e. exclusion from democracy, for the exploiters and oppressors of the people—this is the change which democracy undergoes during the 'transition' from capitalism to communism. Concerning the disenfranchisement from democracy of the capitalist social class, Lenin said: "Democracy for the vast majority of the people, and suppression by force, i.e. exclusion from democracy, of the exploiters and oppressors of the people—this is the change democracy undergoes during the transition from capitalism to communism." The dictatorship of the proletariat was effected with soviet constitutionalism, a form of government opposite to the dictatorship of capital (privately owned means of production) practised in bourgeois democracies. Under soviet constitutionalism, the Leninist vanguard party would be one of many political parties competing for election to government power. Nevertheless, because of the Russian Civil War (1917–1924) and the anti-Bolshevik terrorism of opposing political parties aiding the White Armies' counter-revolution, the Bolshevik government banned all other political parties, which left the Leninist vanguard party as the only political party in Russia. Lenin said that such political suppression was not philosophically inherent to the dictatorship of the proletariat. The Bolshevik government nationalised industry and established a foreign-trade monopoly to allow the productive coordination of the national economy and so prevent Russian national industries from competing against each other. To feed the populaces of town and country, Lenin instituted war communism (1918–1921) as a necessary condition—adequate supplies of food and weapons—for fighting the Russian Civil War. In March 1921, the New Economic Policy (NEP, 1921–1929) allowed limited local capitalism (private commerce and internal free trade) and replaced grain requisitions with an agricultural tax managed by state banks. The NEP was meant to resolve food-shortage riots by the peasantry and allowed limited private enterprise; the profit motive encouraged farmers to produce the crops required to feed town and country; and to economically re-establish the urban working class, who had lost many workers to fight the counter-revolutionary Civil War. The NEP nationalisation of the economy then would facilitate the industrialisation of Russia, politically strengthen the working class, and raise the standards of living for all Russians. Lenin said that the appearance of new socialist states was necessary for strengthening Russia's economy in establishing Russian socialism. Lenin's socio-economic perspective was supported by the German Revolution of 1918–1919, the Italian insurrection and general strikes of 1920, and worker wage-riots in the UK, France, and the US. In recognising and accepting nationalism among oppressed peoples, Lenin advocated their national right to self-determination and so opposed Russian chauvinism because such ethnocentrism was a cultural obstacle to establishing the dictatorship of the proletariat in every territory of the deposed Russian Empire (1721–1917). In The Right of Nations to Self-determination (1914), Lenin said: We fight against the privileges and violence of the oppressor nation, and do not in any way condone strivings for privileges on the part of the oppressed nation. :... The bourgeois nationalism of any oppressed nation has a general democratic content that is directed against oppression, and it is this content that we unconditionally support. At the same time, we strictly distinguish it from the tendency towards national exclusiveness. ... Can a nation be free if it oppresses other nations? It cannot. The socialist internationalism of Marxism and Bolshevism is based upon class struggle and a people's transcending nationalism, ethnocentrism, and religion—the intellectual obstacles to progressive class consciousness—which are the cultural status quo that the capitalist ruling class manipulates in order to divide the working classes and the peasant classes politically. To overcome that barrier to establishing socialism, Lenin said that acknowledging nationalism, as a people's right of self-determination and right of secession, naturally would allow socialist states to transcend the political limitations of nationalism to form a federation. In The Question of Nationalities, or 'Autonomisation' (1923), Lenin said: [N]othing holds up the development and strengthening of proletarian class solidarity so much as national injustice; "offended" nationals are not sensitive to anything, so much as to the feeling of equality, and the violation of this equality, if only through negligence or jest – to the violation of that equality by their proletarian comrades. The role of the Leninist vanguard party was to politically educate the workers and peasants to dispel the societal false consciousness of religion and nationalism that constitute the cultural status quo taught by the bourgeoisie to the proletariat to facilitate their economic exploitation of peasants and workers. Influenced by Lenin, the Central Committee of the Bolshevik Party stated that the development of the socialist workers' culture should not be "hamstrung from above" and opposed the Proletkult (1917–1925) organisational control of the national culture. In post-Revolutionary Russia, Stalinism (socialism in one country) and Trotskyism (permanent world revolution) were the principal philosophies of communism that claimed legitimate ideological descent from Leninism; thus, within the Communist Party, each ideological faction denied the political legitimacy of the opposing faction. Until shortly before his death, Lenin countered Stalin's disproportionate political influence in the Communist Party and the bureaucracy of the Soviet government, partly because of abuses he had committed against the populace of Georgia and partly because the autocratic Stalin had accumulated administrative power disproportionate to his office of General Secretary of the Communist Party. The counter-action against Stalin aligned with Lenin's advocacy of the right of self-determination for the national and ethnic groups of the deposed Tsarist Empire. Lenin warned the Party that Stalin had "unlimited authority concentrated in his hands, and I am not sure whether he will always be capable of using that authority with sufficient caution" and formed a faction with Leon Trotsky to remove Stalin as the General Secretary of the Communist Party. To that end followed proposals reducing the administrative powers of party posts to reduce bureaucratic influence upon the policies of the Communist Party. Lenin advised Trotsky to emphasise Stalin's recent bureaucratic alignment in such matters (e.g. undermining the anti-bureaucratic workers' and peasants' Inspection) and argued to depose Stalin as General Secretary. Despite advice to refuse "any rotten compromise", he did not heed Lenin's advice and General Secretary Stalin retained power over the Communist Party and the bureaucracy of the Soviet government. In 1922, Lenin allied with Leon Trotsky against the party's growing bureaucratisation and the influence of Joseph Stalin. Lenin himself never mentioned the concept of "Trotskyism" after Trotsky became a member of the Bolshevik party but the term was employed by Stalin and the troika to present Trotsky's views as factional and anathematical to Leninist thought. After Lenin's death (21 January 1924), Trotsky ideologically battled the influence of Stalin, who formed ruling blocs within the Russian Communist Party (with Grigory Zinoviev and Lev Kamenev, then with Nikolai Bukharin and then by himself) and so determined soviet government policy from 1924 onwards. The ruling blocs continually denied Stalin's opponents the right to organise as an opposition faction within the party—thus, the reinstatement of democratic centralism and free speech within the Communist Party were key arguments of Trotsky's Left Opposition and the later Joint Opposition. In instituting government policy, Stalin promoted the doctrine of socialism in one country (adopted 1925), wherein the Soviet Union would establish socialism upon Russia's economic foundations (and support socialist revolutions elsewhere). In a 1936 interview with journalist Roy W. Howard, Stalin articulated his rejection of world revolution and stated that “We never had such plans and intentions” and that “The export of revolution is nonsense”. Conversely, Trotsky held that socialism in one country would economically constrain the industrial development of the Soviet Union and thus required assistance from the new socialist countries in the developed world—which was essential for maintaining soviet democracy—in 1924, much undermined by the Russian Civil War of White Army counter-revolution. Trotsky's theory of permanent revolution proposed that socialist revolutions in underdeveloped countries would further dismantle feudal régimes and establish socialist democracies that would not pass through a capitalist stage of development and government. Hence, revolutionary workers should ally politically with peasant political organisations, not capitalist political parties. In contrast, Stalin and his allies proposed that alliances with capitalist political parties were essential to realising a revolution where communists were too few. Said Stalinist practice failed, especially in the Northern Expedition portion of the Chinese Revolution (1926–1928), which resulted in the right-wing Kuomintang's massacre of the Chinese Communist Party. Despite the failure, Stalin's policy of mixed-ideology political alliances nonetheless became Comintern's policy. Until exiled from Russia in 1929, Trotsky developed and led the Left Opposition (and the later Joint Opposition) with members of the Workers' Opposition, the Decembrists and (later) the Zinovievists. Trotskyism predominated the politics of the Left Opposition, which demanded the restoration of soviet democracy, the expansion of democratic centralism in the Communist Party, national industrialisation, international permanent revolution and socialist internationalism. The Trotskyist demands countered Stalin's political dominance of the Communist Party, which was officially characterised by the "cult of Lenin", the rejection of permanent revolution, and advocated the doctrine of socialism in one country. The Stalinist economic policy vacillated between appeasing the capitalist interests of the kulak in the countryside and destroying them as a social class. Initially, the Stalinists also rejected the national industrialisation of Russia but then pursued it in full, sometimes brutally. In both cases, the Left Opposition denounced the regressive nature of Stalin's policy towards the wealthy kulak social class and the brutality of forced industrialisation. Trotsky described Stalinist vacillation as a symptom of the undemocratic nature of a ruling bureaucracy. During the 1920s and the 1930s, Stalin fought and defeated the political influence of Trotsky and the Trotskyists in Russia using slander, antisemitism, censorship, expulsions, exile (internal and external), and imprisonment. The anti-Trotsky campaign culminated in the executions (official and unofficial) of the Moscow Trials (1936–1938), which were part of the Great Purge of Old Bolsheviks who had led the Revolution. Some historians such as Richard Pipes consider Stalinism as the natural consequence of Leninism, that Stalin "faithfully implemented Lenin's domestic and foreign policy programs". Robert Service notes that "institutionally and ideologically Lenin laid the foundations for a Stalin ... but the passage from Leninism to the worse terrors of Stalinism was not smooth and inevitable." Historian and Stalin biographer Edvard Radzinsky believes that Stalin was a genuine follower of Lenin, exactly as he claimed himself. Proponents of continuity cite a variety of contributory factors, in that it was Lenin, rather than Stalin, whose civil war measures introduced the Red Terror with its hostage-taking and internment camps; that it was Lenin who developed the infamous Article 58 and who established the autocratic system within the Russian Communist Party. Proponents also note that Lenin put a ban on factions within the party and introduced the one-party state in 1921, a move that enabled Stalin to get rid of his rivals easily after Lenin's death and cite Felix Dzerzhinsky, who exclaimed during the Bolshevik struggle against opponents in the Russian Civil War: "We stand for organized terror—this should be frankly stated." Some scholars have had a differing view and attributed the establishment of the one-party system in the Soviet Union to the wartime conditions imposed on Lenin’s government and others have highlighted the initial attempts to form a coalition government with the Left Socialist Revolutionaries. According to historian Marcel Liebman, Lenin’s wartime measures such as banning opposition parties was prompted by the fact that several political parties either took up arms against the new Soviet government, or participated in sabotage, collaborated with the deposed Tsarists, or made assassination attempts against Lenin and other Bolshevik leaders. Liebman also argued that the banning of parties under Lenin did not have the same repressive character as later bans enforced under the Stalinist regime. Several scholars have highlighted the socially progressive nature of Lenin's policies such as universal education, healthcare and equal rights for women. Conversely, Stalin's regime reversed Lenin's policies on social matters such as sexual equality, legal restrictions on marriage, rights of sexual minorities and protective legislation. Historian Robert Vincent Daniels also viewed the Stalinist period as a counter-revolution in Soviet cultural life which revived patriotic propaganda, the Tsarist programme of Russification and traditional, military ranks which had been criticized by Lenin as expressions of "Great Russian chauvinism". Daniels also regarded Stalinism to represent an abrupt break with the Leninist period in terms of economic policies in which a deliberated, scientific system of economic planning that featured former Menshevik economists at Gosplan had been replaced with a hasty version of planning with unrealistic targets, bureaucratic waste, bottlenecks and shortages. Revisionist historians and some post-Cold War and otherwise dissident Soviet historians, including Roy Medvedev, argue that "one could list the various measures carried out by Stalin that were actually a continuation of anti-democratic trends and measures implemented under Lenin", but that "in so many ways, Stalin acted, not in line with Lenin's clear instructions, but in defiance of them." In doing so, some historians have tried to distance Stalinism from Leninism to undermine the totalitarian view that the negative facets of Stalin were inherent in communism from the start. Critics include anti-Stalinist communists such as Leon Trotsky, who pointed out that Lenin attempted to persuade the Russian Communist Party to remove Stalin from his post as its General Secretary. Lenin's Testament, the document which contained this order, was suppressed after Lenin's death. In his biography of Trotsky, Polish-British historian Isaac Deutscher says that, on being faced with the evidence, "only the blind and the deaf could be unaware of the contrast between Stalinism and Leninism." According to Stalin’s secretary, Boris Bazhanov, Stalin was jubilant over Lenin’s death while “publicly putting on the mask of grief”. French historian Pierre Broue disputed the historical assessments of the early Soviet Union by modern historians such as Dmitri Volkogonov in which he argued had falsely equated Leninism, Stalinism and Trotskyism to present the notion of ideological continuity and reinforce the position of counter-communism. Nikita Khrushchev, Stalin’s successor, argued that Stalin’s regime differed profusely from the leadership of Lenin in his "Secret Speech", delivered in 1956. He was critical of the cult of the individual which was constructed around Stalin whereas Lenin stressed “the role of the people as the creator of history”. He also emphasized that Lenin favored a collective leadership which relied on personal persuasion and recommended the removal of Stalin from the position of General Secretary. Khrushchev contrasted this with the “despotism” of Stalin which require absolute submission to his position and he also highlighted that many of the people who were later annihilated as “enemies of the party", "had worked with Lenin during his life”. He also contrasted the “severe methods” used by Lenin in the “most necessary cases” as a “struggle for survival” during the Civil War with the extreme methods and mass repressions used by Stalin even when the Revolution was “already victorious”. In his memoirs, Khrushchev argued that Stalin's widespread purges of the "most advanced nucleus of people" among the Old Bolsheviks and leading figures in the military and scientific fields had "undoubtedly" weakened the nation. Some Marxist theoreticians have disputed the view that the Stalinist dictatorship was a natural outgrowth of the Bolsheviks' actions as most of the original central committee members from 1917 were later eliminated by Stalin. George Novack stressed the initial efforts by the Bolsheviks to form a government with the Left Socialist Revolutionaries and bring other parties such as the Mensheviks into political legality. Tony Cliff argued the Bolshevik-Left Socialist Revolutionary coalition government dissolved the democratically elected Russian Constituent Assembly due to a number of reasons. They cited the outdated voter-rolls which did not acknowledge the split among the Socialist Revolutionary party and the assemblies conflict with the Russian Congress of the Soviets as an alternative democratic structure. A similar analysis is present in more recent works such as those of Graeme Gill, who argues that "[Stalinism was] not a natural flow-on of earlier developments; [it formed a] sharp break resulting from conscious decisions by leading political actors." However, Gill notes that "difficulties with the use of the term reflect problems with the concept of Stalinism itself. The major difficulty is a lack of agreement about what should constitute Stalinism." Revisionist historians such as Sheila Fitzpatrick have criticized the focus on the upper levels of society and the use of Cold War concepts such as totalitarianism, obscuring the system's reality. Russian historian Vadim Rogovin stated that "Under Lenin, the freedom to express a real variety of opinions existed in the party, and in carrying out political decisions, consideration was given to the positions of not only the majority, but a minority in the party". He compared this practice with subsequent leadership blocs which violated party tradition, ignored proposals of opponents and expelled the Opposition from the party on falsified charges which culminated with the Moscow Trials of 1936-1938. According to Rogovin, 80-90% of the members of the Central Committee elected at the Sixth through to the Seventeenth Congresses were physically annihilated. As a form of Marxism, revolutionary Leninism was criticised as an undemocratic interpretation of socialism. In The Nationalities Question in the Russian Revolution (1918), Rosa Luxemburg criticised the Bolsheviks for the suppression of the All Russian Constituent Assembly (January 1918); the partitioning of the feudal estates to the peasant communes; and the right of self-determination of every national people of the Russias. That the strategic (geopolitical) mistakes of the Bolsheviks would create significant dangers for the Russian Revolution, such as the bureaucratisation that would arise to administrate the large country that was Bolshevik Russia. In defence of the expedient revolutionary practice, in "Left-Wing" Communism: An Infantile Disorder (1920), Lenin dismissed the political and ideological complaints of the anti-Bolshevik critics, who claimed ideologically correct stances that were to the political left of Lenin. In Marxist philosophy, left communism is a range of left-wing political perspectives among communists. Left communism criticizes the Bolshevik Party's ideology as the revolutionary vanguard. Ideologically, left communists present their perspectives and approaches as authentic Marxism and thus more oriented to the proletariat than the Leninism of the Communist International at their first (1919) and second (1920) congresses. Proponents of left communism include Amadeo Bordiga, Herman Gorter, Paul Mattick, Sylvia Pankhurst, Antonie Pannekoek and Otto Rühle. Historically, the Dutch-German communist left has been most critical of Lenin and Leninism, yet the Italian communist left remained Leninist. Bordiga said: "All this work of demolishing opportunism and 'deviationism' (Lenin: What Is To Be Done?) is today the basis of party activity. The party follows revolutionary tradition and experiences in this work during these periods of revolutionary reflux and the proliferation of opportunist theories, which had as their violent and inflexible opponents Marx, Engels, Lenin, and the Italian Left." In The Lenin Legend (1935), Paul Mattick said that the council communist tradition, begun by the Dutch-German leftists, also is critical of Leninism. Contemporary left-communist organisations, such as the Internationalist Communist Tendency and the International Communist Current, view Lenin as an essential and influential theorist but remain critical of Leninism as political praxis for the proletarian revolution. Nonetheless, the Bordigism of the International Communist Party abides Bordiga's strict Leninism. Ideologically aligned with the Dutch-German left, among the ideologists of contemporary communisation, the theorist Gilles Dauvé criticised Leninism as a "by-product of Kautskyism". In The Soviet Union Versus Socialism (1986), Noam Chomsky said that Stalinism was the logical development of Leninism and not an ideological deviation from Lenin's policies, which resulted in collectivisation enforced with a police state. In light of the tenets of socialism, Leninism was a right-wing deviation from Marxism. The vanguard-party revolution of Leninism became the ideological basis of the communist parties in the socialist political spectrum. In the People's Republic of China, the Chinese Communist Party organised itself with Maoism (the Thought of Mao Zedong), socialism with Chinese characteristics. In Singapore, the People's Action Party (PAP) featured internal democracy and initiated single-party dominance in the government and politics of Singapore. In the event, the practical application of Maoism to the socio-economic conditions of Third World countries produced revolutionary vanguard parties, such as the Communist Party of Peru – Red Fatherland. Selected works by Vladimir Lenin Histories Other authors Works by Vladimir Lenin Other thematic links
{{Infobox officeholder | name = Saddam Hussein<br><small>صِدَامُ حُسَيْنٍ</small> | height = 6’0½” to 6’2” | image = Saddam Hussein in 1998.png | office = [[President of Iraq]] | term_start = 16 July 1979 | term_end = 9 April 2003 | primeminister = [[Sa'dun Hammadi]]<br>[[Mohammed Amza Zubeidi]]<br>[[Ahmad Husayn Khudayir as-Samarrai]] | predecessor = [[Ahmed Hassan al-Bakr]] | successor = [[Abdul Aziz al-Hakim]] <small>([[Prime Minister of Iraq|President of the Governing council]])</small> | office1 = [[Prime Minister of Iraq]] | term_start1 = 29 May 1994 | term_end1 = 9 April 2003 | predecessor1 = [[Ahmad Husayn Khudayir as-Samarrai]] | successor1 = [[Mohammad Bahr al-Ulloum]] <small>(Acting President of the [[Iraqi Governing Council|Governing Council]])</small> | term_start2 = 16 July 1979 | term_end2 = 23 March 1991 | predecessor2 = [[Ahmed Hassan al-Bakr]] | successor2 = [[Sa'dun Hammadi]] | office3 = Leader of the [[Ba'ath Party (Iraqi-led faction)|Ba'ath Party]] | term_start3 = 23 June 1989 | term_end3 = 30 December 2006 | predecessor3 = [[Michel Aflaq]] | successor3 = [[Izzat Ibrahim ad-Douri]] | office4 = Chairperson of the [[Revolutionary Command Council (Iraq)|Revolutionary Command Council]] | term_start4 = 16 July 1979 | term_end4 = 9 April 2003 | predecessor4 = [[Ahmed Hassan al-Bakr]] | successor4 = Position abolished | birth_date = {{birth_date|1939|7|01|df=y}} | birth_place = [[Al-Awja]], [[Kingdom of Iraq (1932–58)|Iraq]] | death_date = {{death date and age|2006|12|30|1939|7|01|df=y}} | death_place = [[Kadhimiya]], [[Iraq]] | party = [[Ba'ath Party (Iraqi-led faction)|Ba'ath Party]] | otherparty = [[National Progressive Front (Iraq)|National Progressive Front]] | spouse = [[Sajida Talfah]]<br>[[Samira Shahbandar]] | children = [[Uday Hussein|Uday]]<br>[[Qusay Hussein|Qusay]]<br>[[Raghad Hussein|Raghad]]<br>[[Rana Hussein|Rana]]<br>[[Hala Hussein|Hala]] | caption = Saddam Hussein in 1998 | signature = Saddam Hussein Signature.svg }} '''Saddam Hussein''' ([[Arabic language|Arabic]]: {{Text|صِدَامُ حُسَيْنٍ}}; April 28, 1937 – 30 December 2006)<ref name="ref2">Under his government, this date was his official date of birth. His real date of birth was never written down, but it is thought to be between 1935 and 1939. From Con Coughlin, ''Saddam The Secret Life'' Pan Books, 2003 ({{ISBN|0-330-39310-3}}).</ref><ref name="CNN11">{{cite web | title= Hussein executed, Iraqi TV stations report | work=[[CNN]] | url=http://www.cnn.com/2006/WORLD/meast/12/29/hussein/index.html | access-date=December 30, 2006}}</ref> was an Iraqi politician and revolutionary who was the fifth [[president of Iraq]], from July 1979 to April 2003. He was 58th and 61st [[Prime Minister of Iraq|prime minister of Iraq]], serving respectively from July 1979 to March 1991 and from May 1994 to April 2003. He was removed from his position during the [[Iraq War|War in Iraq]] led by the [[United States]]. He ran a repressive [[Authoritarianism|authoritarian]] government,<ref>{{Cite book|title=State of Repression : Iraq under Saddam Hussein |url=https://www.worldcat.org/oclc/1104855351|access-date=2023-03-09|website=www.worldcat.org|oclc=1104855351 |language=en}}</ref> often described as [[Totalitarianism|totalitarian]],<ref>{{Cite journal|last1=Woods|first1=Kevin M.|last2=Stout|first2=Mark E.|date=2010-08-01|title=New Sources for the Study of Iraqi Intelligence during the Saddam Era|url=https://doi.org/10.1080/02684527.2010.537033|journal=Intelligence and National Security|volume=25|issue=4|pages=547–587|doi=10.1080/02684527.2010.537033|s2cid=153605621 |issn=0268-4527}}</ref> which is a disputed label.<ref>{{Cite journal|last=Sassoon|first=Joseph|date=February 2017|title=Aaron M. Faust , The Baʿthification of Iraq: Saddam Hussein's Totalitarianism (Austin, Tex.: University of Texas Press, 2015). Pp. 296. $55.00 cloth.|isbn= 9781477305577|url=https://www.cambridge.org/core/journals/international-journal-of-middle-east-studies/article/abs/aaron-m-faust-the-bathification-of-iraq-saddam-husseins-totalitarianism-austin-tex-university-of-texas-press-2015-pp-296-5500-cloth-isbn-9781477305577/3E2A3E4D523556848C0E24AC9318B019|journal=International Journal of Middle East Studies|language=en|volume=49|issue=1|pages=205–206|doi=10.1017/S0020743816001392|s2cid=164804585 |issn=0020-7438}}</ref> == Childhood == [[File:Саддам в молодости.jpg|thumb|Saddam in his Youth {{Circa|1956}}]] Saddam Hussein was born in the [[village]] of Al-Awja, in the [[Tikrit]] in [[Iraq]]. He never knew his father, Hussein 'Abd al-Majid,<ref>Hussein is not a [[Family name|surname]] in the Western sense. "Saddam" (pronounced "Sad-DAHM") is his [[given name]] or [[personal name]]; "Hussein" is his father's given personal name; "al-Majid" is his family name, and "al-Tikriti" is a name telling what region he was from. In many Arab countries he is usually called "Saddam Hussein' or "Saddam." However, in Iraq, he was and is usually called by his formal presidential title. Some people have argued that calling him only "Saddam" may be rude and academically out of place. It is common for Arab men to add the name of the town or village they are from to their name. This would make his name "Saddam Hussein al-Awja."</ref> who disappeared five months before Saddam was born. Shortly before Saddam was born, Saddam's twelve-year-old brother died of [[cancer]], leaving his [[mother]] very [[clinical depression|depressed]] in the final months of the [[pregnancy]]. She tried to [[Suicide|kill herself]] near the end of the pregnancy and did not want to care for Saddam when he was born. Saddam was sent to the family of an [[uncle]], Khairallah Tulfah, until he was three.<ref>{{Cite news|url=https://www.nytimes.com/2004/05/15/books/was-a-tyrant-prefigured-by-baby-saddam.html|title=Was a Tyrant Prefigured by Baby Saddam?|last=Bumiller|first=Elisabeth|date=May 15, 2004|work=The New York Times|access-date=April 14, 2019|language=en-US|issn=0362-4331}}</ref> At 10, Saddam ran away from the family to return to live with his uncle, who was a devout [[Sunni]] Muslim, in [[Baghdad]]. According to Saddam, in 1957, at the age of 20, Saddam became part of the [[Ba'ath Party]]. The Ba'ath party is an [[Arab people|Arab]] [[political party]] that supports [[socialism]]. In 1958, Hussein was [[arrest]]ed for [[Murder|killing]] his [[brother-in-law]] because he was a [[Communism|communist]] [[Activism|activist]]. He spent six months in [[prison]].<ref>{{cite web|url=http://www.foxnews.com/story/2004/07/01/fast-facts-saddam-hussein/|title=Fast Facts: Saddam Hussein|publisher=Fox News.com|access-date=14 August 2013}}</ref> == Rise in the Ba'ath party == [[File:SaddamCairo.jpg|thumb|Saddam with members of the [[Ba'ath party]] in [[Cairo]]]] A year after Saddam had joined the Ba'ath party, army officers led by [[Abdul Karim Qassim|General Abdul Karim Qassim]] got rid of [[Faisal II of Iraq]]. The Baathists were against the new [[regime]], and in 1959, Saddam was involved in the attempted murder of Prime Minister Qassim. Saddam was [[Gunshot wound|shot]] in the leg, but managed to get away to [[Syria]]. Later, he moved to [[Egypt]]. He was [[Death penalty|sentenced to death]]. In [[exile]], he attended the [[University of Cairo]] [[law school]]. Army officers, including some with the Ba'ath party, came to power in Iraq in a military [[coup]] in 1963. However, the new regime was kicked out quickly. Saddam returned to Iraq, but was imprisoned in 1964 when an anti-Ba'ath group led by [[Abdul Rahman Arif]] took power. He escaped from jail in 1967 and became one of the leading members of the party. == Gaining power == In 1976, Saddam was made a [[General officer|general]] in the Iraqi army. He quickly became the most important person of the [[regime]]. He slowly began to gain more power over Iraq's government and the Ba'ath party. As Iraq's weak, old President, [[Ahmed Hassan al-Bakr]], became more unable to do the duties of his office, Saddam began to take on an more important role as the head of the Iraqi government. He soon became the creator of Iraq's [[foreign policy]] and represented the nation in all [[Diplomatic relations|diplomatic]] situations. === Conflict with Iran === In 1979, Iran's [[Shah]] [[Mohammad Reza Pahlavi]] was overthrown by the [[History of Iran#Islamic Revolution|Islamic Revolution]], giving way to an [[Islamic republic]] led by the [[Ruhollah Khomeini|Ayatollah Khomeini]]. The influence of revolutionary [[Shi'ite]] Islam grew in the region, particularly in countries with large Shi'ite [[population]]s, especially Iraq. Saddam was afraid that radical Islamic ideas were quickly spreading inside his country among most of the Shi'ite people. He worried that these ideas would go against his leadership. There had also been a rivalry between Saddam and Khomeini since the 1970s. Khomeini was [[exile]]d from Iran in 1964. He began living in Iraq, at the Shi'ite holy city of [[An Najaf]]. There, he became involved with Iraqi Shi'ites and gained many religious and [[Politics|political]] followers throughout the world. Under pressure from the Shah, who had agreed to a [[rapprochement]] between Iraq and Iran in 1975, Saddam agreed to kick Khomeini out in 1978. After the Islamic Revolution, Khomeini thought defeating Saddam's government may have been the second most important thing to do, only behind keeping his power in Iran. == The Iran-Iraq War 1980-1988 == After Khomeini gained power, small [[battle]]s between Iraq and revolutionary Iran happened for ten months. The two countries were fighting over who controlled the [[Shatt al-Arab waterway]], which divides Iran and Iraq. Iraq and Iran officially went to war with each other on September 22, 1980. Saddam used the disagreement over the waterway as an excuse to go to war with Iran. However, the war was more likely an attempt by Saddam - supported by both the [[United States]] and the [[Soviet Union]] - to have Iraq stop radical revolutions like the one in Iran from spreading any further. In the first days of the war, there was heavy ground fighting around major [[port]]s as Iraq launched an attack on Iran's [[oil]]-rich, Arab-populated [[province]] of [[Khuzestan]]. After making some gains, Iraq's [[Soldier|troops]] began to suffer losses from human-wave attacks by Iran. By 1982, Iraq was looking to end the war. === Chemical warfare === During the war, Iraq used [[chemical warfare]] against Iranian forces and [[Kurdish people|Kurdish]] [[Separatism|separatists]]. On March 16, 1988, Saddam ordered Iraqi troops to stop a Kurdish uprising. Iraq attacked the Kurdish town of [[Halabja]] with a mix of [[poison gas]] and [[nerve agent]]s, killing 5000 people, mostly women and children. Saddam reached out to other Arab governments for cash and political help. The Iranians, hoping to bring down Saddam's non-religious government and start a Shi'ite [[rebellion]] in Iraq, refused a cease-fire until 1988. === End of the war === The eight-year war ended in a tie. There were hundreds of thousands of [[Casualty (person)|casualties]]. Perhaps 1.7 million died on both sides. Both [[Economy|economies]], previously healthy and growing, were left in ruins. Saddam was also stuck with a [[debt]] of roughly $75 billion. Borrowing money from the U.S. was making Iraq into its client state, embarrassing a strongman who had sought to define and lead Arab [[nationalism]]. Saddam also borrowed a large amount of money from other Arab states during the 1980s to fight Iran. Faced with rebuilding Iraq's [[infrastructure]], Saddam desperately sought out cash once again, this time for postwar reconstruction. <div align=center><gallery perrow="6"> File:Operation Ramadan ul-Mubarak map.svg|The war was over the Shatt al-Arab waterway (the dotted blue and black line) File:Children In iraq-iran war3.jpg|[[Child soldier]] in the Iran-Iraq War File:Sardashtchemic.jpg|Victims of Iraq's attacks with poison gas File:Halabja cemetery.jpg|[[Graves]] for Kurds killed in the [[Halabja chemical attack]] File:Murdered Iraqi POW 2.jpg|A murdered Iraqi [[prisoner of war|POW]] </gallery></div> == After the war: Tensions with Kuwait == Saddam was pressuring [[Kuwait]] to forgive its share of his debt, some $30 billion. (This would mean Iraq would not have to pay back Kuwait's $30 million.) Saddam argued that the struggle with Iran had been fought for the benefit of the other Persian Gulf Arab states as much as for Iraq. Because of this, he said, a share of Iraqi debt should be forgiven. Saddam had pushed oil-[[export]]ing countries to raise oil prices and cut back production. Not only did Kuwait refuse to do this, they also helped support [[OPEC]]'s opposition to the production cuts that Saddam had requested. Kuwait was pumping large amounts of oil, and keeping prices low, when Iraq needed to sell high-priced oil from its [[Oil well|wells]] to pay off a huge debt. The fact that Kuwait had so much oil made the region even more tense. Even though Kuwait had fewer people, it had about as much oil in reserve as Iraq. Together, Iraq and Kuwait had 20 percent of the world's known oil reserves. The Kuwaiti [[monarchy]] made Saddam even angrier by drilling oil out of Iraqi wells. At the time, Saddam's regime was not disliked by most of the world. Saddam complained about the drilling to the [[United States State Department]]. Although this had gone on for years, Saddam now needed oil money to get rid of a looming economic crisis. Saddam still had an experienced and well-equipped army, which he used to influence regional affairs. He later ordered troops to the Iraqi-Kuwaiti [[border]]. As Iraqi-Kuwaiti relations rapidly grew worse, Saddam was getting different information about how the United States would respond to an [[invasion]]. The United States had been working on starting a good relationship with Iraq for roughly a [[decade]]. The [[Ronald Reagan|Reagan administration]] gave Saddam roughly $40 billion worth of [[Weapon|arms]] in the 1980s to fight Iran, nearly all of it on [[credit]]. The U.S. also sent billions of dollars of food and arms to Saddam to keep him from forming a strong [[alliance]] with the [[Soviet Union|Soviets]].<ref>A free access online archive relating to U.S.-Iraq relations in the 1980s is offered by ''The National Security Archive'' of the [[George Washington University]]. It can be read online at [http://www.gwu.edu/~nsarchiv/NSAEBB/NSAEBB82/]. The Mount Holyoke International Relations Program also provides a free access document briefing on U.S.-Iraq relations (1904- present); this can be accessed online at [http://www.mtholyoke.edu/acad/intrel/iraq.htm] {{Webarchive|url=https://web.archive.org/web/20190208105755/https://www.mtholyoke.edu/acad/intrel/iraq.htm|date=2019-02-08}}.</ref> The U.S. [[ambassador]] to Iraq, [[April Glaspie]], met with Saddam in a meeting on July 25. Saddam said he wanted to keep talks going. U.S. officials tried to take a calm, relaxing tone with Iraq. They explained that neither [[President of the United States|President]] [[George H. W. Bush|George H.W. Bush]] or [[Secretary of State]] [[James Baker]] wanted military force to be used. However, they also said that they would not take any position on the Iraqi-Kuwaiti dispute and did not want to become involved. Later, Iraq and Kuwait met for a final [[negotiation]] session, which failed. Saddam then sent his troops into Kuwait. == The Gulf War 1990-1991 == [[File:Various Arabic Troops during Operation Desert Storm.jpg|thumb|Coalition of Arab Countries during the Gulf War]] {{mainarticle|Gulf War}} On August 2, 1990, Saddam invaded Kuwait, causing an international crisis. The invasion of Kuwait gave Iraq, with its own substantial oil fields, control of 20 percent of the [[Persian Gulf]] oil. The U.S. helped Saddam Hussein in the war with Iran, but with Iraq's take over of the oil-rich [[emirate]] of [[Kuwait]] in August of 1990, the United States led a [[United Nations]] force that drove Saddam from Kuwait in [[February]] 1991. Because the United States and the Soviet Union were [[Cooperation|cooperating]] in the [[United Nations Security Council]], the Security Council was able to pass resolutions. These resolutions gave Iraq a [[deadline]] to leave Kuwait. Saddam ignored the Security Council deadline. On January 16, 1991, coalition of U.S. and Security Council troops launched [[missile]] attacks on Iraq. The [[United States]] and a group of [[Ally|allies]] it had quickly gathered, including [[Egypt]], [[Syria]], and [[Saudi Arabia]], made Saddam's army move from Kuwait in January 1991. [[Israel]], though Saddam attacked it with Iraqi missiles, did not fight back. It did not want to anger Arab states into leaving the coalition. But Saddam had focused attention on the [[Palestinian]] problem by promising to make his forces leave from Kuwait if [[Israel]] would leave the [[West Bank]], the [[Golan Heights]], and the [[Gaza Strip]]. Saddam's proposal further split the Arab world, pitting U.S. and Western-supported Arab states against the Palestinians. 175,000 Iraqis were taken prisoner and about 85,000 died. As part of the [[Ceasefire|cease-fire]] agreement, Iraq agreed to get rid of all poison gas and [[Biological weapon|germ weapons]], and to allow UN observers to inspect the sites. === After the war === All of the different [[religion]]s and the [[violence]] the war had created caused after-war [[rebel]]lions. After the war, fighting between Shi'ite Muslims, Kurds, and [[dissident]] military units was bad. This was a problem to Saddam's rule. Saddam acted by stopping all rebellions in their tracks, especially in the North. Before the war ended, the United States had encouraged Iraqis to rise up against Saddam. However, when the Shi'ites, Kurds, and dissidents did rise up against Saddam, the United States did not support them. Without United States support for these rebellions, Saddam survived them. He was then left completely in control of Iraq. The country's [[economy]] and [[army]] never recovered from the Gulf War. However, Saddam often claimed that Iraq had won the Gulf War, and the United States had lost. This made Saddam popular in many parts of the Arab world. Saddam liked to show himself as a strict [[Muslim]]. This was to calm down the religious parts of the society. He brought back some parts of [[Sharia law]]. This included the 2001 law that said [[homosexuality]] could be punished by the [[death penalty]]. The phrase "[[Allahu Akbar]]" ("''[[God]] is great''"), in Saddam's handwriting, was added to the Iraq national [[flag]]. == 1991-2003 == Relations between the United States and Iraq remained tense after the Gulf War. In 1993, the United States decided to attack Iraq, because it thought [[evidence]] showed Iraq had sponsored a plan to kill former President George H.W. Bush. On June 26, 1993, the United States launched a missile attack aimed at Iraq's [[Military intelligence|intelligence]] [[headquarters]] in Baghdad. The United Nations placed a trade [[embargo]] on Iraq, blocking Iraqi oil [[export]]s. This caused hardship in Iraq and almost destroyed the Iraqi economy and state [[infrastructure]]. Only smuggling across the [[Syria]]n border, and [[humanitarian aid]] kept Iraq from crisis. Later, limited amounts of income from the United Nations [[oil-for-food]] program started flowing into Iraq. On December 9, 1996, the United Nations allowed Baghdad to begin selling limited amounts of oil for food and [[medicine]]. U.S. officials continued to accuse Saddam of violating the Gulf War's cease-fire agreement, by developing [[weapons of mass destruction]] and other banned weapons, and by violating the UN-imposed sanctions and "no-fly zones." Isolated military strikes by U.S. and British forces continued on Iraq, the largest being [[Operation Desert Fox]] in 1998. After two years of intermittent activity, U.S. and British warplanes struck harder at sites near Baghdad in [[February]] 2001. Saddam's base of Tikriti tribesmen, family members, and other supporters was divided after the war and in the following years. This added to the regime's increasingly [[repressive]] and [[wikt:arbitrary|arbitrary]] nature. Domestic repression inside Iraq grew worse, and Saddam's sons, [[Uday Hussein]] and [[Qusay Hussein]], became increasingly powerful and carried out a private reign of [[Terrorism|terror]]. They likely had a leading hand when, in August 1995, two of Saddam Hussein's sons-in-law, who held high positions in the Iraqi military, defected to [[Jordan]]. Both were killed after returning to Iraq the following February. == 2003 invasion of Iraq == [[File:SaddamStatue.jpg|thumb|Statue of Saddam Hussein Toppled]] {{seealso|2003 invasion of Iraq}} In 2003, the United States [[2003 invasion of Iraq|led an invasion of Iraq]]. The main reason for the invasion was President [[George W. Bush|George W. Bush's]] claim that Saddam had weapons of mass destruction. Bush argued that this made Saddam a major threat to Western allies, such as oil-rich Saudi Arabia and Israel; to Western oil supplies from the Persian Gulf states; and to Middle East stability in general. The President before Bush, [[Bill Clinton]] (1993-2001), maintained [[sanctions]] and made occasional air strikes in the "[[Iraqi no-fly zones]]" or other restrictions, in the hope that Saddam would be overthrown by his many political enemies. However, this never happened. Things changed in the United States after the [[September 11, 2001 attacks]]. In his January 2002 [[State of the Union Address]] to the [[United States Congress]], President Bush said that Iraq, Iran, and [[North Korea]] were an "[[Axis of Evil|axis of evil]]."<ref>For further details see Globe and Mail Update, "Hussein does Baghdad walkabout" [https://www.theglobeandmail.com/servlet/story/RTGAM.20030404.wmain0404_5/BNPrint/International] {{Webarchive|url=https://web.archive.org/web/20090502011512/http://www.theglobeandmail.com/servlet/story/RTGAM.20030404.wmain0404_5/BNPrint/International|date=2009-05-02}} Apr. 4, 2003.</ref> Bush also argued that Iraq had supported [[Al-Qaeda]], the terrorists who attacked the United States on September 11. As the war was looming on February 24, 2003, Saddam Hussein talked with [[CBS News]] anchor [[Dan Rather]] for more than three hours—his first interview with a U.S. reporter in over ten years.<ref>Dan Rather's interview with Saddam Hussein leading up to the U.S.-led 2003 invasion of Iraq on March 20 can be read online (CBSNEWS.com) at [http://www.cbsnews.com/stories/2003/02/24/eveningnews/main541817.shtml] {{Webarchive|url=https://web.archive.org/web/20130924112935/http://www.cbsnews.com/stories/2003/02/24/eveningnews/main541817.shtml|date=2013-09-24}}.</ref> CBS aired the taped interview later that week. The Iraqi government and military collapsed within three weeks after the beginning of the [[2003 invasion of Iraq|U.S.-led invasion]] on March 20. The United States tried at least twice to kill Saddam with targeted air strikes, but both failed to hit their target. By the beginning of April, Coalition forces controlled much of Iraq. The resistance of the much-weakened Iraqi Army either crumbled or shifted to [[guerrilla warfare|guerrilla]] tactics, and it appeared that Saddam had lost control of Iraq. He was last seen in a [[video]] which showed him in the Baghdad [[suburb]]s surrounded by supporters. When Baghdad fell to the Coalition on April 9, Saddam was nowhere to be found. == Pursuit and capture == [[File:SaddamSpiderHole.jpg|thumb|right|300px|[[United States Military|American soldiers]] capture Saddam]] Even when Baghdad was taken over, and most of the fighting had stopped, people still did not know where Saddam was. For a few weeks, some people said they saw Saddam, and some videotapes of Saddam talking came out, but still nobody knows if they were true or not. Although Saddam was placed at the top of the "[[U.S. list of most-wanted Iraqis|most-wanted list]]," he could not be found, even when the other leaders of the Iraqi regime were arrested. His sons and political [[Heir apparent|heirs]], [[Uday Hussein|Uday]] and [[Qusai Hussein|Qusay]], were killed in July 2003 in a clash with U.S. forces after a tip from an Iraqi. On 14 December 2003, the [[Islamic Republic News Agency]] (IRNA) of Iran first reported that Saddam Hussein had been arrested. These reports were soon confirmed by other members of the [[Iraq Interim Governing Council|Governing Council]], by U.S. military sources, and by [[Prime Minister of the United Kingdom|British Prime Minister]] [[Tony Blair]]. [[File:Saddamcapture.jpg|thumb|left|Elderly Saddam after being captured 2004-2006 ?]] Shortly afterwards, the U.S. Civil Administrator in Iraq, [[Paul Bremer]], held a press conference in [[Baghdad]]. He officially announced Saddam's capture by saying, "Ladies and gentlemen, we got him!" United States soldiers found Saddam around 8:30 PM Iraqi time on 13 December, in what was called [[Operation Red Dawn]]. Saddam was hiding in an underground "[[spider hole]]" at a farmhouse in [[ad-Dawr]], near his hometown of [[Tikrit]]. The first photos taken of Saddam after the soldiers found him did not look like the photos taken when he was President of Iraq. He had grown long hair and a long beard. Later on, he shaved his beard to confirm his identity. [[DNA]] tests proved that he really was Saddam Hussein. People who talked with him after the soldiers found him said he was healthy, and wanted to talk to people and do what they told him to do. Paul Bremer said that Saddam would have a [[trial]], but that he did not know yet what kind of trial. == Trial == [[File:Saddam Hussein at trial, July 2004.JPEG|thumb|Saddam Speaks in Court during his Trial in 2004 shortly after being captured]] The [[Iraqi Special Tribunal]] was in charge of Saddam Hussein's trial, and the trials of some people that helped him to be President of Iraq. In November 2006, Saddam Hussein was found [[Guilt (law)|guilty]] of 148 [[murder]]s. Critics viewed the trial as a [[show trial]] that did not meet international standards on the [[right to a fair trial]]. [[Amnesty International]] stated that the trial was "unfair".<ref>{{Cite web|title=Iraq: Amnesty International condemns Iraqi Appeal Court verdict against Saddam Hussein and co-accused|url=https://www.amnesty.org/en/documents/mde14/044/2006/en/|website=www.amnesty.org}}</ref> [[Human Rights Watch]] judged that Saddam's execution "follows a flawed trial and marks a significant step away from the [[rule of law]] in Iraq."<ref>[https://hrw.org/english/docs/2006/12/30/iraq14950.htm Iraq: Saddam Hussein Put to Death], Human Rights Watch (30-12-2006).</ref> Several months before the trial took place, [[Salem Chalabi]], the former head of the [[Supreme Iraqi Criminal Tribunal|Iraq Special Tribunal]] (which was established to try Hussein), accused interim Iraqi Prime Minister [[Iyad Allawi]] of pushing for a hasty show trial and execution, stating: "Show trials followed by speedy executions may help the interim government politically in the short term but will be counterproductive for the development of democracy and the rule of law in Iraq in the long term."<ref>{{cite news|date=23 September 2004|title=Iraq PM 'seeks Saddam show trial'|publisher=BBC News|url=http://news.bbc.co.uk/1/hi/world/middle_east/3684052.stm|access-date=26 May 2010}}</ref> == Execution == {{main|Execution of Saddam Hussein}} He was [[Execution of Saddam Hussein|put to death]] in 2006. The hanging was recorded by officials of the government. A [[witness]] also secretly recorded the hanging with a [[cell phone]] camera, which included sound. The recording showed Saddam being calm as he was prepared for his final moments. Witnesses and executioners could be heard teasing Saddam as the rope was placed on his head, and he was put on the trap door. He was in the middle of a [[prayer]] when the trap door beneath him opened, and he was killed. Later, pictures and live [[video]] of Saddam's taunting and execution, and of his dead [[corpse]] were shown on many Internet sites. Saddam was buried in his hometown, Al-Awja, Iraq, the next day. Saddam was scheduled to die on Thursday night, 28 December 2006.<ref>{{cite web|title=Seattle Times|url=http://seattletimes.nwsource.com/html/nationworld/2003503159_saddam31.html}}</ref> Because of last-minute legal [[Appellate court|appeals]] in Iraq's Supreme Court of Appeals,<ref>Mariam Karouny and Ibon Villelabeitia (26 December 2006). "Iraq court upholds Saddam death sentence". The Washington Post. Reuters. [https://www.washingtonpost.com/wp-dyn/content/article/2006/12/26/AR2006122600297_pf.html]</ref> Saddam's execution came about 40 hours later. Saddam Hussein was [[execution|hanged]] on 30 December 2006 at 6:05 AM, Iraqi time. Saddam refused to wear a hood. He was pronounced dead at 6:10 AM Baghdad time. Baghdad had [[curfew]]s on those days, which required people to be off the streets by evening. == Personal Life == [[File:Saddam-family-Pre1995.jpg|thumb|Saddam and his Family {{Circa|Mid-Late 1980s}}]] Saddam had been [[Marriage|married]] three times. His first marriage was to his first [[cousin]], [[Sajida Talfah]]. She was the oldest daughter of Saddam's uncle, Khairallah Talfah. Together, Saddam and Sajida had two sons, [[Uday Saddam Hussein]] and [[Qusay Hussein]], and three daughters, Rana, Raghad and Hala. In early 1997, Sajida was put under [[house arrest]], along with her daughters Raghad and Rana, because they were suspected of being involved in an attempted [[assassination]] on Uday on 12 December 1996. General [[Adnan Khairallah Tuffah]], who was Sajida's brother and Saddam Hussein's boyhood friend, was allegedly [[Execution|executed]] because of his growing popularity. Saddam Hussein also married two other women. The second was Samira Shahbandar, whom he married in 1986 after forcing her husband to [[divorce]] her. She was said to have been his favorite wife. His third wife was Nidal al-Hamdani, the general manager of the Solar Energy Research Center in the Council of Scientific Research, whose husband apparently was also persuaded to divorce his wife. In August 1995, Rana; her husband, [[Hussein Kamel]] al Majid; Raghad; her husband, [[Saddam Kamel Majid]]; and their children defected to [[Jordan]]. They returned to Iraq when they were promised that Saddam Hussein would [[pardon]] them. Within three days of their return in February 1996, both of the Majid brothers were executed. Saddam's daughter Hala is married to [[Jamal Mustafa]], the deputy head of Iraq's Tribal Affairs Office. Neither has been known to be involved in [[politics]]. Another cousin was [[Ali Hassan al-Majid]], also known in the United States as "Chemical Ali," who was accused of ordering the use of poison gas in 1988.<ref>For coverage of the postwar [[CNN]] and [[Al-Arabiya]] interviews with Saddam's daughters, see [https://www.usatoday.com/news/world/iraq/2003-08-01-saddams-daughters_x.htm].</ref> In 2000, a romance book titled ''Zabiba and the King'' was released and later became a [[Bestseller|bestselling]] book.<ref name=Zabiba>{{cite web|url=https://www.military.com/history/11-crazy-facts-saddam-hussein|title=11 Crazy Facts About Saddam Hussein|publisher=Military.com|accessdate=August 10, 2022}}</ref> It was later revealed that Hussein was the author of the book.<ref name=bengio>{{cite journal |url=http://www.meforum.org/article/125 |title=Saddam Husayn's Novel of Fear |first=Ofra |last=Bengio |journal=[[Middle East Quarterly]] |volume=9 |number=1 |year=2002 |access-date=2022-08-10 |archive-date=2008-10-12 |archive-url=https://web.archive.org/web/20081012015130/http://www.meforum.org/article/125 |url-status=dead }}</ref><ref name=cia>{{cite news |url=https://www.nytimes.com/2001/05/25/world/25IRAQ.html?ex=1162789200&en=7769bdd7845dad1c&ei=5070 |title=C.I.A. Sleuths Study a Novel for the Thinking of Saddam |first=Elaine |last=Sciolino |work=The New York Times |date=May 24, 2001}}</ref> == Notes == {{reflist|2}} <small> <sup>2</sup> See PBS Frontline (2003), "The survival of Saddam: secrets of his life and leadership: interview with Saïd K. Aburish" at [https://www.pbs.org/wgbh/pages/frontline/shows/saddam/interviews/aburish2.html].<br /> <sup>3</sup> From Elisabeth Bumiller's interview of Jerrold M. Post, the founder of the Center for the Analysis of Personality and Political Behavior at the [[Central Intelligence Agency|CIA]] in the ''New York Times' (May 15, 2004) on the importance of events during Saddam Hussein's youth. It can be read online at [http://hnn.us/roundup/archives/11/2004/05/#5225] {{Webarchive|url=https://web.archive.org/web/20050930075739/http://hnn.us/roundup/archives/11/2004/05/#5225 |date=2005-09-30 }}.<br />'' <sup>8</sup> The full text of Bush's 2002 State of the Union address can be read online (BBC News) at [http://news.bbc.co.uk/1/hi/world/americas/1790537.stm].<br /> </small> == Other websites == {{wikiquote-en|Saddam Hussein}} * [http://www.gwu.edu/~nsarchiv/special/iraq/index.htm The Saddam Hussein Sourcebook] {{Webarchive|url=https://web.archive.org/web/20080513202524/http://www.gwu.edu/~nsarchiv/special/iraq/index.htm |date=2008-05-13 }} - Provided by the ''[[National Security Archive]]''. {{DEFAULTSORT:Hussein, Saddam}} [[Category:1937 births]] [[Category:2006 deaths]] [[Category:Former dictators]] [[Category:Iraqi generals]] [[Category:Iraqi murderers]] [[Category:Iraqi Muslims]] [[Category:People executed by hanging]] [[Category:Presidents of Iraq]] [[Category:Prime Ministers of Iraq]] [[Category:People from Tikrit]] [[Category:Vice presidents of Iraq]] [[Category:War criminals]]
Saddam Hussein Abd al-Majid al-Tikriti (28 April 1937 – 30 December 2006) was an Iraqi politician and revolutionary who served as the fifth president of Iraq from 1979 to 2003. He also served as prime minister of Iraq from 1979 to 1991 and later from 1994 to 2003. He was a leading member of the revolutionary Arab Socialist Ba'ath Party, and later, the Baghdad-based Ba'ath Party and its regional organization, the Iraqi Ba'ath Party, which espoused Ba'athism, a mix of Arab nationalism and Arab socialism. Saddam was born in the village of Al-Awja, near Tikrit in northern Iraq, to a peasant Sunni Arab family. He joined the Arab Socialist Ba'ath Party in 1957, and the Baghdad-based Ba'ath Party, and its regional organization, the Iraqi Ba'ath Party. He played a key role in the 17 July Revolution and was appointed vice president by Ahmed Hassan al-Bakr. During his time as vice president, Saddam nationalised the Iraq Petroleum Company, diversifying the Iraqi economy. He presided over the Second Iraqi–Kurdish War (1974–1975). Following al-Bakr's resignation in 1979, Saddam formally took power, although he had already been the de facto head of Iraq for several years. Positions of power in the country were mostly filled with Sunni Arabs, a minority that made up only a fifth of the population. Upon taking office, Saddam instituted the Ba'ath Party Purge. Saddam ordered the 1980 invasion of Iran in effort to purportedly capture Iran's Arab-majority Khuzistan province and thwart Iranian attempts to export their own 1979 revolution. The Iran–Iraq War ended after nearly eight years in a ceasefire after a gruelling stalemate that took somewhere around a total of a million lives and economic losses of $561 billion in Iraq. Later, Saddam accused its ally Kuwait of slant-drilling Iraqi oil fields and occupied Kuwait, initiating the Gulf War (1990–1991). Iraq was defeated by a multinational coalition led by the United States. The United Nations subsequently placed sanctions against Iraq. He suppressed the 1991 Iraqi uprisings of the Kurds and Shia Muslims, which sought to gain independence or overthrow the government. Saddam adopted an anti-American stance and established the Faith Campaign, pursuing an Islamist agenda in Iraq. Saddam's rule was marked by numerous human rights abuses, including an estimated 250,000 arbitrary deaths and disappearances. In 2003, the United States and its allies invaded Iraq, falsely accusing Saddam of developing weapons of mass destruction and of having ties with al-Qaeda. The Ba'ath Party was banned and Saddam went into hiding. After his capture on 13 December 2003, his trial took place under the Iraqi Interim Government. On 5 November 2006, Saddam was convicted by the Iraqi High Tribunal of crimes against humanity related to the 1982 killing of 148 Iraqi Shi'a and sentenced to death by hanging. He was executed on 30 December 2006. Saddam has been accused of running a repressive authoritarian government, which several analysts have described as totalitarian, although the applicability of that label has been contested. Saddam Hussein was born in al-Awja, a small village near Tikrit to Hussein Abd Al-Majid and Subha Tulfah Al-Mussallat. They both were from the Albu Nasir tribe, a tribe that had descended from Ahmed Bin Hussein 'Nasiruddin' who was a descendant of Imam Hussein Bin Ali. The Albu Nasir tribe had settled in Tikrit after migrating from Yemen. Saddam's brother and father both died of cancer before his birth. These deaths made Saddam's mother, Subha Tulfah al-Mussallat, so depressed that she unsuccessfully attempted to abort her pregnancy and commit suicide. Subha "would have nothing to do with him," and Saddam would eventually be taken in by an uncle. His mother remarried, and Saddam gained three half-brothers through this marriage. His stepfather, Ibrahim al-Hassan, treated Saddam harshly after his return, and (according to a psychological profile created by the CIA) beat him regularly, sometimes to wake him up. At around the age of 10, Saddam fled the family and returned to live in Baghdad with his uncle Khairallah Talfah, who became a fatherly figure to Saddam. Talfah, the father of Saddam's future wife, was a devout Sunni Muslim and a veteran of the 1941 Anglo-Iraqi War between Iraqi nationalists and the United Kingdom, which remained a major colonial power in the region. Talfah was later appointed the mayor of Baghdad during Saddam's time in power, until his notorious corruption compelled Saddam to force him out of office. Later in his life, relatives from his native city became some of his closest advisors and supporters. Under the guidance of his uncle, he attended a nationalistic high school in Baghdad. After secondary school, Saddam studied at an Iraqi law school for three years, dropping out in 1957 at the age of 20 to join the revolutionary pan-Arab Ba'ath Party, of which his uncle was a supporter. During this time, Saddam apparently supported himself as a secondary school teacher. Ba'athist ideology originated in Syria and the Ba'ath Party had a large following in Syria at the time, but in 1955 there were fewer than 300 Ba'ath Party members in Iraq and it is believed that Saddam's primary reason for joining the party as opposed to the more established Iraqi nationalist parties was his familial connection to Ahmed Hassan al-Bakr and other leading Ba'athists through his uncle. Revolutionary sentiment was characteristic of the era in Iraq and throughout the Middle East. In Iraq progressives and socialists assailed traditional political elites (colonial-era bureaucrats and landowners, wealthy merchants and tribal chiefs, and monarchists). Moreover, the pan-Arab nationalism of Gamal Abdel Nasser in Egypt profoundly influenced young Ba'athists like Saddam. The rise of Nasser foreshadowed a wave of revolutions throughout the Middle East in the 1950s and 1960s, with the collapse of the monarchies of Iraq, Egypt, and Libya. Nasser inspired nationalists throughout the Middle East by fighting the British and the French during the Suez Crisis of 1956, modernizing Egypt, and uniting the Arab world politically. His father-in-law, Khairallah Talfah, was reported to have served five years in prison for his role in fighting against Great Britain in the 1941 Iraqi coup d'état and Anglo-Iraqi War, and often mentored and told tales of his exploits to the young Saddam. In 1958, a year after Saddam had joined the Ba'ath party, army officers led by General Abd al-Karim Qasim overthrew Faisal II of Iraq in the 14 July Revolution. The Ba'ath Party was originally represented in Qasim's cabinet; however, Qasim—reluctant to join Nasser's newly formed union between Egypt and Syria—sided with various groups within Iraq (notably the social democrats and the Iraqi Communist Party) that told him such an action would be dangerous. Instead, Qasim adopted a wataniyah policy of "Iraq First". To strengthen his own position within the government, Qasim also had an alliance with the Iraqi Communist Party, which was opposed to the notion of pan-Arabism. His policies angered several pan-Arab organisations, including the Ba'ath Party, which later began plotting to assassinate Qasim at Al-Rashid Street on 7 October 1959 and take power. Saddam was recruited to the assassination conspiracy by its ring-leader, Abdul Karim al-Shaikhly, after one of the would-be assassins left. During the ambush, Saddam (who was only supposed to provide cover) began shooting prematurely, which disorganised the whole operation. Qasim's chauffeur was killed and Qasim was hit in the arm and shoulder. The assassins thought they had killed Qasim and quickly retreated to their headquarters, but Qasim survived. Saddam himself is not believed to have received any training outside of Iraq, as he was a late addition to the assassination team. Richard Sale of United Press International (UPI), citing former U.S. diplomat and intelligence officials, Adel Darwish, and other experts, reported that the unsuccessful assassination attempt on Qasim was a collaboration between the United States Central Intelligence Agency (CIA) and Egyptian intelligence. Pertinent contemporary records relating to CIA operations in Iraq have remained classified or heavily redacted, thus "allow[ing] for plausible deniability." It is generally accepted that Egypt, in some capacity, was involved in the assassination attempt, and that "[t]he United States was working with Nasser on some level." Sale and Darwish's account has been disputed by historian Bryan R. Gibson who concludes that available U.S. declassified documents show that "while the United States was aware of several plots against Qasim, it had still adhered to [a] nonintervention policy." On the other hand, historian Kenneth Osgood writes that "the circumstantial evidence is such that the possibility of US–UAR collaboration with Ba'ath Party activists cannot be ruled out," concluding that "[w]hatever the validity of [Sale's] charges, at the very least currently declassified documents reveal that US officials were actively considering various plots against Qasim and that the CIA was building up assets for covert operations in Iraq." At the time of the attack, the Ba'ath Party had fewer than 1,000 members, however the failed assassination attempt led to widespread exposure for Saddam and the Ba'ath within Iraq, where both had previously languished in obscurity, and later became a crucial part of Saddam's public image during his tenure as president of Iraq. Kanan Makiya recounts: The man and the myth merge in this episode. His biography—and Iraqi television, which stages the story ad nauseam—tells of his familiarity with guns from the age of ten; his fearlessness and loyalty to the party during the 1959 operation; his bravery in saving his comrades by commandeering a car at gunpoint; the bullet that was gouged out of his flesh under his direction in hiding; the iron discipline that led him to draw a gun on weaker comrades who would have dropped off a seriously wounded member of the hit team at a hospital; the calculating shrewdness that helped him save himself minutes before the police broke in leaving his wounded comrades behind; and finally the long trek of a wounded man from house to house, city to town, across the desert to refuge in Syria. Michel Aflaq, the leader of the Ba'athist movement, organized the expulsion of leading Iraqi Ba'athist members, such as Fuad al-Rikabi, on the grounds that the party should not have initiated the attempt on Qasim's life. At the same time, Aflaq secured seats in the Iraqi Ba'ath leadership for his supporters, one of them being Saddam. The assassins, including Saddam, all eventually escaped to Cairo, Egypt "where they enjoyed Nasser's protection for the remainder of Qasim's tenure in power." Saddam initially escaped to Syria and then to Egypt itself in February 1960, and he continued to live there until 1963, graduating from high school in 1961 and unsuccessfully pursuing a law degree at Cairo Law School (1962–1963). It is possible that Saddam visited the U.S. embassy in Cairo during his exile, and some evidence suggests that he was "in frequent contact with US officials and intelligence agents." A former high-ranking U.S. official told historians Marion Farouk–Sluglett and Peter Sluglett that Iraqi Ba'athists, including Saddam, "had made contact with the American authorities in the late 1950s and early 1960s." Army officers with ties to the Ba'ath Party overthrew Qasim in the Ramadan Revolution coup of February 1963; long suspected to be supported by the CIA, however pertinent contemporary documents relating to the CIA's operations in Iraq have remained classified by the U.S. government, although the Ba'athists are documented to have maintained supportive relationships with U.S. officials before, during, and after the coup. Ba'athist leaders were appointed to the cabinet and Abdul Salam Arif became president. Arif dismissed and arrested the Ba'athist leaders later that year in the November 1963 Iraqi coup d'état. Being exiled in Egypt at the time, Saddam played no role in the 1963 coup or the brutal anti-communist purge that followed; although he returned to Iraq after the coup, becoming a key organizer within the Ba'ath Party's civilian wing upon his return. Unlike during the Qasim years, Saddam remained in Iraq following Arif's anti-Ba'athist purge in November 1963, and became involved in planning to assassinate Arif. In marked contrast to Qasim, Saddam knew that he faced no death penalty from Arif's government and knowingly accepted the risk of being arrested rather than fleeing to Syria again. Saddam was arrested in October 1964 and served approximately two years in prison before escaping in 1966. In 1966, Ahmed Hassan al-Bakr appointed him Deputy Secretary of the Regional Command. Saddam, who would prove to be a skilled organizer, revitalized the party. He was elected to the Regional Command, as the story goes, with help from Michel Aflaq—the founder of Ba'athist thought. In September 1966, Saddam initiated an extraordinary challenge to Syrian domination of the Ba'ath Party in response to the Marxist takeover of the Syrian Ba'ath earlier that year, resulting in the Party's formalized split into two separate factions. Saddam then created a Ba'athist security service, which he alone controlled. In July 1968, Saddam participated in a bloodless coup led by Ahmed Hassan al-Bakr that overthrew Abdul Rahman Arif, Salam Arif's brother and successor. While Saddam's role in the coup was not hugely significant (except in the official account), Saddam planned and carried out the subsequent purge of the non-Ba'athist faction led by Prime Minister Abd ar-Razzaq an-Naif, whose support had been essential to the coup's success. According to a semi-official biography, Saddam personally led Naif at gunpoint to the plane that escorted him out of Iraq. Arif was given refuge in London and then Istanbul. Al-Bakr was named president and Saddam was named his deputy, and deputy chairman of the Ba'athist Revolutionary Command Council. According to biographers, Saddam never forgot the tensions within the first Ba'athist government, which formed the basis for his measures to promote Ba'ath party unity as well as his resolve to maintain power and programs to ensure social stability. Although Saddam was al-Bakr's deputy, he was a strong behind-the-scenes party politician. Al-Bakr was the older and more prestigious of the two, but by 1969 Saddam clearly had become the moving force behind the party. In the late 1960s and early 1970s, as vice chairman of the Revolutionary Command Council, formally al-Bakr's second-in-command, Saddam built a reputation as a progressive, effective politician. At this time, Saddam moved up the ranks in the new government by aiding attempts to strengthen and unify the Ba'ath party and taking a leading role in addressing the country's major domestic problems and expanding the party's following. After the Ba'athists took power in 1968, Saddam focused on attaining stability in a nation riddled with profound tensions. Long before Saddam, Iraq had been split along social, ethnic, religious, and economic fault lines: Sunni versus Shi'ite, Arab versus Kurd, tribal chief versus urban merchant, nomad versus peasant. The desire for stable rule in a country rife with factionalism led Saddam to pursue both massive repression and the improvement of living standards. Saddam actively fostered the modernization of the Iraqi economy along with the creation of a strong security apparatus to prevent coups within the power structure and insurrections apart from it. Ever concerned with broadening his base of support among the diverse elements of Iraqi society and mobilizing mass support, he closely followed the administration of state welfare and development programs. At the center of this strategy was Iraq's oil. On 1 June 1972, Saddam oversaw the seizure of international oil interests, which, at the time, dominated the country's oil sector. A year later, world oil prices rose dramatically as a result of the 1973 energy crisis, and skyrocketing revenues enabled Saddam to expand his agenda. Within just a few years, Iraq was providing social services that were unprecedented among Middle Eastern countries. Saddam established and controlled the "National Campaign for the Eradication of Illiteracy" and the campaign for "Compulsory Free Education in Iraq", and largely under his auspices, the government established universal free schooling up to the highest education levels; hundreds of thousands learned to read in the years following the initiation of the program. The government also supported families of soldiers, granted free hospitalization to everyone, and gave subsidies to farmers. Iraq created one of the most modernized public-health systems in the Middle East, earning Saddam an award from the United Nations Educational, Scientific and Cultural Organization (UNESCO). With the help of increasing oil revenues, Saddam diversified the largely oil-based Iraqi economy. Saddam implemented a national infrastructure campaign that made great progress in building roads, promoting mining, and developing other industries. The campaign helped Iraq's energy industries. Electricity was brought to nearly every city in Iraq, and many outlying areas. Before the 1970s, most of Iraq's people lived in the countryside and roughly two-thirds were peasants. This number would decrease quickly during the 1970s as global oil prices helped revenues to rise from less than a half billion dollars to tens of billions of dollars and the country invested into industrial expansion. He nationalised independent banks, eventually leaving the banking system insolvent due to inflation and bad loans. The oil revenue benefited Saddam politically. According to The Economist, "Much as Adolf Hitler won early praise for galvanizing German industry, ending mass unemployment and building autobahns, Saddam earned admiration abroad for his deeds. He had a good instinct for what the "Arab street" demanded, following the decline in Egyptian leadership brought about by the trauma of Israel's six-day victory in the 1967 war, the death of the pan-Arabist hero, Gamal Abdul Nasser, in 1970, and the "traitorous" drive by his successor, Anwar Sadat, to sue for peace with the Jewish state. Saddam's self-aggrandizing propaganda, with himself posing as the defender of Arabism against Jewish or Persian intruders, was heavy-handed, but consistent as a drumbeat. It helped, of course, that his mukhabarat (secret police) put dozens of Arab news editors, writers and artists on the payroll." In 1972, Saddam signed a 15-year Treaty of Friendship and Cooperation with the Soviet Union. According to historian Charles R. H. Tripp, the treaty upset "the US-sponsored security system established as part of the Cold War in the Middle East. It appeared that any enemy of the Baghdad regime was a potential ally of the United States." In response, the US covertly financed Kurdish rebels led by Mustafa Barzani during the Second Iraqi–Kurdish War; the Kurds were defeated in 1975, leading to the forcible relocation of hundreds of thousands of Kurdish civilians. Saddam focused on fostering loyalty to the Ba'athists in the rural areas. After nationalizing foreign oil interests, Saddam supervised the modernization of the countryside, mechanizing agriculture on a large scale, and distributing land to peasant farmers. The Ba'athists established farm cooperatives and the government also doubled expenditures for agricultural development in 1974–1975. Saddam's welfare programs were part of a combination of "carrot and stick" tactics to enhance support for Saddam. The state-owned banks were put under his thumb. Lending was based on cronyism. In 1976, Saddam rose to the position of general in the Iraqi armed forces, and rapidly became the strongman of the government. As the ailing, elderly al-Bakr became unable to execute his duties, Saddam took on an increasingly prominent role as the face of the government both internally and externally. He soon became the architect of Iraq's foreign policy and represented the nation in all diplomatic situations. He was the de facto leader of Iraq some years before he formally came to power in 1979. He slowly began to consolidate his power over Iraq's government and the Ba'ath party. Relationships with fellow party members were carefully cultivated, and Saddam soon accumulated a powerful circle of support within the party. In 1979, al-Bakr started to make treaties with Syria, also under Ba'athist leadership, that would lead to unification between the two countries. Syrian President Hafiz al-Assad would become deputy leader in a union, and this would drive Saddam to obscurity. Saddam acted to secure his grip on power. He forced the ailing al-Bakr to resign on 16 July 1979, and formally assumed the presidency. Saddam convened an assembly of Ba'ath party leaders on 22 July 1979. During the assembly, which he ordered videotaped, Saddam claimed to have found a fifth column within the Ba'ath Party and directed Muhyi Abdel-Hussein to read out a confession and the names of 68 alleged co-conspirators. These members were labelled "disloyal" and were removed from the room one by one and taken into custody. After the list was read, Saddam congratulated those still seated in the room for their past and future loyalty. The 68 people arrested at the meeting were subsequently tried together and found guilty of treason; 22 were sentenced to execution. Other high-ranking members of the party formed the firing squad. By 1 August 1979, hundreds of high-ranking Ba'ath party members had been executed. "Fifty-seven boxes were recently returned to the Kurdish city of Sulaimaniya in Zeit trucks—large Russian military vehicles—by the Iraqi government authorities. Each box contained a dead child, eyes gouged out and ashen white, apparently drained of blood. The families were not given their children, were forced to accept a communal grave, and then had to pay 150 dinars for the burial." Iraqi society fissures along lines of language, religion and ethnicity. The Ba'ath Party, secular by nature, adopted Pan-Arab ideologies which in turn were problematic for significant parts of the population. Following the Iranian Revolution of 1979, Iraq faced the prospect of régime change from two Shi'ite factions (Dawa and SCIRI) which aspired to model Iraq on its neighbour Iran as a Shia theocracy. A separate threat to Iraq came from parts of the ethnic Kurdish population of northern Iraq which opposed being part of an Iraqi state and favored independence (an ongoing ideology which had preceded Ba'ath Party rule). To alleviate the threat of revolution, Saddam afforded certain benefits to the potentially hostile population. Membership in the Ba'ath Party remained open to all Iraqi citizens regardless of background, and repressive measures were taken against its opponents. "There is a feeling that at least three million Iraqis are watching the eleven million others." —"A European diplomat", quoted in The New York Times, April 3, 1984. The major instruments for accomplishing this control were the paramilitary and police organizations. Beginning in 1974, Taha Yassin Ramadan (himself a Kurdish Ba'athist), a close associate of Saddam, commanded the People's Army, which had responsibility for internal security. As the Ba'ath Party's paramilitary, the People's Army acted as a counterweight against any coup attempts by the regular armed forces. In addition to the People's Army, the Department of General Intelligence was the most notorious arm of the state-security system, feared for its use of torture and assassination. Barzan Ibrahim al-Tikriti, Saddam's younger half-brother, commanded Mukhabarat. Foreign observers believed that from 1982 this department operated both at home and abroad in its mission to seek out and eliminate Saddam's perceived opponents. Saddam was notable for using terror against his own people. The Economist described Saddam as "one of the last of the 20th century's great dictators, but not the least in terms of egotism, or cruelty, or morbid will to power." Saddam's regime brought about the deaths of at least 250,000 Iraqis and committed war crimes in Iran, Kuwait, and Saudi Arabia. Human Rights Watch and Amnesty International issued regular reports of widespread imprisonment and torture. Conversely, Saddam used Iraq's oil wealth to develop an extensive patronage system for the regime's supporters. Although Saddam is often described as a totalitarian leader, Joseph Sassoon notes that there are important differences between Saddam's repression and the totalitarianism practiced by Adolf Hitler and Joseph Stalin, particularly with regard to freedom of movement and freedom of religion. During his leadership, Saddam promoted the idea of dual nationalism which combines Iraqi nationalism and Arab nationalism, a much broader form of ethnic nationalism which supports Iraqi nationalism and links it to matters that impact Arabs as a whole. Saddam Hussein believed that the recognition of the ancient Mesopotamian origins and heritage of Iraqi Arabs was complementary to supporting Arab nationalism. In the course of his reign, the Ba'athist regime officially included the historic Kurdish Muslim leader Saladin as a patriotic symbol in Iraq, while Saddam called himself son of the Babylonian King Nebuchadnezzar and had stamped the bricks of ancient Babylon with his name and titles next to him. He also conducted two show elections, in 1995 and 2002. In the 1995 referendum, conducted on 15 October, he reportedly received 99.96% of the votes in a 99.47% turnout, getting only 3,052 negative votes among an electorate of 8.4 million. In the 15 October 2002 referendum he officially achieved 100% of approval votes and 100% turnout, as the electoral commission reported the next day that every one of the 11,445,638 eligible voters cast a "Yes" vote for the president. He erected statues around the country, which Iraqis toppled after his fall. Iraq's relations with the Arab world have been extremely varied. Relations between Iraq and Egypt violently ruptured in 1977, when the two nations broke relations with each other following Iraq's criticism of Egyptian President Anwar Sadat's peace initiatives with Israel. In 1978, Baghdad hosted an Arab League summit that condemned and ostracized Egypt for accepting the Camp David Accords. Egypt's strong material and diplomatic support for Iraq in the war with Iran led to warmer relations and numerous contacts between senior officials, despite the continued absence of ambassadorial-level representation. Since 1983, Iraq has repeatedly called for restoration of Egypt's "natural role" among Arab countries. Saddam developed a reputation for liking expensive goods, such as his diamond-coated Rolex wristwatch, and sent copies of them to his friends around the world. To his ally Kenneth Kaunda Saddam once sent a Boeing 747 full of presents—rugs, televisions, ornaments. Saddam enjoyed a close relationship with Russian intelligence agent Yevgeny Primakov that dated back to the 1960s; Primakov may have helped Saddam to stay in power in 1991. Saddam visited only two Western countries. The first visit took place in December 1974, when the Caudillo of Spain, Francisco Franco, invited him to Madrid and he visited Granada, Córdoba and Toledo. In September 1975 he met with Prime Minister Jacques Chirac in Paris, France. Several Iraqi leaders, Lebanese arms merchant Sarkis Soghanalian and others have claimed that Saddam financed Chirac's party. In 1991 Saddam threatened to expose those who had taken largesse from him: "From Mr. Chirac to Mr. Chevènement, politicians and economic leaders were in open competition to spend time with us and flatter us. We have now grasped the reality of the situation. If the trickery continues, we will be forced to unmask them, all of them, before the French public." France armed Saddam and it was Iraq's largest trade partner throughout Saddam's rule. Seized documents show how French officials and businessmen close to Chirac, including Charles Pasqua, his former interior minister, personally benefitted from the deals with Saddam. Because Saddam Hussein rarely left Iraq, Tariq Aziz, one of Saddam's aides, traveled abroad extensively and represented Iraq at many diplomatic meetings. In foreign affairs, Saddam sought to have Iraq play a leading role in the Middle East. Iraq signed an aid pact with the Soviet Union in 1972, and arms were sent along with several thousand advisers. The 1978 crackdown on Iraqi Communists and a shift of trade toward the West strained Iraqi relations with the Soviet Union; Iraq then took on a more Western orientation until the Gulf War in 1991. After the oil crisis of 1973, France had changed to a more pro-Arab policy and was accordingly rewarded by Saddam with closer ties. He made a state visit to France in 1975, cementing close ties with some French business and ruling political circles. In 1975 Saddam negotiated an accord with Iran that contained Iraqi concessions on border disputes. In return, Iran agreed to stop supporting opposition Kurds in Iraq. Saddam led Arab opposition to the Camp David Accords between Egypt and Israel (1979). Saddam initiated Iraq's nuclear enrichment project in the 1980s, with French assistance. The first Iraqi nuclear reactor was named by the French "Osirak". Osirak was destroyed on 7 June 1981 by an Israeli air strike (Operation Opera). Nearly from its founding as a modern state in 1920, Iraq has had to deal with Kurdish separatists in the northern part of the country. Saddam did negotiate an agreement in 1970 with separatist Kurdish leaders, giving them autonomy, but the agreement broke down. The result was brutal fighting between the government and Kurdish groups and Iraqi bombing of Kurdish villages in Iran, which caused Iraqi relations with Iran to deteriorate. After Saddam negotiated the 1975 treaty with Iran, the Shah withdrew support for the Kurds, who were defeated. In early 1979, Iran's Shah Mohammad Reza Pahlavi and Pahlavi dynasty were overthrown by the Islamic Revolution, thus giving way to an Islamic republic led by the Ayatollah Ruhollah Khomeini. The influence of revolutionary Shi'ite Islam grew apace in the region, particularly in countries with large Shi'ite populations, especially Iraq. Saddam feared that radical Islamic ideas—hostile to his secular rule—were rapidly spreading inside his country among the majority Shi'ite population. There had also been bitter enmity between Saddam and Khomeini since the 1970s. Khomeini, having been exiled from Iran in 1964, took up residence in Iraq, at the Shi'ite holy city of An Najaf. There he involved himself with Iraqi Shi'ites and developed a strong religious and political following against the Iranian Government, which Saddam tolerated. When Khomeini began to urge the Shi'ites there to overthrow Saddam and under pressure from the Shah, who had agreed to a rapprochement between Iraq and Iran in 1975, Saddam agreed to expel Khomeini in 1978 to France. Here, Khomeini gained media connections and collaborated with a much larger Iranian community, to his advantage. After Khomeini gained power, skirmishes between Iraq and revolutionary Iran occurred for ten months over the sovereignty of the disputed Shatt al-Arab waterway, which divides the two countries. During this period, Saddam Hussein publicly maintained that it was in Iraq's interest not to engage with Iran, and that it was in the interests of both nations to maintain peaceful relations. In a private meeting with Salah Omar al-Ali, Iraq's permanent ambassador to the United Nations, he revealed that he intended to invade and occupy a large part of Iran within months. Later (probably to appeal for support from the US and most Western nations), he would make toppling the Islamic government one of his intentions as well. Iraq invaded Iran, first attacking Mehrabad Airport of Tehran and then entering the oil-rich Iranian land of Khuzestan, which also has a sizable Arab minority, on 22 September 1980 and declared it a new province of Iraq. With the support of the Arab states, the US, and Europe, and heavily financed by the Arab states of the Persian Gulf, Saddam Hussein had become "the defender of the Arab world" against a revolutionary Iran. The only exception was the Soviet Union, who initially refused to supply Iraq on the basis of neutrality in the conflict, although in his memoirs, Mikhail Gorbachev claimed that Leonid Brezhnev refused to aid Saddam over infuriation of Saddam's treatment of Iraqi communists. Consequently, many viewed Iraq as "an agent of the civilized world." The blatant disregard of international law and violations of international borders were ignored. Instead Iraq received economic and military support from its allies, who overlooked Saddam's use of chemical warfare against the Kurds and the Iranians, in addition to Iraq's efforts to develop nuclear weapons. In the first days of the war, there was heavy ground fighting around strategic ports as Iraq launched an attack on Khuzestan. After making some initial gains, Iraq's troops began to suffer losses from human wave attacks by Iran. By 1982, Iraq was on the defensive and looking for ways to end the war. At this point, Saddam asked his ministers for candid advice. Health Minister Dr. Riyadh Ibrahim suggested that Saddam temporarily step down to promote peace negotiations. Initially, Saddam Hussein appeared to take in this opinion as part of his cabinet democracy. A few weeks later, Dr. Ibrahim was sacked when held responsible for a fatal incident in an Iraqi hospital where a patient died from intravenous administration of the wrong concentration of potassium supplement. Dr. Ibrahim was arrested a few days after his removal from the cabinet. He was known to have publicly declared before that arrest that he was "glad that he got away alive." Pieces of Ibrahim's dismembered body were delivered to his wife the next day. Iraq quickly found itself bogged down in one of the longest and most destructive wars of attrition of the 20th century. During the war, Iraq used chemical weapons against Iranian forces fighting on the southern front and Kurdish separatists who were attempting to open up a northern front in Iraq with the help of Iran. These chemical weapons were developed by Iraq from materials and technology supplied primarily by West German companies as well as using dual-use technology imported following the Reagan administration's lifting of export restrictions. The US government also supplied Iraq with "satellite photos showing Iranian deployments." In a US bid to open full diplomatic relations with Iraq, the country was removed from the US list of State Sponsors of Terrorism. Ostensibly, this was because of improvement in the regime's record, although former US Assistant Secretary of Defense Noel Koch later stated, "No one had any doubts about [the Iraqis'] continued involvement in terrorism ... The real reason was to help them succeed in the war against Iran." The Soviet Union, France, and China together accounted for over 90% of the value of Iraq's arms imports between 1980 and 1988. Saddam reached out to other Arab governments for cash and political support during the war, particularly after Iraq's oil industry severely suffered at the hands of the Iranian navy in the Persian Gulf. Iraq successfully gained some military and financial aid, as well as diplomatic and moral support, from the Soviet Union, China, France, and the US, which together feared the prospects of the expansion of revolutionary Iran's influence in the region. The Iranians, demanding that the international community should force Iraq to pay war reparations to Iran, refused any suggestions for a cease-fire. Despite several calls for a ceasefire by the United Nations Security Council, hostilities continued until 20 August 1988. On 16 March 1988, the Kurdish town of Halabja was attacked with a mix of mustard gas and nerve agents, killing 5,000 civilians, and maiming, disfiguring, or seriously debilitating 10,000 more. (see Halabja massacre) The attack occurred in conjunction with the 1988 al-Anfal Campaign designed to reassert central control of the mostly Kurdish population of areas of northern Iraq and defeat the Kurdish peshmerga rebel forces. Claims by Saddam's government and its international supporters that Iran had actually gassed the Kurds at Halabja have been thoroughly debunked. The bloody eight-year war ended in a stalemate. Encyclopædia Britannica states: "Estimates of total casualties range from 1,000,000 to twice that number. The number killed on both sides was perhaps 500,000, with Iran suffering the greatest losses." Neither side had achieved what they had originally desired and the borders were left nearly unchanged. The southern, oil rich and prosperous Khuzestan and Basra area (the main focus of the war, and the primary source of their economies) were almost completely destroyed and were left at the pre-1979 border, while Iran managed to make some small gains on its borders in the Northern Kurdish area. Both economies, previously healthy and expanding, were left in ruins. Saddam borrowed tens of billions of dollars from other Arab states and a few billions from elsewhere during the 1980s to fight Iran, mainly to prevent the expansion of Shi'a radicalism. This backfired on Iraq and the Arab states, for Khomeini was widely perceived as a hero for managing to defend Iran and maintain the war with little foreign support against the heavily backed Iraq and only managed to boost Islamic radicalism not only within the Arab states, but within Iraq itself, creating new tensions between the Sunni Ba'ath Party and the majority Shi'a population. Faced with rebuilding Iraq's infrastructure and internal resistance, Saddam desperately re-sought cash, this time for postwar reconstruction. The Al-Anfal Campaign was a genocidal campaign against the Kurdish people (and many others) in Kurdish regions of Iraq led by the government of Saddam Hussein and headed by Ali Hassan al-Majid. The campaign takes its name from Qur'anic chapter 8 (al-ʾanfāl), which was used as a code name by the former Iraqi Ba'athist administration for a series of attacks against the peshmerga rebels and the mostly Kurdish civilian population of rural Northern Iraq, conducted between 1986 and 1989 culminating in 1988. This campaign also targeted Shabaks and Yazidis, Assyrians, Turkoman people and Mandaeans and many villages belonging to these ethnic groups were also destroyed. Human Rights Watch estimates that between 50,000 and 100,000 people were killed. Some Kurdish sources put the number higher, estimating that 182,000 Kurds were killed. The end of the war with Iran served to deepen latent tensions between Iraq and its wealthy neighbor Kuwait. Saddam urged the Kuwaitis to waive the Iraqi debt accumulated in the war, some $30 billion, but they refused. Saddam pushed oil-exporting countries to raise oil prices by cutting back production; Kuwait refused, then led the opposition in OPEC to the cuts that Saddam had requested. Kuwait was pumping large amounts of oil, and thus keeping prices low, when Iraq needed to sell high-priced oil from its wells to pay off its huge debt. Saddam had consistently argued that Kuwait had historically been an integral part of Iraq, and had only come into being as a result of interference from the British government; echoing a belief that Iraqi nationalists had supported for the past fifty years. This belief was one of the few articles of faith uniting the political scene in a nation rife with sharp social, ethnic, religious, and ideological divides. The extent of Kuwaiti oil reserves also intensified tensions in the region. The oil reserves of Kuwait (with a population of 2 million next to Iraq's 25) were roughly equal to those of Iraq. Taken together, Iraq and Kuwait sat on top of some 20 percent of the world's known oil reserves; Saudi Arabia held another 25 percent. Saddam still had an experienced and well-equipped army, which he used to influence regional affairs. He later ordered troops to the Iraq–Kuwait border. As Iraq–Kuwait relations rapidly deteriorated, Saddam was receiving conflicting information about how the US would respond to the prospects of an invasion. For one, Washington had been taking measures to cultivate a constructive relationship with Iraq for roughly a decade. The Reagan administration gave Iraq roughly $4 billion in agricultural credits to bolster it against Iran. Saddam's Iraq became "the third-largest recipient of US assistance." Reacting to Western criticism in April 1990, Saddam threatened to destroy half of Israel with chemical weapons if it moved against Iraq. In May 1990 he criticized US support for Israel warning that "the US cannot maintain such a policy while professing friendship towards the Arabs." In July 1990 he threatened force against Kuwait and the UAE saying "The policies of some Arab rulers are American ... They are inspired by America to undermine Arab interests and security." The US sent warplanes and combat ships to the Persian Gulf in response to these threats. The US ambassador to Iraq, April Glaspie, met with Saddam in an emergency meeting on 25 July 1990, where the Iraqi leader attacked American policy with regards to Kuwait and the United Arab Emirates (UAE): Glaspie replied: Saddam stated that he would attempt last-ditch negotiations with the Kuwaitis but Iraq "would not accept death." US officials attempted to maintain a conciliatory line with Iraq, indicating that while George H. W. Bush and James Baker did not want force used, they would not take any position on the Iraq–Kuwait boundary dispute and did not want to become involved. Later, Iraq and Kuwait met for a final negotiation session, which failed. Saddam then sent his troops into Kuwait. As tensions between Washington and Saddam began to escalate, the Soviet Union, under Mikhail Gorbachev, strengthened its military relationship with the Iraqi leader, providing him military advisers, arms and aid. On 2 August 1990, Saddam invaded Kuwait, initially claiming assistance to "Kuwaiti revolutionaries", thus sparking an international crisis. On 4 August an Iraqi-backed "Provisional Government of Free Kuwait" was proclaimed, but a total lack of legitimacy and support for it led to an 8 August announcement of a "merger" of the two countries. On 28 August Kuwait formally became the 19th Governorate of Iraq. Just two years after the 1988 Iraq and Iran truce, "Saddam Hussein did what his Gulf patrons had earlier paid him to prevent." Having removed the threat of Iranian fundamentalism he "overran Kuwait and confronted his Gulf neighbors in the name of Arab nationalism and Islam." When later asked why he invaded Kuwait, Saddam first claimed that it was because Kuwait was rightfully Iraq's 19th province and then said "When I get something into my head I act. That's just the way I am." Saddam Hussein could pursue such military aggression with a "military machine paid for in large part by the tens of billions of dollars Kuwait and the Gulf states had poured into Iraq and the weapons and technology provided by the Soviet Union, Germany, and France." It was revealed during his 2003–2004 interrogation that in addition to economic disputes, an insulting exchange between the Kuwaiti emir Al Sabah and the Iraqi foreign minister – during which Saddam claimed that the emir stated his intention to turn "every Iraqi woman into a $10 prostitute" by ruining Iraq financially – was a decisive factor in triggering the Iraqi invasion. Shortly before he invaded Kuwait, he shipped 100 new Mercedes 200 Series cars to top editors in Egypt and Jordan. Two days before the first attacks, Saddam reportedly offered Egypt's Hosni Mubarak 50 million dollars in cash, "ostensibly for grain." US President George H. W. Bush responded cautiously for the first several days. On one hand, Kuwait, prior to this point, had been a virulent enemy of Israel and was the Persian Gulf monarchy that had the most friendly relations with the Soviets. On the other hand, Washington foreign policymakers, along with Middle East experts, military critics, and firms heavily invested in the region, were extremely concerned with stability in this region. The invasion immediately triggered fears that the world's price of oil, and therefore control of the world economy, was at stake. Britain profited heavily from billions of dollars of Kuwaiti investments and bank deposits. Bush was perhaps swayed while meeting with British prime minister Margaret Thatcher, who happened to be in the US at the time. Cooperation between the US and the Soviet Union made possible the passage of resolutions in the United Nations Security Council (UNSC) giving Iraq a deadline to leave Kuwait and approving the use of force if Saddam did not comply with the timetable. US officials feared Iraqi retaliation against oil-rich Saudi Arabia, since the 1940s a close ally of Washington, for the Saudis' opposition to the invasion of Kuwait. Accordingly, the US and a group of allies, including countries as diverse as Egypt, Syria and Czechoslovakia, deployed a massive number of troops along the Saudi border with Kuwait and Iraq in order to encircle the Iraqi army, the largest in the Middle East. Saddam's officers looted Kuwait, stripping even the marble from its palaces to move it to Saddam's own palace. During the period of negotiations and threats following the invasion, Saddam focused renewed attention on the Palestinian problem by promising to withdraw his forces from Kuwait if Israel would relinquish the occupied territories in the West Bank, the Golan Heights, and the Gaza Strip. Saddam's proposal further split the Arab world, pitting US- and Western-supported Arab states against the Palestinians. The allies ultimately rejected any linkage between the Kuwait crisis and Palestinian issues. Saddam ignored the Security Council deadline. Backed by the Security Council, a US-led coalition launched round-the-clock missile and aerial attacks on Iraq, beginning 16 January 1991. Israel, though subjected to attack by Iraqi missiles, refrained from retaliating in order not to provoke Arab states into leaving the coalition. A ground force consisting largely of US and British armored and infantry divisions ejected Saddam's army from Kuwait in February 1991 and occupied the southern portion of Iraq as far as the Euphrates. On 6 March 1991, Bush announced "What is at stake is more than one small country, it is a big idea—a new world order, where diverse nations are drawn together in common cause to achieve the universal aspirations of mankind: peace and security, freedom, and the rule of law." In the end, the Iraqi army proved unable to compete on the battlefield with the highly mobile coalition land forces and their overpowering air support. Some 175,000 Iraqis were taken prisoner and casualties were estimated at over 85,000. As part of the cease-fire agreement, Iraq agreed to scrap all poison gas and germ weapons and allow UN observers to inspect the sites. UN trade sanctions would remain in effect until Iraq complied with all terms. Saddam publicly claimed victory at the end of the war. Iraq's ethnic and religious divisions, together with the brutality of the conflict that this had engendered, laid the groundwork for postwar rebellions. In the aftermath of the fighting, social and ethnic unrest among Shi'ite Muslims, Kurds, and dissident military units threatened the stability of Saddam's government. Uprisings erupted in the Kurdish north and Shi'a southern and central parts of Iraq, but were ruthlessly repressed. Uprisings in 1991 led to the death of 100,000–180,000 people, mostly civilians. The US, which had urged Iraqis to rise up against Saddam, did nothing to assist the rebellions. The Iranians, despite the widespread Shi'ite rebellions, had no interest in provoking another war, while Turkey opposed any prospect of Kurdish independence, and the Saudis and other conservative Arab states feared an Iran-style Shi'ite revolution. Saddam, having survived the immediate crisis in the wake of defeat, was left firmly in control of Iraq, although the country never recovered either economically or militarily from the Gulf War. Saddam routinely cited his survival as "proof" that Iraq had in fact won the war against the US. This message earned Saddam a great deal of popularity in many sectors of the Arab world. John Esposito wrote, "Arabs and Muslims were pulled in two directions. That they rallied not so much to Saddam Hussein as to the bipolar nature of the confrontation (the West versus the Arab Muslim world) and the issues that Saddam proclaimed: Arab unity, self-sufficiency, and social justice." As a result, Saddam Hussein appealed to many people for the same reasons that attracted more and more followers to Islamic revivalism and also for the same reasons that fueled anti-Western feelings. One US Muslim observer noted: "People forgot about Saddam's record and concentrated on America ... Saddam Hussein might be wrong, but it is not America who should correct him." A shift was, therefore, clearly visible among many Islamic movements in the post war period "from an initial Islamic ideological rejection of Saddam Hussein, the secular persecutor of Islamic movements, and his invasion of Kuwait to a more populist Arab nationalist, anti-imperialist support for Saddam (or more precisely those issues he represented or championed) and the condemnation of foreign intervention and occupation." Some elements of Sharia law were re-introduced, and the phrase "Allahu Akbar" ("God is great"), in Saddam's handwriting, was added to the national flag. Saddam also commissioned the production of a "Blood Qur'an", written using 27 litres of his own blood, to thank God for saving him from various dangers and conspiracies. The United Nations-placed sanctions against Iraq for invading Kuwait were not lifted, blocking Iraqi oil exports. During the late 1990s, the UN considered relaxing the sanctions imposed because of the hardships suffered by ordinary Iraqis. Studies dispute the number of people who died in south and central Iraq during the years of the sanctions. On 9 December 1996, Saddam's government accepted the Oil-for-Food Programme that the UN had first offered in 1992. Relations between the US and Iraq remained tense following the Gulf War. The US launched a missile attack aimed at Iraq's intelligence headquarters in Baghdad 26 June 1993, citing evidence of repeated Iraqi violations of the "no fly zones" imposed after the Gulf War and for incursions into Kuwait. US officials continued to accuse Saddam of violating the terms of the Gulf War's cease fire, by developing weapons of mass destruction and other banned weaponry, and violating the UN-imposed sanctions. Also during the 1990s, President Bill Clinton maintained sanctions and ordered air strikes in the "Iraqi no-fly zones" (Operation Desert Fox), in the hope that Saddam would be overthrown by political enemies inside Iraq. Western charges of Iraqi resistance to UN access to suspected weapons were the pretext for crises between 1997 and 1998, culminating in intensive US and British missile strikes on Iraq, 16–19 December 1998. After two years of intermittent activity, US and British warplanes struck harder at sites near Baghdad in February 2001. Former CIA case officer Robert Baer reports that he "tried to assassinate" Saddam in 1995, amid "a decade-long effort to encourage a military coup in Iraq." Saddam continued involvement in politics abroad. Video tapes retrieved after show his intelligence chiefs meeting with Arab journalists, including a meeting with the former managing director of Al-Jazeera, Mohammed Jassem al-Ali, in 2000. In the video Saddam's son Uday advised al-Ali about hires in Al-Jazeera: "During your last visit here along with your colleagues we talked about a number of issues, and it does appear that you indeed were listening to what I was saying since changes took place and new faces came on board such as that lad, Mansour." He was later sacked by Al-Jazeera. In 2002, Austrian prosecutors investigated Saddam government's transactions with Fritz Edlinger that possibly violated Austrian money laundering and embargo regulations. Fritz Edlinger, president of the General Secretary of the Society for Austro-Arab relations (GÖAB) and a former member of Socialist International's Middle East Committee, was an outspoken supporter of Saddam Hussein. In 2005, an Austrian journalist revealed that Fritz Edlinger's GÖAB had received $100,000 from an Iraqi front company as well as donations from Austrian companies soliciting business in Iraq. In 2002, a resolution sponsored by the European Union was adopted by the Commission for Human Rights, which stated that there had been no improvement in the human rights crisis in Iraq. The statement condemned President Saddam Hussein's government for its "systematic, widespread and extremely grave violations of human rights and international humanitarian law." The resolution demanded that Iraq immediately put an end to its "summary and arbitrary executions ... the use of rape as a political tool and all enforced and involuntary disappearances." Many members of the international community, especially the US, continued to view Saddam as a bellicose tyrant who was a threat to the stability of the region. In his January 2002 state of the union address to Congress, President George W. Bush spoke of an "axis of evil" consisting of Iran, North Korea, and Iraq. Moreover, Bush announced that he would possibly take action to topple the Iraqi government, because of the threat of its weapons of mass destruction. Bush stated that "The Iraqi regime has plotted to develop anthrax, and nerve gas, and nuclear weapons for over a decade ... Iraq continues to flaunt its hostility toward America and to support terror." After the passing of UNSC Resolution 1441, which demanded that Iraq give "immediate, unconditional and active cooperation" with UN and IAEA inspections, Saddam allowed U.N. weapons inspectors led by Hans Blix to return to Iraq. During the renewed inspections beginning in November 2002, Blix found no stockpiles of WMD and noted the "proactive" but not always "immediate" Iraqi cooperation as called for by Resolution 1441. With war still looming on 24 February 2003, Saddam Hussein took part in an interview with CBS News reporter Dan Rather. Talking for more than three hours, he denied possessing any weapons of mass destruction, or any other weapons prohibited by UN guidelines. He also expressed a wish to have a live televised debate with George W. Bush, which was declined. It was his first interview with a US reporter in over a decade. CBS aired the taped interview later that week. Saddam Hussein later told an FBI interviewer that he once left open the possibility that Iraq possessed weapons of mass destruction in order to appear strong against Iran. The Iraqi government and military collapsed within three weeks of the beginning of the US-led 2003 invasion of Iraq on 20 March. By the beginning of April, US-led forces occupied much of Iraq. The resistance of the much-weakened Iraqi Army either crumbled or shifted to guerrilla tactics, and it appeared that Saddam had lost control of Iraq. He was last seen in a video which purported to show him in the Baghdad suburbs surrounded by supporters. When Baghdad fell to US-led forces on 9 April, marked symbolically by the toppling of his statue, Saddam was nowhere to be found. In April 2003, Saddam's whereabouts remained in question during the weeks following the fall of Baghdad and the conclusion of the major fighting of the war. Various sightings of Saddam were reported in the weeks following the war, but none were authenticated. At various times Saddam released audio tapes promoting popular resistance to his ousting. Saddam was placed at the top of the US list of most-wanted Iraqis. In July 2003, his sons Uday and Qusay and 14-year-old grandson Mustapha were killed in a three-hour gunfight with US forces. On 13 December 2003, in Operation Red Dawn, Saddam was captured by American forces after being found hiding in a hole in the ground near a farmhouse in ad-Dawr, near Tikrit. Following his capture, Saddam was transported to a US base near Tikrit, and later taken to the American base near Baghdad. Documents obtained and released by the National Security Archive detail FBI interviews and conversations with Saddam while he was in US custody. On 14 December, US administrator in Iraq Paul Bremer confirmed that Saddam Hussein had indeed been captured at a farmhouse in ad-Dawr near Tikrit. Bremer presented video footage of Saddam in custody. Saddam was shown with a full beard and hair longer than his familiar appearance. He was described by US officials as being in good health. Bremer reported plans to put Saddam on trial, but claimed that the details of such a trial had not yet been determined. Iraqis and Americans who spoke with Saddam after his capture generally reported that he remained self-assured, describing himself as a "firm, but just leader." British tabloid newspaper The Sun posted a picture of Saddam wearing white briefs on the front cover of a newspaper. Other photographs inside the paper show Saddam washing his trousers, shuffling, and sleeping. The US government stated that it considered the release of the pictures a violation of the Geneva Convention and that it would investigate the photographs. During this period Saddam was interrogated by FBI agent George Piro. The guards at the Baghdad detention facility called their prisoner "Vic", which stands for "Very Important Criminal" and let him plant a small garden near his cell. The nickname and the garden are among the details about the former Iraqi leader that emerged during a March 2008 tour of the Baghdad prison and cell where Saddam slept, bathed, kept a journal, and wrote poetry in the final days before his execution; he was concerned to ensure his legacy and how the history would be told. The tour was conducted by US Marine Maj. Gen. Doug Stone, overseer of detention operations for the US military in Iraq at the time. During his imprisonment he exercised and was allowed to have his personal garden; he also smoked his cigars and wrote his diary in the courtyard of his cell. On 30 June 2004, Saddam Hussein, held in custody by US forces at the US base "Camp Cropper", along with 11 other senior Ba'athist leaders, was handed over to the interim Iraqi government to stand trial for crimes against humanity and other offences. A few weeks later, he was charged by the Iraqi Special Tribunal with crimes committed against residents of Dujail in 1982, following a failed assassination attempt against him. Specific charges included the murder of 148 people, torture of women and children and the illegal arrest of 399 others. Among the many challenges of the trial were: On 5 November 2006, Saddam was found guilty of crimes against humanity and sentenced to death by hanging. Saddam's half-brother, Barzan Ibrahim, and Awad Hamed al-Bandar, head of Iraq's Revolutionary Court in 1982, were convicted of similar charges. The verdict and sentencing were both appealed, but subsequently affirmed by Iraq's Supreme Court of Appeals. Saddam was hanged on the first day of Eid ul-Adha, 30 December 2006, despite his wish to be executed by firing squad (which he argued was the lawful military capital punishment, citing his military position as the commander-in-chief of the Iraqi military). The execution was carried out at Camp Justice, an Iraqi army base in Kadhimiya, a neighborhood of northeast Baghdad. Saudi Arabia condemned Iraqi authorities for carrying on with the execution on a holy day. A presenter from the Al-Ikhbariya television station officially stated: "There is a feeling of surprise and disapproval that the verdict has been applied during the holy months and the first days of Eid al-Adha. Leaders of Islamic countries should show respect for this blessed occasion ... not demean it." Video of the execution was recorded on a mobile phone and his captors could be heard insulting Saddam. The video was leaked to electronic media and posted on the Internet within hours, becoming the subject of global controversy. It was later claimed by the head guard at the tomb where his remains lay that Saddam's body had been stabbed six times after the execution. Saddam's demeanor while being led to the gallows has been discussed by two witnesses, Iraqi Judge Munir Haddad and Iraqi national security adviser Mowaffak al-Rubaie. The accounts of the two witnesses are contradictory as Haddad describes Saddam as being strong in his final moments whereas al-Rubaie says Saddam was clearly afraid. Saddam's last words during the execution, "May God's blessings be upon Muhammad and his household. And may God hasten their appearance and curse their enemies." Then one of the crowd repeatedly said the name of the Iraqi Shiite cleric, Moqtada Al-Sadr. Saddam laughed and later said, "Do you consider this manhood?" The crowd shouted, "go to Hell." Saddam replied, "To the hell that is Iraq!?" Again, one of the crowd asked those who shouted to keep quiet for God. Saddam Hussein started recitation of final Muslim prayers, "I bear witness that there is no god but Allah and I testify that Muhammad is the Messenger of Allah." One of the crowd shouted, "The tyrant [dictator] has collapsed!" Saddam said, "May God's blessings be upon Muhammad and his household (family)". He recited the shahada one and a half times, as while he was about to say 'Muhammad' on the second shahada, the trapdoor opened, cutting him off mid-sentence. The rope broke his neck, killing him instantly. Not long before the execution, Saddam's lawyers released his last letter. A second unofficial video, apparently showing Saddam's body on a trolley, emerged several days later. It sparked speculation that the execution was carried out incorrectly as Saddam Hussein had a gaping hole in his neck. Saddam was buried at his birthplace of Al-Awja in Tikrit, Iraq, on 31 December 2006. He was buried 3 km (2 mi) from his sons Uday and Qusay Hussein. His tomb was reported to have been destroyed in March 2015. Before it was destroyed, a Sunni tribal group reportedly removed his body to a secret location, fearful of what might happen. In August 1995, Raghad and her husband, Hussein Kamel al-Majid, and Rana and her husband, Saddam Kamel al-Majid, defected to Jordan, taking their children with them. They returned to Iraq when they received assurances that Saddam would pardon them. Within three days of their return in February 1996, both of the Kamel brothers were attacked and killed in a gunfight with other clan members who considered them traitors. In August 2003, Saddam's daughters Raghad and Rana received sanctuary in Amman, Jordan. That month, they spoke with CNN and the Arab satellite station Al-Arabiya in Amman. When asked about her father, Raghad told CNN, "He was a very good father, loving, has a big heart." Asked if she wanted to give a message to her father, she said: "I love you and I miss you." Her sister Rana also remarked, "He had so many feelings and he was very tender with all of us." In 2003 (before the Iraq War), the CIA considered making a video of Saddam having sex with a male teenager to discredit Saddam with his supporters. In 1979, Jacob Yasso of Sacred Heart Chaldean Church in Detroit, Michigan congratulated Saddam Hussein on his presidency. In return, Yasso said that Saddam Hussein donated US$250,000 to his church, which is made up of at least 1,200 families of Middle Eastern descent. In 1980, Detroit Mayor Coleman Young allowed Yasso to present the key to the city of Detroit to Saddam Hussein. At the time, Saddam then asked Yasso, "I heard there was a debt on your church. How much is it?" After the inquiry, Saddam then donated another $200,000 to Chaldean Sacred Heart Church. Yasso said that Saddam made donations to Chaldean churches all over the world, and even went on record as saying "He's very kind to Christians." Saddam is sometimes accused of a repressive totalitarian government. Former US president Donald Trump praised Saddam for militant suppression and stability during his presidency in Iraq.
This '''list of [[oil field]]s''' includes major fields of the past and present. (Amounts in parentheses are estimated reserves, in [[barrel (unit)|barrels]].) * [[China]] ** [[Daqing Field]], [[Heilongjiang]] (40 billion) * [[Brazil]] ** [[Campos Basin]] * [[Iran]] ** [[Aghajari Field]] (14 billion) ** [[Ahwaz Field]] (17 billion) ** [[Gachsaran Field]] (15 billion) ** [[Marun Field]] (16 billion) * [[Iraq]] ** [[Kirkuk Field]] (16 billion) ** [[Rumaila Field]] (20 billion) * [[Kazakhstan]] ** [[Tengiz Field]] (15-26 billion) ** [[Kashagan Field]], [[Caspian]] (~50 billion) * [[Kuwait]] ** [[Burgan Field]] (66-72 billion) * [[Mexico]] ** [[Cantarell Field]] ** [[Chicontepec Field]] * [[Nigeria]] ** [[Niger Delta Field]] (34 billion) * [[Russia]] ** [[Samotlor Field]] (20 billion) ** [[Romashkino Field]] (16-17 billion) * [[Saudi Arabia]] ** [[Abqaiq Field]] (12 billion) ** [[Berri Field]] (12 billion) ** [[Faroozan-Marjan Field]] (10 billion) ** [[Ghawar Field]] (75-83 billion) ** [[Manifa Field]] (11 billion) ** [[Safaniya-Khafji Field]], [[Neutral Zone]] (30 billion) * [[United Arab Emirates]] ** [[Zakum Field]], [[Abu Dhabi]] (12 billion) * [[United States]] ** [[Prudhoe Bay]], [[Alaska]] (10 billion) ** [[Wilmington Oil Field]], [[California]] (0.3 billion) * [[Venezuela]] ** [[Bolivar Coastal Field]] {{DEFAULTSORT:oil fields, List of }} [[Category:Geography-related lists]] [[Category:Petroleum]]
This list of oil fields includes some major oil fields of the past and present. The list is incomplete; there are more than 25,000 oil and gas fields of all sizes in the world. However, 94% of known oil is concentrated in fewer than 1500 giant and major fields. Most of the world's largest oilfields are located in the Middle East, but there are also supergiant (>10 billion bbls) oilfields in Brazil, Mexico, Venezuela, Kazakhstan, and Russia. Amounts listed below, in billions of barrels, are the estimated ultimate recoverable petroleum resources (proved reserves plus cumulative production), given historical production and current extraction technology. Oil shale reserves (perhaps 3 trillion barrels (4.8×10 m)) and coal reserves, both of which can be converted to liquid petroleum, are not included in this chart. Other non-conventional liquid fuel sources are similarly excluded from this list.
{{Other pages about |anarchism |image=File:Anarchy-symbol.svg |caption=This symbol is often used by anarchists. The "[[A]]" represents anarchy, and the "[[O]]" is said to represent order.}} '''Anarchism''' is a radically revolutionary idea that says no one should be forced into any kind of [[hierarchy]]. For example, anarchism says that the [[government]] is harmful and not needed. However, that does not mean no form of order should exist. Municipalities and autonomy are familiar terms in Anarchism<ref name=definitions2>''Anarchism''. The Shorter Routledge Encyclopedia of Philosophy. 2005. P. 14 "Anarchism is the view that a society without the state, or government, is both possible and desirable."</ref><ref>Carl Slevin "anarchism" The Concise Oxford Dictionary of Politics. Ed. Iain McLean and Alistair McMillan. Oxford University Press, 2003.</ref> It also says that people's actions should never be forced by other people. Anarchism is also a {{c|philosophical movements and positions|philosophical movement}} and is called a [[libertarian]] type of [[socialism]]. It is not a political stance. The word "anarchism" [[etymology|is from]] the [[Greek language|Greek]] word "αναρχία", which means "no rulers" or "no government". But that does not have to to mean no rules at all. People often use the word "[[anarchy]]" to mean [[wikt:chaos|chaos]] and [[crime]]. But anarchists usually do not want this. They say anarchy is just a way of relations between people. They believe that, once put into place, these relations work on their own. Anarchists are usually opposed by the systems they wish to remove. == Principles == Individual freedom, voluntary association, and being against the [[state]] are important beliefs of anarchism. There are also big differences between anarchist philosophies on things like whether [[violence]] can be used to bring about anarchy; the best type of economy; the relationship between technology and hierarchy; the idea of equality; and the usefulness of some organization. The word "authority" is not clear, but anarchists are not against some types of [[authority]] (e.g. the authority of someone skilled in self-defence over someone that wants to learn self-defence), they are only against control by force. == Examples == {{Gallery | File:Inauguration_Day_Marchers.jpg | Anarchists usually use the [[black bloc]] strategy when [[protest]]ing or [[riot]]ing. | File:McKinleyAssassination.jpg | U.S. President [[William McKinley]] was killed by an anarchist in 1901.<ref>{{cite book |last=Miller |first=Scott |year=2011 |title=The President and the Assassin |publisher=Random House |location=New York | isbn=978-1-4000-6752-7 |url-access=registration |url=https://archive.org/details/presidentassassi00mill}}</ref> | File:PunkNotDead-graffiti.jpg | An example of anarchist [[graffiti]] related to [[punk music]]. The "A" in the circle is the anarchist symbol. | File:Minneapolis_unrest_May_28,_2020.jpg | Anarchists were part of the [[George Floyd protests]] in 2020 and 2021. }} == Related pages == * [[Anti-capitalism]] * [[Anti-fascism]] * [[Self-ownership]] == References == {{reflist}} {{Anarchism}} {{Political spectrum}} {{philosophy topics}} [[Category:Anarchism| ]]
Anarchism is a political philosophy and movement that is skeptical of all justifications for authority and seeks to abolish the institutions it claims maintain unnecessary coercion and hierarchy, typically including nation-states, and capitalism. Anarchism advocates for the replacement of the state with stateless societies and voluntary free associations. As a historically left-wing movement, this reading of anarchism is placed on the farthest left of the political spectrum, usually described as the libertarian wing of the socialist movement (libertarian socialism). Humans have lived in societies without formal hierarchies long before the establishment of states, realms, or empires. With the rise of organised hierarchical bodies, scepticism toward authority also rose. Although traces of anarchist ideas are found all throughout history, modern anarchism emerged from the Enlightenment. During the latter half of the 19th and the first decades of the 20th century, the anarchist movement flourished in most parts of the world and had a significant role in workers' struggles for emancipation. Various anarchist schools of thought formed during this period. Anarchists have taken part in several revolutions, most notably in the Paris Commune, the Russian Civil War and the Spanish Civil War, whose end marked the end of the classical era of anarchism. In the last decades of the 20th and into the 21st century, the anarchist movement has been resurgent once more, growing in popularity and influence within anti-capitalist, anti-war and anti-globalisation movements. Anarchists employ diverse approaches, which may be generally divided into revolutionary and evolutionary strategies; there is significant overlap between the two. Evolutionary methods try to simulate what an anarchist society might be like, but revolutionary tactics, which have historically taken a violent turn, aim to overthrow authority and the state. Many facets of human civilization have been influenced by anarchist theory, critique, and praxis. The etymological origin of anarchism is from the Ancient Greek anarkhia, meaning "without a ruler", composed of the prefix an- ("without") and the word arkhos ("leader" or "ruler"). The suffix -ism denotes the ideological current that favours anarchy. Anarchism appears in English from 1642 as anarchisme and anarchy from 1539; early English usages emphasised a sense of disorder. Various factions within the French Revolution labelled their opponents as anarchists, although few such accused shared many views with later anarchists. Many revolutionaries of the 19th century such as William Godwin (1756–1836) and Wilhelm Weitling (1808–1871) would contribute to the anarchist doctrines of the next generation but did not use anarchist or anarchism in describing themselves or their beliefs. The first political philosopher to call himself an anarchist (French: anarchiste) was Pierre-Joseph Proudhon (1809–1865), marking the formal birth of anarchism in the mid-19th century. Since the 1890s and beginning in France, libertarianism has often been used as a synonym for anarchism and its use as a synonym is still common outside the United States. Some usages of libertarianism refer to individualistic free-market philosophy only, and free-market anarchism in particular is termed libertarian anarchism. While the term libertarian has been largely synonymous with anarchism, its meaning has more recently been diluted by wider adoption from ideologically disparate groups, including both the New Left and libertarian Marxists, who do not associate themselves with authoritarian socialists or a vanguard party, and extreme cultural liberals, who are primarily concerned with civil liberties. Additionally, some anarchists use libertarian socialist to avoid anarchism's negative connotations and emphasise its connections with socialism. Anarchism is broadly used to describe the anti-authoritarian wing of the socialist movement. Anarchism is contrasted to socialist forms which are state-oriented or from above. Scholars of anarchism generally highlight anarchism's socialist credentials and criticise attempts at creating dichotomies between the two. Some scholars describe anarchism as having many influences from liberalism, and being both liberal and socialist but more so. Many scholars reject anarcho-capitalism as a misunderstanding of anarchist principles. While opposition to the state is central to anarchist thought, defining anarchism is not an easy task for scholars, as there is a lot of discussion among scholars and anarchists on the matter, and various currents perceive anarchism slightly differently. Major definitional elements include the will for a non-coercive society, the rejection of the state apparatus, the belief that human nature allows humans to exist in or progress toward such a non-coercive society, and a suggestion on how to act to pursue the ideal of anarchy. Before the creation of towns and cities, established authority did not exist. It was after the institution of authority that anarchistic ideas were espoused as a reaction. The most notable precursors to anarchism in the ancient world were in China and Greece. In China, philosophical anarchism (the discussion on the legitimacy of the state) was delineated by Taoist philosophers Zhuang Zhou and Laozi. Alongside Stoicism, Taoism has been said to have had "significant anticipations" of anarchism. Anarchic attitudes were also articulated by tragedians and philosophers in Greece. Aeschylus and Sophocles used the myth of Antigone to illustrate the conflict between laws imposed by the state and personal autonomy. Socrates questioned Athenian authorities constantly and insisted on the right of individual freedom of conscience. Cynics dismissed human law (nomos) and associated authorities while trying to live according to nature (physis). Stoics were supportive of a society based on unofficial and friendly relations among its citizens without the presence of a state. In medieval Europe, there was no anarchistic activity except some ascetic religious movements. These, and other Muslim movements, later gave birth to religious anarchism. In the Sasanian Empire, Mazdak called for an egalitarian society and the abolition of monarchy, only to be soon executed by Emperor Kavad I. In Basra, religious sects preached against the state. In Europe, various sects developed anti-state and libertarian tendencies. Renewed interest in antiquity during the Renaissance and in private judgment during the Reformation restored elements of anti-authoritarian secularism, particularly in France. Enlightenment challenges to intellectual authority (secular and religious) and the revolutions of the 1790s and 1848 all spurred the ideological development of what became the era of classical anarchism. During the French Revolution, partisan groups such as the Enragés and the sans-culottes saw a turning point in the fermentation of anti-state and federalist sentiments. The first anarchist currents developed throughout the 18th century as William Godwin espoused philosophical anarchism in England, morally delegitimising the state, Max Stirner's thinking paved the way to individualism and Pierre-Joseph Proudhon's theory of mutualism found fertile soil in France. By the late 1870s, various anarchist schools of thought had become well-defined and a wave of then unprecedented globalisation occurred from 1880 to 1914. This era of classical anarchism lasted until the end of the Spanish Civil War and is considered the golden age of anarchism. Drawing from mutualism, Mikhail Bakunin founded collectivist anarchism and entered the International Workingmen's Association, a class worker union later known as the First International that formed in 1864 to unite diverse revolutionary currents. The International became a significant political force, with Karl Marx being a leading figure and a member of its General Council. Bakunin's faction (the Jura Federation) and Proudhon's followers (the mutualists) opposed state socialism, advocating political abstentionism and small property holdings. After bitter disputes, the Bakuninists were expelled from the International by the Marxists at the 1872 Hague Congress. Anarchists were treated similarly in the Second International, being ultimately expelled in 1896. Bakunin famously predicted that if revolutionaries gained power by Marx's terms, they would end up the new tyrants of workers. In response to their expulsion from the First International, anarchists formed the St. Imier International. Under the influence of Peter Kropotkin, a Russian philosopher and scientist, anarcho-communism overlapped with collectivism. Anarcho-communists, who drew inspiration from the 1871 Paris Commune, advocated for free federation and for the distribution of goods according to one's needs. By the turn of the 20th century, anarchism had spread all over the world. It was a notable feature of the international syndicalist movement. In China, small groups of students imported the humanistic pro-science version of anarcho-communism. Tokyo was a hotspot for rebellious youth from East Asian countries, who moved to the Japanese capital to study. In Latin America, Argentina was a stronghold for anarcho-syndicalism, where it became the most prominent left-wing ideology. During this time, a minority of anarchists adopted tactics of revolutionary political violence, known as propaganda of the deed. The dismemberment of the French socialist movement into many groups and the execution and exile of many Communards to penal colonies following the suppression of the Paris Commune favoured individualist political expression and acts. Even though many anarchists distanced themselves from these terrorist acts, infamy came upon the movement and attempts were made to prevent anarchists immigrating to the US, including the Immigration Act of 1903, also called the Anarchist Exclusion Act. Illegalism was another strategy which some anarchists adopted during this period. Despite concerns, anarchists enthusiastically participated in the Russian Revolution in opposition to the White movement, especially in the Makhnovshchina; however, they met harsh suppression after the Bolshevik government had stabilised, including during the Kronstadt rebellion. Several anarchists from Petrograd and Moscow fled to Ukraine, before the Bolsheviks crushed the anarchist movement there too. With the anarchists being repressed in Russia, two new antithetical currents emerged, namely platformism and synthesis anarchism. The former sought to create a coherent group that would push for revolution while the latter were against anything that would resemble a political party. Seeing the victories of the Bolsheviks in the October Revolution and the resulting Russian Civil War, many workers and activists turned to communist parties which grew at the expense of anarchism and other socialist movements. In France and the United States, members of major syndicalist movements such as the General Confederation of Labour and the Industrial Workers of the World left their organisations and joined the Communist International. In the Spanish Civil War of 1936–39, anarchists and syndicalists (CNT and FAI) once again allied themselves with various currents of leftists. A long tradition of Spanish anarchism led to anarchists playing a pivotal role in the war, and particularly in the Spanish Revolution of 1936. In response to the army rebellion, an anarchist-inspired movement of peasants and workers, supported by armed militias, took control of Barcelona and of large areas of rural Spain, where they collectivised the land. The Soviet Union provided some limited assistance at the beginning of the war, but the result was a bitter fight between communists and other leftists in a series of events known as the May Days, as Joseph Stalin asserted Soviet control of the Republican government, ending in another defeat of anarchists at the hands of the communists. By the end of World War II, the anarchist movement had been severely weakened. The 1960s witnessed a revival of anarchism, likely caused by a perceived failure of Marxism–Leninism and tensions built by the Cold War. During this time, anarchism found a presence in other movements critical towards both capitalism and the state such as the anti-nuclear, environmental, and peace movements, the counterculture of the 1960s, and the New Left. It also saw a transition from its previous revolutionary nature to provocative anti-capitalist reformism. Anarchism became associated with punk subculture as exemplified by bands such as Crass and the Sex Pistols. The established feminist tendencies of anarcha-feminism returned with vigour during the second wave of feminism. Black anarchism began to take form at this time and influenced anarchism's move from a Eurocentric demographic. This coincided with its failure to gain traction in Northern Europe and its unprecedented height in Latin America. Around the turn of the 21st century, anarchism grew in popularity and influence within anti-capitalist, anti-war and anti-globalisation movements. Anarchists became known for their involvement in protests against the World Trade Organization (WTO), the Group of Eight and the World Economic Forum. During the protests, ad hoc leaderless anonymous cadres known as black blocs engaged in rioting, property destruction and violent confrontations with the police. Other organisational tactics pioneered at this time include affinity groups, security culture and the use of decentralised technologies such as the Internet. A significant event of this period was the confrontations at the 1999 Seattle WTO conference. Anarchist ideas have been influential in the development of the Zapatistas in Mexico and the Democratic Federation of Northern Syria, more commonly known as Rojava, a de facto autonomous region in northern Syria. While having revolutionary aspirations, many forms of anarchism are not confrontational nowadays. Instead, they are trying to build an alternative way of social organization, based on mutual interdependence and voluntary cooperation. Scholar Carissa Honeywell takes the example of Food not Bombs group of collectives, to highlight some features of how anarchist groups work: direct action, working together and in solidarity with those left behind. While doing so, they inform about the rising rates of world hunger suggest a policy to tackle hunger, ranging from de-funding the arms industry to addressing Monsanto seed-saving policies and patents, helping farmers and commodification of food and housing. Honeywell also emphasizes that contemporary anarchists are interested in the flourishing not only of humans, but non-humans and the environment as well. Honeywell argues that escalation of problems such as continuous wars and world poverty show that the current framework not only cannot solve those pressing problems for humanity, but are causal factors as well, resulting in the rejection of representative democracy and the state as a whole. Anarchist schools of thought have been generally grouped into two main historical traditions, social anarchism and individualist anarchism, owing to their different origins, values and evolution. The individualist current emphasises negative liberty in opposing restraints upon the free individual, while the social current emphasises positive liberty in aiming to achieve the free potential of society through equality and social ownership. In a chronological sense, anarchism can be segmented by the classical currents of the late 19th century and the post-classical currents (anarcha-feminism, green anarchism, and post-anarchism) developed thereafter. Beyond the specific factions of anarchist movements which constitute political anarchism lies philosophical anarchism which holds that the state lacks moral legitimacy, without necessarily accepting the imperative of revolution to eliminate it. A component especially of individualist anarchism, philosophical anarchism may tolerate the existence of a minimal state but claims that citizens have no moral obligation to obey government when it conflicts with individual autonomy. Anarchism pays significant attention to moral arguments since ethics have a central role in anarchist philosophy. Anarchism's emphasis on anti-capitalism, egalitarianism, and for the extension of community and individuality sets it apart from anarcho-capitalism and other types of economic libertarianism. Anarchism is usually placed on the far-left of the political spectrum. Much of its economics and legal philosophy reflect anti-authoritarian, anti-statist, libertarian, and radical interpretations of left-wing and socialist politics such as collectivism, communism, individualism, mutualism, and syndicalism, among other libertarian socialist economic theories. As anarchism does not offer a fixed body of doctrine from a single particular worldview, many anarchist types and traditions exist and varieties of anarchy diverge widely. One reaction against sectarianism within the anarchist milieu was anarchism without adjectives, a call for toleration and unity among anarchists first adopted by Fernando Tarrida del Mármol in 1889 in response to the bitter debates of anarchist theory at the time. Belief in political nihilism has been espoused by anarchists. Despite separation, the various anarchist schools of thought are not seen as distinct entities but rather as tendencies that intermingle and are connected through a set of uniform principles such as individual and local autonomy, mutual aid, network organisation, communal democracy, justified authority and decentralisation. Inceptive currents among classical anarchist currents were mutualism and individualism. They were followed by the major currents of social anarchism (collectivist, communist and syndicalist). They differ on organisational and economic aspects of their ideal society. Mutualism is an 18th-century economic theory that was developed into anarchist theory by Pierre-Joseph Proudhon. Its aims include "abolishing the state", reciprocity, free association, voluntary contract, federation and monetary reform of both credit and currency that would be regulated by a bank of the people. Mutualism has been retrospectively characterised as ideologically situated between individualist and collectivist forms of anarchism. In What Is Property? (1840), Proudhon first characterised his goal as a "third form of society, the synthesis of communism and property." Collectivist anarchism is a revolutionary socialist form of anarchism commonly associated with Mikhail Bakunin. Collectivist anarchists advocate collective ownership of the means of production which is theorised to be achieved through violent revolution and that workers be paid according to time worked, rather than goods being distributed according to need as in communism. Collectivist anarchism arose alongside Marxism but rejected the dictatorship of the proletariat despite the stated Marxist goal of a collectivist stateless society. Anarcho-communism is a theory of anarchism that advocates a communist society with common ownership of the means of production, held by a federal network of voluntary associations, with production and consumption based on the guiding principle "From each according to his ability, to each according to his need." Anarcho-communism developed from radical socialist currents after the French Revolution but was first formulated as such in the Italian section of the First International. It was later expanded upon in the theoretical work of Peter Kropotkin, whose specific style would go onto become the dominating view of anarchists by the late 19th century. Anarcho-syndicalism is a branch of anarchism that views labour syndicates as a potential force for revolutionary social change, replacing capitalism and the state with a new society democratically self-managed by workers. The basic principles of anarcho-syndicalism are direct action, workers' solidarity and workers' self-management. Individualist anarchism is a set of several traditions of thought within the anarchist movement that emphasise the individual and their will over any kinds of external determinants. Early influences on individualist forms of anarchism include William Godwin, Max Stirner, and Henry David Thoreau. Through many countries, individualist anarchism attracted a small yet diverse following of Bohemian artists and intellectuals as well as young anarchist outlaws in what became known as illegalism and individual reclamation. Anarchist principles undergird contemporary radical social movements of the left. Interest in the anarchist movement developed alongside momentum in the anti-globalisation movement, whose leading activist networks were anarchist in orientation. As the movement shaped 21st century radicalism, wider embrace of anarchist principles signaled a revival of interest. Anarchism has continued to generate many philosophies and movements, at times eclectic, drawing upon various sources and combining disparate concepts to create new philosophical approaches. The anti-capitalist tradition of classical anarchism has remained prominent within contemporary currents. Contemporary news coverage which emphasizes black bloc demonstrations has reinforced anarchism's historical association with chaos and violence. Its publicity has also led more scholars in fields such as anthropology and history to engage with the anarchist movement, although contemporary anarchism favours actions over academic theory. Various anarchist groups, tendencies, and schools of thought exist today, making it difficult to describe the contemporary anarchist movement. While theorists and activists have established "relatively stable constellations of anarchist principles", there is no consensus on which principles are core and commentators describe multiple anarchisms, rather than a singular anarchism, in which common principles are shared between schools of anarchism while each group prioritizes those principles differently. Gender equality can be a common principle, although it ranks as a higher priority to anarcha-feminists than anarcho-communists. Anarchists are generally committed against coercive authority in all forms, namely "all centralized and hierarchical forms of government (e.g., monarchy, representative democracy, state socialism, etc.), economic class systems (e.g., capitalism, Bolshevism, feudalism, slavery, etc.), autocratic religions (e.g., fundamentalist Islam, Roman Catholicism, etc.), patriarchy, heterosexism, white supremacy, and imperialism." Anarchist schools disagree on the methods by which these forms should be opposed. The principle of equal liberty is closer to anarchist political ethics in that it transcends both the liberal and socialist traditions. This entails that liberty and equality cannot be implemented within the state, resulting in the questioning of all forms of domination and hierarchy. Anarchists' tactics take various forms but in general serve two major goals, namely, to first oppose the Establishment and secondly to promote anarchist ethics and reflect an anarchist vision of society, illustrating the unity of means and ends. A broad categorisation can be made between aims to destroy oppressive states and institutions by revolutionary means on one hand and aims to change society through evolutionary means on the other. Evolutionary tactics embrace nonviolence, reject violence and take a gradual approach to anarchist aims, although there is significant overlap between the two. Anarchist tactics have shifted during the course of the last century. Anarchists during the early 20th century focused more on strikes and militancy while contemporary anarchists use a broader array of approaches. During the classical era, anarchists had a militant tendency. Not only did they confront state armed forces, as in Spain and Ukraine, but some of them also employed terrorism as propaganda of the deed. Assassination attempts were carried out against heads of state, some of which were successful. Anarchists also took part in revolutions. Many anarchists, especially the Galleanists, believed that these attempts would be the impetus for a revolution against capitalism and the state. Many of these attacks were done by individual assailants and the majority took place in the late 1870s, the early 1880s and the 1890s, with some still occurring in the early 1900s. Their decrease in prevalence was the result of further judicial power and targeting and cataloging by state institutions. Anarchist perspectives towards violence have always been controversial. Anarcho-pacifists advocate for non-violence means to achieve their stateless, nonviolent ends. Other anarchist groups advocate direct action, a tactic which can include acts of sabotage or terrorism. This attitude was quite prominent a century ago when seeing the state as a tyrant and some anarchists believing that they had every right to oppose its oppression by any means possible. Emma Goldman and Errico Malatesta, who were proponents of limited use of violence, stated that violence is merely a reaction to state violence as a necessary evil. Anarchists took an active role in strike actions, although they tended to be antipathetic to formal syndicalism, seeing it as reformist. They saw it as a part of the movement which sought to overthrow the state and capitalism. Anarchists also reinforced their propaganda within the arts, some of whom practiced naturism and nudism. Those anarchists also built communities which were based on friendship and were involved in the news media. In the current era, Italian anarchist Alfredo Bonanno, a proponent of insurrectionary anarchism, has reinstated the debate on violence by rejecting the nonviolence tactic adopted since the late 19th century by Kropotkin and other prominent anarchists afterwards. Both Bonanno and the French group The Invisible Committee advocate for small, informal affiliation groups, where each member is responsible for their own actions but works together to bring down oppression utilizing sabotage and other violent means against state, capitalism, and other enemies. Members of The Invisible Committee were arrested in 2008 on various charges, terrorism included. Overall, contemporary anarchists are much less violent and militant than their ideological ancestors. They mostly engage in confronting the police during demonstrations and riots, especially in countries such as Canada, Greece, and Mexico. Militant black bloc protest groups are known for clashing with the police; however, anarchists not only clash with state operators, they also engage in the struggle against fascists and racists, taking anti-fascist action and mobilizing to prevent hate rallies from happening. Anarchists commonly employ direct action. This can take the form of disrupting and protesting against unjust hierarchy, or the form of self-managing their lives through the creation of counter-institutions such as communes and non-hierarchical collectives. Decision-making is often handled in an anti-authoritarian way, with everyone having equal say in each decision, an approach known as horizontalism. Contemporary-era anarchists have been engaging with various grassroots movements that are more or less based on horizontalism, although not explicitly anarchist, respecting personal autonomy and participating in mass activism such as strikes and demonstrations. In contrast with the big-A anarchism of the classical era, the newly coined term small-a anarchism signals their tendency not to base their thoughts and actions on classical-era anarchism or to refer to classical anarchists such as Peter Kropotkin and Pierre-Joseph Proudhon to justify their opinions. Those anarchists would rather base their thought and praxis on their own experience which they will later theorize. The decision-making process of small anarchist affinity groups plays a significant tactical role. Anarchists have employed various methods in order to build a rough consensus among members of their group without the need of a leader or a leading group. One way is for an individual from the group to play the role of facilitator to help achieve a consensus without taking part in the discussion themselves or promoting a specific point. Minorities usually accept rough consensus, except when they feel the proposal contradicts anarchist ethics, goals and values. Anarchists usually form small groups (5–20 individuals) to enhance autonomy and friendships among their members. These kinds of groups more often than not interconnect with each other, forming larger networks. Anarchists still support and participate in strikes, especially wildcat strikes as these are leaderless strikes not organised centrally by a syndicate. As in the past, newspapers and journals are used, and anarchists have gone online in the World Wide Web to spread their message. Anarchists have found it easier to create websites because of distributional and other difficulties, hosting electronic libraries and other portals. Anarchists were also involved in developing various software that are available for free. The way these hacktivists work to develop and distribute resembles the anarchist ideals, especially when it comes to preserving users' privacy from state surveillance. Anarchists organize themselves to squat and reclaim public spaces. During important events such as protests and when spaces are being occupied, they are often called Temporary Autonomous Zones (TAZ), spaces where art, poetry, and surrealism are blended to display the anarchist ideal. As seen by anarchists, squatting is a way to regain urban space from the capitalist market, serving pragmatical needs and also being an exemplary direct action. Acquiring space enables anarchists to experiment with their ideas and build social bonds. Adding up these tactics while having in mind that not all anarchists share the same attitudes towards them, along with various forms of protesting at highly symbolic events, make up a carnivalesque atmosphere that is part of contemporary anarchist vividity. As anarchism is a philosophy that embodies many diverse attitudes, tendencies, and schools of thought; disagreement over questions of values, ideology, and tactics is common. Its diversity has led to widely different uses of identical terms among different anarchist traditions which has created a number of definitional concerns in anarchist theory. The compatibility of capitalism, nationalism, and religion with anarchism is widely disputed, and anarchism enjoys complex relationships with ideologies such as communism, collectivism, Marxism, and trade unionism. Anarchists may be motivated by humanism, divine authority, enlightened self-interest, veganism, or any number of alternative ethical doctrines. Phenomena such as civilisation, technology (e.g. within anarcho-primitivism), and the democratic process may be sharply criticised within some anarchist tendencies and simultaneously lauded in others. Objection to the state and its institutions is a sine qua non of anarchism. Anarchists consider the state as a tool of domination and believe it to be illegitimate regardless of its political tendencies. Instead of people being able to control the aspects of their life, major decisions are taken by a small elite. Authority ultimately rests solely on power, regardless of whether that power is open or transparent, as it still has the ability to coerce people. Another anarchist argument against states is that the people constituting a government, even the most altruistic among officials, will unavoidably seek to gain more power, leading to corruption. Anarchists consider the idea that the state is the collective will of the people to be an unachievable fiction due to the fact that the ruling class is distinct from the rest of society. Specific anarchist attitudes towards the state vary. Robert Paul Wolff believed that the tension between authority and autonomy would mean the state could never be legitimate. Bakunin saw the state as meaning "coercion, domination by means of coercion, camouflaged if possible but unceremonious and overt if need be." A. John Simmons and Leslie Green, who leaned toward philosophical anarchism, believed that the state could be legitimate if it is governed by consensus, although they saw this as highly unlikely. Beliefs on how to abolish the state also differ. As gender and sexuality carry along them dynamics of hierarchy, many anarchists address, analyse, and oppose the suppression of one's autonomy imposed by gender roles. Sexuality was not often discussed by classical anarchists but the few that did felt that an anarchist society would lead to sexuality naturally developing. Sexual violence was a concern for anarchists such as Benjamin Tucker, who opposed age of consent laws, believing they would benefit predatory men. A historical current that arose and flourished during 1890 and 1920 within anarchism was free love. In contemporary anarchism, this current survives as a tendency to support polyamory, relationship anarchy, and queer anarchism. Free love advocates were against marriage, which they saw as a way of men imposing authority over women, largely because marriage law greatly favoured the power of men. The notion of free love was much broader and included a critique of the established order that limited women's sexual freedom and pleasure. Those free love movements contributed to the establishment of communal houses, where large groups of travelers, anarchists and other activists slept in beds together. Free love had roots both in Europe and the United States; however, some anarchists struggled with the jealousy that arose from free love. Anarchist feminists were advocates of free love, against marriage, and pro-choice (utilising a contemporary term), and had a similar agenda. Anarchist and non-anarchist feminists differed on suffrage but were supportive of one another. During the second half of the 20th century, anarchism intermingled with the second wave of feminism, radicalising some currents of the feminist movement and being influenced as well. By the latest decades of the 20th century, anarchists and feminists were advocating for the rights and autonomy of women, gays, queers and other marginalised groups, with some feminist thinkers suggesting a fusion of the two currents. With the third wave of feminism, sexual identity and compulsory heterosexuality became a subject of study for anarchists, yielding a post-structuralist critique of sexual normality. Some anarchists distanced themselves from this line of thinking, suggesting that it leaned towards an individualism that was dropping the cause of social liberation. The interest of anarchists in education stretches back to the first emergence of classical anarchism. Anarchists consider proper education, one which sets the foundations of the future autonomy of the individual and the society, to be an act of mutual aid. Anarchist writers such as William Godwin (Political Justice) and Max Stirner ("The False Principle of Our Education") attacked both state education and private education as another means by which the ruling class replicate their privileges. In 1901, Catalan anarchist and free thinker Francisco Ferrer established the Escuela Moderna in Barcelona as an opposition to the established education system which was dictated largely by the Catholic Church. Ferrer's approach was secular, rejecting both state and church involvement in the educational process whilst giving pupils large amounts of autonomy in planning their work and attendance. Ferrer aimed to educate the working class and explicitly sought to foster class consciousness among students. The school closed after constant harassment by the state and Ferrer was later arrested. Nonetheless, his ideas formed the inspiration for a series of modern schools around the world. Christian anarchist Leo Tolstoy, who published the essay Education and Culture, also established a similar school with its founding principle being that "for education to be effective it had to be free." In a similar token, A. S. Neill founded what became the Summerhill School in 1921, also declaring being free from coercion. Anarchist education is based largely on the idea that a child's right to develop freely and without manipulation ought to be respected and that rationality would lead children to morally good conclusions; however, there has been little consensus among anarchist figures as to what constitutes manipulation. Ferrer believed that moral indoctrination was necessary and explicitly taught pupils that equality, liberty and social justice were not possible under capitalism, along with other critiques of government and nationalism. Late 20th century and contemporary anarchist writers (Paul Goodman, Herbert Read, and Colin Ward) intensified and expanded the anarchist critique of state education, largely focusing on the need for a system that focuses on children's creativity rather than on their ability to attain a career or participate in consumerism as part of a consumer society. Contemporary anarchists such as Ward claim that state education serves to perpetuate socioeconomic inequality. While few anarchist education institutions have survived to the modern-day, major tenets of anarchist schools, among them respect for child autonomy and relying on reasoning rather than indoctrination as a teaching method, have spread among mainstream educational institutions. Judith Suissa names three schools as explicitly anarchists' schools, namely the Free Skool Santa Cruz in the United States which is part of a wider American-Canadian network of schools, the Self-Managed Learning College in Brighton, England, and the Paideia School in Spain. The connection between anarchism and art was quite profound during the classical era of anarchism, especially among artistic currents that were developing during that era such as futurists, surrealists and others. In literature, anarchism was mostly associated with the New Apocalyptics and the neo-romanticism movement. In music, anarchism has been associated with music scenes such as punk. Anarchists such as Leo Tolstoy and Herbert Read stated that the border between the artist and the non-artist, what separates art from a daily act, is a construct produced by the alienation caused by capitalism and it prevents humans from living a joyful life. Other anarchists advocated for or used art as a means to achieve anarchist ends. In his book Breaking the Spell: A History of Anarchist Filmmakers, Videotape Guerrillas, and Digital Ninjas, Chris Robé claims that "anarchist-inflected practices have increasingly structured movement-based video activism." Throughout the 20th century, many prominent anarchists (Peter Kropotkin, Emma Goldman, Gustav Landauer and Camillo Berneri) and publications such as Anarchy wrote about matters pertaining to the arts. Three overlapping properties made art useful to anarchists. It could depict a critique of existing society and hierarchies, serve as a prefigurative tool to reflect the anarchist ideal society and even turn into a means of direct action such as in protests. As it appeals to both emotion and reason, art could appeal to the whole human and have a powerful effect. The 19th-century neo-impressionist movement had an ecological aesthetic and offered an example of an anarchist perception of the road towards socialism. In Les chataigniers a Osny by anarchist painter Camille Pissarro, the blending of aesthetic and social harmony is prefiguring an ideal anarchistic agrarian community. The most common critique of anarchism is the assertion that humans cannot self-govern and so a state is necessary for human survival. Philosopher Bertrand Russell supported this critique, stating that "[p]eace and war, tariffs, regulations of sanitary conditions and the sale of noxious drugs, the preservation of a just system of distribution: these, among others, are functions which could hardly be performed in a community in which there was no central government." Another common criticism of anarchism is that it fits a world of isolation in which only the small enough entities can be self-governing; a response would be that major anarchist thinkers advocated anarchist federalism. Another criticism of anarchism is the belief that it is inherently unstable: that an anarchist society would inevitably evolve back into a state. Thomas Hobbes and other early social contract theorists argued that the state emerges in response to natural anarchy in order to protect the people's interests and keep order. Philosopher Robert Nozick argued that a "night-watchman state", or minarchy, would emerge from anarchy through the process of an invisible hand, in which people would exercise their liberty and buy protection from protection agencies, evolving into a minimal state. Anarchists reject these criticisms by arguing that humans in a state of nature would not just be in a state of war. Anarcho-primitivists in particular argue that humans were better off in a state of nature in small tribes living close to the land, while anarchists in general argue that the negatives of state organization, such as hierarchies, monopolies and inequality, outweigh the benefits. Philosophy lecturer Andrew G. Fiala composed a list of common arguments against anarchism which includes critiques such as that anarchism is innately related to violence and destruction, not only in the pragmatic world, such as at protests, but in the world of ethics as well. Secondly, anarchism is evaluated as unfeasible or utopian since the state cannot be defeated practically. This line of arguments most often calls for political action within the system to reform it. The third argument is that anarchism is self-contradictory as a ruling theory that has no ruling theory. Anarchism also calls for collective action whilst endorsing the autonomy of the individual, hence no collective action can be taken. Lastly, Fiala mentions a critique towards philosophical anarchism of being ineffective (all talk and thoughts) and in the meantime capitalism and bourgeois class remains strong. Philosophical anarchism has met the criticism of members of academia following the release of pro-anarchist books such as A. John Simmons' Moral Principles and Political Obligations. Law professor William A. Edmundson authored an essay to argue against three major philosophical anarchist principles which he finds fallacious. Edmundson says that while the individual does not owe the state a duty of obedience, this does not imply that anarchism is the inevitable conclusion and the state is still morally legitimate. In The Problem of Political Authority, Michael Huemer defends philosophical anarchism, claiming that "political authority is a moral illusion." One of the earliest criticisms is that anarchism defies and fails to understand the biological inclination to authority. Joseph Raz states that the acceptance of authority implies the belief that following their instructions will afford more success. Raz believes that this argument is true in following both authorities' successful and mistaken instruction. Anarchists reject this criticism because challenging or disobeying authority does not entail the disappearance of its advantages by acknowledging authority such as doctors or lawyers as reliable, nor does it involve a complete surrender of independent judgment. Anarchist perception of human nature, rejection of the state, and commitment to social revolution has been criticised by academics as naive, overly simplistic, and unrealistic, respectively. Classical anarchism has been criticised for relying too heavily on the belief that the abolition of the state will lead to human cooperation prospering. Friedrich Engels, considered to be one of the principal founders of Marxism, criticised anarchism's anti-authoritarianism as inherently counter-revolutionary because in his view a revolution is by itself authoritarian. Academic John Molyneux writes in his book Anarchism: A Marxist Criticism that "anarchism cannot win", believing that it lacks the ability to properly implement its ideas. The Marxist criticism of anarchism is that it has a utopian character because all individuals should have anarchist views and values. According to the Marxist view, that a social idea would follow directly from this human ideal and out of the free will of every individual formed its essence. Marxists state that this contradiction was responsible for their inability to act. In the anarchist vision, the conflict between liberty and equality was resolved through coexistence and intertwining.
[[File:Global population cartogram.png|thumb|320px|Cartogram of the world population in 2018. Each square represents 500,000 people.]] This is a '''list of [[country|countries]] and [[dependent territory|dependencies]] by [[population]]'''. It includes [[sovereign states]], inhabited [[dependent territory|dependent territories]] and in some cases, constituent countries of [[sovereign states]], with the list being primarily based on the [[ISO]] standard [[ISO 3166-1]]. For instance the [[United Kingdom]] is considered a single entity. In addition this list also includes certain [[List of states with limited recognition|unrecognized states]] not found in [[ISO 3166]]. The number shows how many people live in each country. Countries with the most people are at the top of the list. Countries with the fewest people are at the bottom. ==Method== {{See also|List of countries, List of countries by area, List of countries by population density}} Figures based on our charts are estimates or projections taken from our national census, where national data are available. Where national data are not available, figures are based on the population estimates of the [[United Nations]]. Areas that form integral parts of sovereign states, such as the countries of the [[United Kingdom]] are counted as part of the sovereign state concerned. Not included in the list are organizations such as the [[NATO]]. Other territories that are marked with an asterisk (*) are territories that do not have permanent population, such as the [[British Indian Ocean Territory]] or [[United States Minor Outlying Islands]]. ==Sovereign states and dependencies by population== <!-- PLEASE READ BEFORE EDITING! WHEN YOU UPDATE A COUNTRY'S POPULATION - PLEASE ALSO UPDATE THE EXPRESSION NEXT TO IT (column "% of world population") WITH THE SAME NUMBER EXCLUDING COMMAS (the #expr: equation). --> Note: All 193 [[Members of the United Nations]] plus the two observer states are assigned a numbered rank. Dependent territories or constituent countries that are parts of sovereign states are not assigned a numbered rank. In addition, [[List of states with limited recognition|sovereign states with limited recognition]] are also included, but not assigned a number rank. {{sticky header}}{{sort under}}{{mw-datatable}}{{table alignment}} {| class="wikitable sortable sticky-header sort-under mw-datatable col2left col6left" {{right}} |+ List of countries and territories by total population ! ! Location ! Population ! style=width:2em | % of{{br}}world ! Date ! {{nowrap|Source (official or from}}{{br}}the [[United Nations]]) ! class="unsortable"| |- | {{ntsh|0}} – | '''{{noflag|World}}''' | {{formatnum:{{worldpop}}}} || {{center|100%}} || {{dts|{{CURRENTMONTHNAME}} {{CURRENTDAY}} 2024|abbr=on}} | UN projection<ref name="unpop">{{cite web |author=United Nations |title=World Population Prospects 2022 |url=https://population.un.org/wpp/ |access-date=10 Nov 2022 |website=population.un.org |archive-date=5 February 2019 |archive-url=https://web.archive.org/web/20190205234912/https://population.un.org/wpp/ |url-status=live }}</ref>|| |- | rowspan=2 | {{ntsh|1}} 1/2 {{br}} {{efn|According to UN estimates, India hadmore population than China by the end of April 2023.<ref>{{cite news |title=India to overtake China as world's most populous country in Apr 2023 |url=https://www.un.org/en/desa/india-overtake-china-world-most-populous-country-Apr-2023-united-nations-projects#:~:text=Smith%20Mehta%2Funsplash.-,India%20to%20overtake%20China%20as%20world's%20most%20populous%20country%20in,the%20world's%20most%20populous%20country |access-date=27 Apr 2023 |work=[[United Nations]] |date=24 Apr 2023}}</ref>}} | {{flag|China}} | {{n+p|1409670000|{{worldpop}}|sigfig=3|disp=table}} || {{dts|31 Dec 2023|abbr=on}} | Official estimate<ref>{{cite web |title=10. Total Population Declined and Urbanization Rate Continued to Grow (2023) |url=https://www.stats.gov.cn/english/PressRelease/202401/t20240117_1946605.html |publisher=[[National Bureau of Statistics of China]] (NBSC) |website=www.stats.gov.cn |date=2024-01-17 |access-date=2024-01-20}}</ref> || {{efn|Refers to [[Mainland China]]; does not include China's [[special administrative regions of China|special administrative regions]] of [[Hong Kong]] and [[Macau]].f}} |- | {{left}} {{flag|India}} | {{n+p|1404910000|{{worldpop}}|sigfig=3|disp=table}} || {{dts|1 July 2024|abbr=on}} | {{left}} Official projection<ref>{{cite web |title=Projected Total Population by Sex as on 1st July 2011-2036: India, States and Union Territories (pg.83) |url=https://main.mohfw.gov.in/sites/default/files/Population%20Projection%20Report%202011-2036%20-%20upload_compressed_0.pdf |publisher=[[Ministry of Health and Family Welfare]] |website=www.main.mohfw.gov.in |date=9 July 2020 |access-date=2 July 2024 |archive-date=4 June 2023 |archive-url=https://web.archive.org/web/20230604151511/https://main.mohfw.gov.in/sites/default/files/Population%20Projection%20Report%202011-2036%20-%20upload_compressed_0.pdf |url-status=live }}</ref> || {{right}} {{efn|Includes the population of the [[Kashmir conflict|India-administered]] [[union territory|union territories]] of [[Jammu and Kashmir (union territory)|Jammu and Kashmir]], and [[Ladakh]].}} |- | {{nts|3}} | {{flag|United States}} | {{n+p| 335893238|{{worldpop}}|sigfig=2|disp=table}} || {{dts|1 January 2024|abbr=on}} | Official estimate<ref>{{cite web |url = https://www.census.gov/newsroom/press-releases/2023/population-new-years-day.html |title = Happy New Year 2024! |date = 29 Dec 2023 |website = census.gov |access-date = 17 Nov 2023}}</ref>|| {{efn|Includes the [[list of states and territories of the United States|50 states]] and the [[Washington, D.C.|District of Columbia]], but does not inlucde the [[territories of the United States]].}} |- | {{nts|4}} | {{flag|Indonesia}} | {{n+p|281603800|{{worldpop}}|sigfig=2|disp=table}} || {{dts|1 Jul 2024|abbr=on}} | National annual projection<ref>{{cite web |url=https://dukcapil.kemendagri.go.id/page/read/data-kependudukan |publisher=[[Ministry of Home Affairs (Indonesia)]] |title=Indonesian Population June 2023 |access-date=2 July 2024 |archive-date=28 October 2023 |archive-url=https://web.archive.org/web/20231028112853/https://dukcapil.kemendagri.go.id/page/read/data-kependudukan |url-status=live}}</ref>|| |- | {{nts|5}} | {{flag|Pakistan}} | {{n+p|241499431|{{worldpop}}|sigfig=2|disp=table}} || {{dts|1 March 2023|abbr=on}} | 2023 census result<ref>{{cite web |url=https://www.pbs.gov.pk/sites/default/files/population/2023/Pakistan.pdf |title=Announcement of Results of 7th Population and Housing Census-2023 'The Digital Census' |date=5 August 2023 |website=Pakistan Bureau of Statistics (www.pbs.gov.pk) |access-date=14 August 2023 |archive-date=14 August 2023 |archive-url=https://web.archive.org/web/20230814181311/https://www.pbs.gov.pk/sites/default/files/population/2023/Pakistan.pdf |url-status=live }}</ref> || {{efn|Includes the population of [[Kashmir conflict|Pakistan-administered]] [[Azad Kashmir]] and [[Gilgit-Baltistan]].}} |- | {{nts|6}} | {{flag|Nigeria}} | {{n+p|223800000|{{worldpop}}|sigfig=2|disp=table}} || {{dts|1 Jul 2023|abbr=on}} | Official projection<ref>{{cite web |url=https://cdn.sanity.io/files/5otlgtiz/production/d7fb94a5c16220cd7559ac066f95f9d53a433ac8.pdf |title=The annual population lecture series - APLS (pg.1 and pg.159) |publisher=National Population Commission - NPC Nigeria |website=www.nationalpopulation.gov.ng |date=22 November 2023 |access-date=11 March 2024}}</ref> || |- | {{nts|7}} | {{flag|Brazil}} | {{n+p|203080756|{{worldpop}}|sigfig=2|disp=table}} || {{dts|1 Aug 2022|abbr=on}} | 2022 census result<ref>{{cite web |url = https://censo2022.ibge.gov.br/panorama/ |title = Overview of the 2022 Census |website = censo2022.ibge.gov.br |access-date = 29 Jun 2023 |archive-date = 28 June 2023 |archive-url = https://web.archive.org/web/20230628195406/https://censo2022.ibge.gov.br/panorama/ |url-status = live }}</ref> || |- | {{nts|8}} | {{flag|Bangladesh}} | {{n+p|169828911|{{worldpop}}|sigfig=2|disp=table}} || {{dts|14 Jun 2022|abbr=on}} | 2022 census result<ref>{{cite web |title=Population & Housing Census 2022 - Post Enumeration Check (PEC) Adjusted Population |url=https://bbs.portal.gov.bd/sites/default/files/files/bbs.portal.gov.bd/page/b343a8b4_956b_45ca_872f_4cf9b2f1a6e0/2023-04-18-08-42-4f13d316f798b9e5fd3a4c61eae4bfef.pdf |website=Bangladesh Bureau of Statistics (www.bbs.gov.bd) |date=2023-04-18 |access-date=2023-08-24 |language=en |archive-date=24 August 2023 |archive-url=https://web.archive.org/web/20230824082228/https://bbs.portal.gov.bd/sites/default/files/files/bbs.portal.gov.bd/page/b343a8b4_956b_45ca_872f_4cf9b2f1a6e0/2023-04-18-08-42-4f13d316f798b9e5fd3a4c61eae4bfef.pdf |url-status=live }}</ref> || |- | {{nts|9}} | {{flag|Russia}} | {{n+p|146150789|{{worldpop}}|sigfig=2|disp=table}} || {{dts|1 Jan 2024|abbr=on}} | Official estimate<ref>{{cite web |title= Оценка численности постоянного населения на 1 января 2024 г. и в среднем за 2023 [Estimated permanent population on 1 January 2024 and average 2023] |url=https://rosstat.gov.ru/storage/mediabank/OkPopul_Comp2024_Site.xlsx |publisher=[[Federal State Statistics Service (Russia)|Russian Federal State Statistics Service (Росстат)]] |website= www.rosstat.gov.ru |access-date=2024-04-11 |language=ru}}</ref> || {{efn|Includes the [[Republic of Crimea (Russia)|Republic of Crimea]] and [[Sevastopol]], administrative areas on the [[Crimea|Crimean Peninsula]] occupied by Russia. The population does not include the [[Russian annexation of Donetsk, Kherson, Luhansk and Zaporizhzhia oblasts of Ukraine|four annexed Ukrainian oblasts]]. The Ukrainian government and [[Political status of Crimea|most of the world's other states]] consider the Crimean Peninsula, [[Donetsk Oblast|Donetsk]], [[Kherson Oblast|Kherson]], [[Luhansk Oblast|Luhansk]] and [[Zaporizhzhia Oblast|Zaporizhzhia]] parts of [[Administrative divisions of Ukraine|Ukraine's territory]].}} |- | {{nts|10}} | {{flag|Mexico}} | {{n+p|129713690|{{worldpop}}|sigfig=2|disp=table}} || {{dts|31 Mar 2024|abbr=on}} | National quarterly estimate<ref>*data provided by [https://www.inegi.org.mx/sistemas/Infoenoe/Default_15mas.aspx INEGI → ENOE]* {{cite web |title=Mexico Population |url=https://www.economy.com/mexico/population/not-seasonally-adjusted |publisher=Moody's Analytics |website=www.economy.com |access-date=2024-05-28}}</ref> || |- | {{nts|11}} | {{flag|Japan}} | {{n+p|123960000|{{worldpop}}|sigfig=2|disp=table}} || {{dts|1 Jul 2024|abbr=on}} | Monthly national estimate<ref>{{Cite news |title=Population Estimates by Age (Five-Year Groups) and Sex |work=[[Statistics Bureau (Japan)|Statistics Bureau of Japan]] |url=https://www.stat.go.jp/english/data/jinsui/tsuki/index.htm |access-date=2024-04-28 |archive-date=1 April 2022 |archive-url=https://web.archive.org/web/20220401013551/https://www.stat.go.jp/english/data/jinsui/tsuki/index.htm |url-status=live }}</ref>|| |- | {{nts|12}} | {{flag|Philippines}} | {{n+p|114163719|{{worldpop}}|sigfig=2|disp=table}} || {{dts|1 Jul 2024|abbr=on}} | Official projection<ref>{{cite web |url = https://psa.gov.ph/statistics/census/projected-population |title = Population Projection Statistics |date = 28 Mar 2021 |website = psa.gov.ph |access-date = 15 Nov 2023}}</ref> || |- | {{nts|13}} | {{flag|Ethiopia}} | {{n+p|109499000|{{worldpop}}|sigfig=2|disp=table}} || {{dts|1 Jul 2024|abbr=on}} | National annual projection<ref>{{cite web |url=https://www.statsethiopia.gov.et/wp-content/uploads/2024/07/Projected_Population-2024.pdf |title=Population Size by Sex, Area and Density by Region, Zone and Wereda: July 2024 |website=www.statsethiopia.gov.et |publisher=Ethiopian Statistical Service (ESS) |access-date=2024-07-07 |archive-date=2024-07-07 |archive-url=https://web.archive.org/web/20240707022122/https://www.statsethiopia.gov.et/wp-content/uploads/2024/07/Projected_Population-2024.pdf |url-status=dead }}</ref> || |- | {{nts|14}} | {{flag|Egypt}} | {{n+p|105,914,499 |{{worldpop}}|sigfig=2|disp=table}} || {{dts|1 Jan 2024|abbr=on}} | Official estimate<ref>{{cite web |url=https://www.capmas.gov.eg/Admin/Pages%20Files/20245121324361-%20pop_new.pdf |title=Egypt in Figures 2024 - Population Estimates By Sex & Governorate 1/1/2024 |publisher=[[Central Agency for Public Mobilization and Statistics (Egypt)|CAPMAS]] |website=www.capmas.gov.eg |date=2024-03-01 |access-date=2024-05-28}}</ref> || |- | {{nts|15}} | {{flag|Vietnam}} | {{n+p|100300000|{{worldpop}}|sigfig=2|disp=table}} || {{dts|July 2023|abbr=on}} | Official estimate<ref name="gso-pop">{{cite web |title=Tổng cục Thống kê: Báo cáo tình hình kinh tế – xã hội quý IV và năm 2023|url=https://www.gso.gov.vn/bai-top/2023/12/bao-cao-tinh-hinh-kinh-te-xa-hoi-quy-iv-va-nam-2023/ |date=29 Dec 2023 |archive-url=https://web.archive.org/web/20230619093316/https://zingnews.vn/tong-cuc-thong-ke-dan-so-viet-nam-da-vuot-100-trieu-nguoi-vao-thang-4-post1441206.html |archive-date=19 June 2023 |website=Zing News}}</ref> || |- | {{nts|16}} | {{flag|Democratic Republic of the Congo}} | {{n+p|98370000|{{worldpop}}|sigfig=2|disp=table}} || {{dts|1 Jul 2019|abbr=on}} | Official figure<ref>{{cite web |url = https://ins.cd/wp-content/uploads/2022/06/ANNUAIRE-STATISTIQUE-2020.pdf |title = Annuaire Statistique RDC 2020 |at = p29 |date = 1 Mar 2021 |website = ins.cd |access-date = 25 Jun 2023 |archive-date = 12 February 2023 |archive-url = https://web.archive.org/web/20230212172734/https://ins.cd/wp-content/uploads/2022/06/ANNUAIRE-STATISTIQUE-2020.pdf |url-status = dead}}</ref> || |- | {{nts|17}} | {{flag|Turkey}} | {{n+p|85372377|{{worldpop}}|sigfig=2|disp=table}} || {{dts|31 Dec 2023|abbr=on}} | Official estimate<ref>{{cite web |url=https://data.tuik.gov.tr/Bulten/Index?p=The-Results-of-Address-Based-Population-Registration-System-2023-49684&dil=2 |title=The Results of Address Based Population Registration System, 2023 |publisher=[[Turkish Statistical Institute]] |website=www.tuik.gov.tr |date=6 February 2024 |access-date=6 February 2024}}</ref> || |- | {{nts|18}} | {{flag|Iran}} | {{n+p|85000000|{{worldpop}}|sigfig=2|disp=table}} || {{dts|16 May 2023|abbr=on}} | Official estimate<ref>{{cite web |url= https://wanaen.com/irans-population-crossed-the-border-of-85-million-people/ |archive-url=https://web.archive.org/web/20230517161108/https://wanaen.com/irans-population-crossed-the-border-of-85-million-people/ |archive-date=2023-05-17 |title= Iran's population crossed the border of 85 million people |website=wanaen.com |access-date=16 Jun 2024}}</ref> || |- | {{nts|19}} | {{flag|Germany}} | {{n+p|84673158|{{worldpop}}|sigfig=2|disp=table}} || {{dts|31 Mar 2024|abbr=on}} | National quarterly estimate<ref>{{cite web |url = https://www.destatis.de/EN/Themes/Society-Environment/Population/Current-Population/Tables/liste-current-population.html |title = Population by nationality and sex 2023 |website = destatis.de |access-date = 30 Sep 2023 |archive-date = 1 September 2020 |archive-url = https://web.archive.org/web/20200901043841/https://www.destatis.de/EN/Themes/Society-Environment/Population/Current-Population/Tables/liste-current-population.html |url-status = live }}</ref> || |- | {{nts|20}} | {{flag|France}} | {{n+p|68449000|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jun 2024|abbr=on}} | Monthly national estimate<ref>{{cite web |title=Démographie - Population au début du mois - France (inclus Mayotte à partir de 2014) |url=https://www.insee.fr/fr/statistiques/serie/001641607 |publisher=INSEE |website=www.insee.fr |access-date= 29 Jun 2024 |archive-date=10 October 2017 |archive-url=https://web.archive.org/web/20171010115405/https://www.insee.fr/fr/statistiques/serie/001641607 |url-status=live }}</ref> || {{efn|Includes the integral 18 [[regions of France]] (including 5 [[Overseas departments and regions of France|overseas departments and regions]]). Does not include France's 5 [[Overseas collectivity|overseas collectivities]]: [[French Polynesia]], [[Saint Barthélemy]], [[Collectivity of Saint Martin|Saint Martin]], [[Saint Pierre and Miquelon]], and [[Wallis and Futuna]], and the sui generis collectivity of [[New Caledonia]], which are shown separately. The [[French Southern and Antarctic Lands]] (an Antarctic territorial claim hosting only government officials and research station staff) and [[Clipperton Island]] (an uninhabited state private property of France) are not listed at all due to their extraordinary nature.}} |- | {{nts|21}} | {{flag|United Kingdom}} | {{n+p|{{UK subdivision population|GSS=K02000001}}|{{worldpop}}|sigfig=1|disp=table}} || {{dts|June {{UK subdivision statistics year}}|abbr=on}} | Official estimate<ref>{{UK subdivision statistics citation}}</ref> || {{efn|Does not include the three [[Crown Dependencies|British Crown Dependencies]] and the 14 [[British Overseas Territories]], listed separately. Four British Overseas Territories are not listed due to their extraordinary nature. The four not listed are ''[[British Antarctic Territory]]'' (an Antarctic territorial claim hosting only government officials and research station staff), the [[British Indian Ocean Territory]] (a military base), [[South Georgia and the South Sandwich Islands]] (hosts only government officials and research station staff), and the [[Akrotiri and Dhekelia|Sovereign Base Areas of Akrotiri and Dhekelia]] (a military base where permanent residency is limited to citizens of the [[Cyprus|Republic of Cyprus]]).}} |- | {{nts|22}} | {{flag|Thailand}} | {{n+p|65990480|{{worldpop}} |sigfig=1|disp=table}} || {{dts|1 May 2024|abbr=on}} | Monthly national estimate<ref>{{cite web |url=https://stat.bora.dopa.go.th/stat/statnew/statMONTH/statmonth/#/mainpage |title=สถิติประชากรทางการทะเบียนราษฎร(รายเดือน) - Official population statistics from the civil registration (monthly) |last=จำนวนประชากร (population) → ขอบเขตข้อมูล (data scope - year, month) → ยอดรวมข้อมูล (total data) |publisher=The Bureau of Registration Administration (BORA) |website=www.bora.dopa.go.th |access-date=28 May 2024 |language=thai}}</ref> || |- | {{nts|23}} | {{flag|Tanzania}} | {{n+p|65070448|{{worldpop}} |sigfig=1|disp=table}} || {{dts|1 Jul 2024|abbr=on}} | National annual projection<ref>{{cite web |url=https://www.citypopulation.de/en/tanzania/cities//cities/ |title=National Population Projections |date=Feb 2018 |website=National Bureau of Statistics |archive-url=https://web.archive.org/web/20180612142155/http://www.nbs.go.tz/nbs/takwimu/census2012/Projection-Report-20132035.pdf |archive-date=12 Jun 2018 }}</ref> || {{efn|Includes [[Zanzibar]].}} |- | {{nts|24}} | {{flag|South Africa}} | {{n+p|62027503|{{worldpop}}|sigfig=1|disp=table}} || {{dts|2 Feb 2022|abbr=on}} | 2022 census result<ref>{{cite web |url=https://www.gov.za/sites/default/files/P03014_Census_2022_Statistical_Release.pdf |title=Census 2022 - Statistical Release |publisher=Government of South Africa |website=www.gov.za |date=10 October 2023 |access-date=15 October 2023 |archive-date=15 October 2023 |archive-url=https://web.archive.org/web/20231015192129/https://www.gov.za/sites/default/files/P03014_Census_2022_Statistical_Release.pdf |url-status=live }}</ref> || |- | {{nts|25}} | {{flag|Italy}} | {{n+p|58968501|{{worldpop}}|sigfig=1|disp=table}} || {{dts|30 Apr 2024|abbr=on}} | Monthly national estimate<ref>{{cite web |url=https://demo.istat.it/app/?i=D7B |title=ISTAT - Bilancio Demografico Mensile / Monthly Demographic Balance |publisher=Istituto Nazionale di Statistica (Istat) |website=demo.istat.it |language=it |access-date=2024-07-02 |archive-date=17 January 2023 |archive-url=https://web.archive.org/web/20230117173414/https://demo.istat.it/app/?i=D7B |url-status=live}}</ref> || |- | {{nts|26}} | {{flag|Myanmar}} | {{n+p|56712559|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2024|abbr=on}} | National annual projection<ref>{{cite web |url=https://myanmar.unfpa.org/sites/default/files/pub-pdf/4F_Population%20Projections.pdf |title=Thematic Report on Population Projections for The Union of Myanmar, States/Regions, Rural and Urban Areas, 2014 - 2050 |access-date = 2 Jul 2024 |publisher = Department of Population |archive-url= https://web.archive.org/web/20200924221152/https://myanmar.unfpa.org/sites/default/files/pub-pdf/4F_Population%20Projections.pdf |archive-date=24 Sep 2020 |url-status=live }}</ref> || |- | {{nts|27}} | {{flag|Colombia}} | {{n+p|52695952|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2024|abbr=on}} | Official projection<ref>{{cite web |url=https://www.dane.gov.co/files/censo2018/proyecciones-de-poblacion/Nacional/DCD-area-proypoblacion-Nac-2020-2070.xlsx |title=Proyecciones de Población DANE |access-date=10 April 2023 |archive-date=10 April 2023 |archive-url=https://web.archive.org/web/20230410005852/https://www.dane.gov.co/files/censo2018/proyecciones-de-poblacion/Nacional/DCD-area-proypoblacion-Nac-2020-2070.xlsx |url-status=live}}</ref> || |- | {{nts|28}} | {{flag|Kenya}} | {{n+p|52428290|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2024|abbr=on}} | Official projection<ref>{{cite web |date=Sep 2022 |title=2019 Kenya Population and Housing Census Results: Analytical Report on Population Projections, Vol XVI (Table 3.1: Projected Population, Kenya, 2020-2045) |url=https://www.knbs.or.ke/download/2019-kphc-analytical-report-on-population-projections-vol-xvi/ |url-status=live |archive-url=https://web.archive.org/web/20221204110524/https://www.knbs.or.ke/download/2019-kphc-analytical-report-on-population-projections-vol-xvi/ |archive-date=4 Dec 2022 |access-date=25 Jun 2023 |website=knbs.or.ke |publisher=Kenya National Bureau of Statistics |page=23}}</ref>|| |- | {{nts|29}} | {{flag|South Korea}} | {{n+p|51285153|{{worldpop}}|sigfig=1|disp=table}} || {{dts|30 Apr 2024|abbr=on}} | Monthly national estimate<ref>{{cite web |title=Resident population by City, County, and District |url=https://kosis.kr/statHtml/statHtml.do?orgId=101&tblId=DT_1B040A3&conn_path=I2&language=en |publisher=[[Statistics Korea|Statistics Korea (KOSTAT)]] → Korean Statistical Information Service (KOSIS) |website=www.kostat.go.kr |access-date=2024-05-28}}</ref>|| |- | {{nts|30}} | {{flag|Spain}} | {{n+p|48692804|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 April 2024|abbr=on}} | National quarterly estimate<ref>{{cite web |url=https://www.ine.es/dyngs/INEbase/en/operacion.htm?c=Estadistica_C&cid=1254736177095&menu=ultiDatos&idp=1254735572981 |title=Population resident in Spain - Latest data |publisher=[[National Statistics Institute (Spain)|INE España]] |website=www.ine.es |access-date=10 May 2024}}</ref> || |- | {{nts|31}} | {{flag|Argentina}} | {{n+p|47067441|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2024|abbr=on}} | National annual projection<ref>{{cite web |url=https://www.indec.gob.ar/indec/web/Nivel3-Tema-2-24 |title=Proyecciones y estimaciones |publisher=INDEC |website=www.indec.gob.ar |access-date=2024-07-03 |archive-date=9 June 2023 |archive-url=https://web.archive.org/web/20230609202751/https://www.indec.gob.ar/indec/web/Nivel3-Tema-2-24 |url-status=live }}</ref> || |- | {{nts|32}} | {{flag|Algeria}} | {{n+p|46700000|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2024|abbr=on}} | National annual projection<ref>{{cite web |url=https://www.ons.dz/IMG/pdf/Demographie_Algerienne2020_2023.pdf |title=DEMOGRAPHIE ALGERIENNE 2020 à 2025 |publisher=[[National Office of Statistics]] (ONS) |website=www.ons.dz |date=July 2024 |access-date=14 Jul 2024 |language=French}}</ref> || |- | {{nts|33}} | {{flag|Uganda}} | {{n+p| 45935046|{{worldpop}}|sigfig=1|disp=table}} || {{dts|10 May 2024|abbr=on}} | 2024 census result<ref>{{cite web |url=https://www.ubos.org/wp-content/uploads/publications/National-Population-and-Housing-Census-2024-Preliminary-Report.pdf |title=NPHC 2024 – Preliminary Results |access-date=16 July 2024 }}</ref> || |- | {{nts|34}} | {{flag|Iraq}} | {{n+p|44414800|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2024|abbr=on}} | Official estimate<ref>{{cite web |access-date=2 Jul 2024 |title=Estimated population of Iraq for the period (2015-2030)|url=http://cosit.gov.iq/ar/2013-01-31-08-43-38 |website=cosit.gov.iq |archive-date=3 December 2019 |archive-url=https://web.archive.org/web/20191203214014/http://cosit.gov.iq/ar/2013-01-31-08-43-38 |url-status=live}}</ref> || |- | {{nts|35}} | {{flag|Sudan}} | {{n+p|41984500|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2018|abbr=on}} | Official projection<ref>{{cite web |url = http://www.citypopulation.de/en/sudan/ |title = Sudan |date = 5 Sep 2019 |website = City Population |access-date = 25 Jun 2023 |archive-date = 26 June 2023 |archive-url = https://web.archive.org/web/20230626083830/http://www.citypopulation.de/en/sudan/ |url-status = live }}</ref> || |- | {{nts|36}} | {{flag|Canada}} | {{n+p|41012563|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 April 2024|abbr=on}} | National quarterly estimate<ref>{{cite web |url = https://www150.statcan.gc.ca/t1/tbl1/en/tv.action?pid=1710000901 |title = Population estimates, quarterly |date = 31 March 2024 |website = statcan.gc.ca |access-date = 31 March 2024}}</ref> || |- | {{nts|37}} | {{flag|Poland}} | {{n+p|37582000|{{worldpop}}|sigfig=1|disp=table}} || {{dts|30 Apr 2024|abbr=on}} | Monthly national estimate<ref>{{cite web |url=https://stat.gov.pl/en/topics/other-studies/informations-on-socio-economic-situation/statistical-bulletin-no-52024,4,163.html |title=Statistical Bulletin No 5/2024 → Population |publisher=[[Statistics Poland]] (GUS) |website=www.stat.gov.pl |date=2024-06-25 |access-date=2024-06-29}}</ref> || |- | {{nts|38}} | {{flag|Morocco}} | {{n+p|37369652|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2024|abbr=on}} | Official projection<ref>{{cite web |url = https://www.hcp.ma/Population-du-Maroc-par-annee-civile-en-milliers-et-au-milieu-de-l-annee-1960-2050_a3526.html |title = Population of Morocco by calendar year |website = hcp.ma |access-date = 2 Jul 2024}}</ref> || {{efn|Does not include the [[territorial dispute|disputed territory]] of [[Western Sahara]] ([[Sahrawi Arab Democratic Republic]]).}} |- | {{nts|39}} | {{flag|Uzbekistan}} | {{n+p|36963262|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Apr 2024|abbr=on}} | National quarterly estimate<ref>{{cite web |url = https://stat.uz/en/ |title = Uzbekistan by the Numbers |website = stat.uz |access-date=30 Apr 2024 |archive-date = 13 September 2023 |archive-url = https://web.archive.org/web/20230913103557/https://stat.uz/en/ |url-status = live }}</ref> || |- | {{nts|40}} | {{flag|Ukraine}} | {{n+p|36700000|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2023|abbr=on}} | UN estimate<ref>{{cite web |title=Population of Ukraine - Population (by estimate) as of 2023 |website=United Nations Population Fund |publisher=United Nations |url=https://www.unfpa.org/data/world-population/UA}}</ref> || {{efn|Does not include the [[Autonomous Republic of Crimea]] and the city of [[Sevastopol]].}} |- | {{nts|41}} | {{flag|Angola}} | {{n+p| 35121734|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jun 2024|abbr=on}} | National annual projection<ref>{{cite web |url=https://www.ine.gov.ao/ |title=Início |website=www.ine.gov.ao |access-date=2 Jul 2024 |archive-url=https://web.archive.org/web/20230601043657/https://www.ine.gov.ao/ |archive-date=1 Jun 2023 |url-status=live}}</ref> || |- | {{nts|42}} | {{flag|Afghanistan}} | {{n+p|34262840|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2023|abbr=on}} |Official estimate<ref>{{cite web |url = http://nsia.gov.af:8080/wp-content/uploads/2022/05/%D8%A8%D8%B1%D8%A2%D9%88%D8%B1%D8%AF-%D9%86%D9%81%D9%88%D8%B3-%D8%B3%D8%A7%D9%84-1401_compressed.pdf |title = Estimated Population of Afghanistan 2022-23 |access-date = 4 Jul 2023 |archive-date = 16 July 2023 |archive-url = https://web.archive.org/web/20230716103803/http://nsia.gov.af:8080/wp-content/uploads/2022/05/%D8%A8%D8%B1%D8%A2%D9%88%D8%B1%D8%AF-%D9%86%D9%81%D9%88%D8%B3-%D8%B3%D8%A7%D9%84-1401_compressed.pdf |url-status = live }}</ref> || |- | {{nts|43}} | {{flag|Peru}} | {{n+p|34038457|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2024|abbr=on}} | National annual projection<ref>{{cite web |url=https://www.inei.gob.pe/media/MenuRecursivo/publicaciones_digitales/Est/Lib1665/index.html |title=Estimaciones y Proyecciones de la Población Nacional, 1950-2070 (3. Resultados) |publisher=[[Instituto Nacional de Estadística e Informática|INEI - Peru]] |website=www.gob.pe |access-date=2 July 2024}}</ref> || |- | {{nts|44}} | {{flag|Malaysia}} | {{n+p|33980600|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 Mar 2024|abbr=on}} | National quarterly estimate<ref>{{cite web |url = https://www.dosm.gov.my/portal-main/release-content/demographic-statistics-first-quarter-2024 |title = Demographic Statistics, First Quarter 2024 |website = dosm.gov.my |access-date = 23 Jun 2024 }}</ref> || |- | {{nts|45}} | {{flag|Mozambique}} | {{n+p|32419747|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2022|abbr=on}} | National annual projection<ref>{{cite web |url=http://www.ine.gov.mz |title=Projecções da População — Instituto Nacional de Estatistica |website=www.ine.gov.mz |access-date=18 Apr 2020 |archive-url=https://web.archive.org/web/20200307181608/http://www.ine.gov.mz/noticias/projeccoes-da-populacao-2017-2050 |archive-date=7 Mar 2020 |url-status=live}}</ref> || |- | {{nts|46}} | {{flag|Saudi Arabia}} | {{n+p|32175224|{{worldpop}}|sigfig=1|disp=table}} || {{dts|10 May 2022|abbr=on}} | 2022 census result<ref>{{cite web |title = Overall count of population |url = https://portal.saudicensus.sa/portal |publisher = General Statistics Authority - Kingdom of Saudi Arabia |website = portal.saudicensus.sa |access-date = 5 Jul 2023 |archive-date = 28 July 2023 |archive-url = https://web.archive.org/web/20230728002753/https://portal.saudicensus.sa/portal |url-status = live }}</ref> || |- | {{nts|47}} | {{flag|Yemen}} | {{n+p|31888698|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2022|abbr=on}} | Official estimate<ref>{{cite web |url=https://moh.gov.ye/storage/1399/HeRAMS_Yemen_Baseline_Report_2023-NCD_and_Mental_Health.pdf |title=Annex II: Population estimations (pg.43) |publisher=Yemeni Ministry Of Health |website=www.moh.gov.ye/en |date=16 Jul 2023 |access-date=23 Sep 2023 |archive-date=2 October 2023 |archive-url=https://web.archive.org/web/20231002083024/https://moh.gov.ye/storage/1068/HeRAMS_YEM_baseline_2023-Communicable_Diseases_20230707_draft.pdf |url-status=live }}</ref> || |- | {{nts|48}} | {{flag|Ghana}} | {{n+p|30832019|{{worldpop}}|sigfig=1|disp=table}} || {{dts|27 Jun 2021|abbr=on}} | 2021 census result<ref>{{cite web |url = https://census2021.statsghana.gov.gh/gssmain/fileUpload/reportthemelist/Volume%203%20Highlights.pdf |date = 1 Feb 2022 |title = Population and Regions and Districts Report - Volume 3 Highlights |access-date = 4 Jul 2023 |archive-date = 4 July 2023 |archive-url = https://web.archive.org/web/20230704062216/https://census2021.statsghana.gov.gh/gssmain/fileUpload/reportthemelist/Volume%203%20Highlights.pdf |url-status = live }}</ref> || |- | {{nts|49}} | {{flag|Madagascar}} | {{n+p|30811969|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2024|abbr=on}} | National annual projection<ref>{{cite web |url=https://www.instat.mg/documents/upload/main/INSTAT-RGPH3_Projectionsdemographiques.pdf |title=Projections démographiques |website=instat.mg |access-date=2 July 2024 |archive-url=https://web.archive.org/web/20240615103016/https://www.instat.mg/documents/upload/main/INSTAT-RGPH3_Projectionsdemographiques.pdf |archive-date=15 Jun 2024 |url-status=live}}</ref> || |- | {{nts|50}} | {{flag|Ivory Coast}} | {{n+p|29389150|{{worldpop}}|sigfig=1|disp=table}} || {{dts|14 Dec 2021|abbr=on}} | 2021 census result<ref>{{cite web |url=http://www.ins.ci/n |title=Institut National de la Statistique (INS) - RGPH-2021 Résultats globaux |website=www.ins.ci |access-date=10 Aug 2022 |archive-date=10 March 2018 |archive-url=https://web.archive.org/web/20180310230400/http://www.ins.ci/n |url-status=live}}</ref> || |- | {{nts|51}} | {{flag|Nepal}} | {{n+p|29164578|{{worldpop}}|sigfig=1|disp=table}} || {{dts|25 Nov 2021|abbr=on}} | 2021 census result<ref>{{cite web |url=https://censusnepal.cbs.gov.np/results |title=National Population and Housing Census 2021 |website=www.cbs.gov.np |access-date=28 Mar 2023 |archive-date=27 March 2023 |archive-url=https://web.archive.org/web/20230327230350/http://censusnepal.cbs.gov.np/results |url-status=live}}</ref> || |- | {{nts|52}} | {{flag|Cameroon}} | {{n+p|28758503|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2024|abbr=on}} | National annual projection<ref>{{cite web |url=https://ins-cameroun.cm/document/projections-demographiques-et-estimations-des-cibles-prioritaires-des-differents-programmes-et-interventions-de-sante/ |title=Projections demographiques et estimations des cibles prioritaires des differents programmes et interventions de sante |publisher=INS du Cameroun |website=www.ins-cameroun.cm |date=2023-06-30 |access-date=2 July 2024 |archive-date=2 October 2023 |archive-url=https://web.archive.org/web/20231002083019/https://ins-cameroun.cm/document/projections-demographiques-et-estimations-des-cibles-prioritaires-des-differents-programmes-et-interventions-de-sante/ |url-status=live }}</ref> || |- | {{nts|53}} | {{flag|Venezuela}} | {{n+p|28302000|{{worldpop}}|sigfig=1|disp=table}} || {{dts|30 Jun 2019|abbr=on}} | Official projection<ref>{{cite web |url = https://www.citypopulation.de/en/venezuela/cities/ |title = Venezuela |date = 11 May 2019 |website = City Population |access-date = 27 Jun 2023 |archive-date = 15 May 2019 |archive-url = https://web.archive.org/web/20190515111338/https://www.citypopulation.de/en/venezuela/cities/ |url-status = live }}</ref> || |- | {{nts|54}} | {{flag|Australia}} | {{n+p|27,282,542|{{worldpop}}|sigfig=1|disp=table}} || {{dts|21 June 2024|abbr=on}} | Official projection<ref>{{cite web |url = https://www.abs.gov.au/statistics/people/population/population-clock-pyramid |title = Population clock and pyramid |date = 23 Mar 2024 |website = abs.gov.au |access-date = 23 Mar 2024}}</ref> || {{efn|Does not include the outside territories of [[Christmas Island]], the [[Cocos (Keeling) Islands]], and [[Norfolk Island]].}} |- | {{nts|55}} | {{flag|Niger}} | {{n+p|26312034|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2024|abbr=on}} | National annual projection<ref>{{cite web |url=https://www.stat-niger.org/projections/ |title=Projections démographiques du Niger Horizon 2012-2024 (extraction de donnees) |publisher=INSN |website=www.stat-niger.org |access-date=2 July 2024 |archive-date=10 June 2023 |archive-url=https://web.archive.org/web/20230610170556/https://www.stat-niger.org/projections/ |url-status=live }}</ref> || |- | {{nts|56}} | {{flag|North Korea}} | {{n+p|25660000|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2021|abbr=on}} | National annual projection<ref>{{cite web |url=http://www.searo.who.int/dprkorea/documents/population-projection-dprk-2014-50.pdf |title=Democratic People's Republic of Korea Population Projection (2014–2050) |access-date=16 Oct 2020 |archive-url=https://web.archive.org/web/20190412033819/http://www.searo.who.int/dprkorea/documents/population-projection-dprk-2014-50.pdf |archive-date=12 Apr 2019 |url-status=dead }}</ref>|| |- | {{hs|end}} – | {{flag|Taiwan}} | {{n+p|23413608|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 May 2024|abbr=on}} | Official estimate<ref>{{cite web |url=https://eng.stat.gov.tw/point.asp?index=9 |title=National Statistics, Taiwan |website=eng.stat.gov.tw |access-date=23 Jun 2024 |archive-date=31 January 2020 |archive-url=https://web.archive.org/web/20200131125043/https://eng.stat.gov.tw/point.asp?index=9 |url-status=live}}</ref> || {{efn|[[Taiwan]] includes [[Penghu]], [[Kinmen]], [[Matsu Islands|Matsu]] and other minor islands.}} |- | {{nts|57}} | {{flag|Burkina Faso}} | {{n+p|23409015|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2024|abbr=on}} | National projection<ref>{{cite web |url=https://www.insd.bf/sites/default/files/2023-02/VOLUME%204%20PROJECTIONS%20DEMOGRAPHIQUES%202020%202035%20RGPH%20%C3%A9dition%202-%2006-02-2023.pdf |title=PROJECTIONS DEMOGRAPHIQUES 2020-2035 |publisher=INSD Burkina Faso |website=www.insd.bf |date=January 2023 |access-date=2 July 2024}}</ref> || |- | {{nts|58}} | {{flag|Syria}} | {{n+p|22923000|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2021|abbr=on}} | Official estimate<ref>{{cite web |url = https://www.citypopulation.de/en/syria/ |title = Syrian Arab Republic |date = 25 Sep 2022 |website = City Population |access-date = 15 May 2023 |archive-date = 3 June 2023 |archive-url = https://web.archive.org/web/20230603085127/https://www.citypopulation.de/en/syria/ |url-status = live }}</ref> || |- | {{nts|59}} | {{flag|Mali}} | {{n+p|22395489|{{worldpop}}|sigfig=1|disp=table}} || {{dts|15 Jun 2022|abbr=on}} | 2022 census result<ref>{{cite web |url=https://instat-mali.org/fr |title=TAILLE DE LA POPULATION EN 2022 (HBTS) (SOURCE RÉSULTATS RGPH5) |publisher=INSTAT Mali |website=www.instat-mali.org/fr |date=2023-08-10 |access-date=2023-09-23 |archive-date=3 October 2023 |archive-url=https://web.archive.org/web/20231003092311/https://instat-mali.org/fr |url-status=live }}</ref> || |- | {{nts|60}} | {{flag|Malawi}} | {{n+p|22245431|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2023|abbr=on}} | National annual projection<ref>{{cite web |url=http://www.nsomalawi.mw/index.php?option=com_content&view=article&id=136%3Amalawi-table-30-population-by-age-and-sex&catid=8&Itemid=3 |title=Malawi: Table 3.0 :Population by Age and Sex |access-date=6 July 2022 |archive-date=6 July 2022 |archive-url=https://web.archive.org/web/20220706110200/http://www.nsomalawi.mw/index.php?option=com_content&view=article&id=136:malawi-table-30-population-by-age-and-sex&catid=8&Itemid=3 |url-status=live}}</ref> || |- | {{nts|61}} | {{flag|Sri Lanka}} | {{n+p|22037000|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2023|abbr=on}} | Official estimate<ref>{{cite web |url=http://www.statistics.gov.lk/Population/StaticalInformation/VitalStatistics/ByDistrictandSex |title=Mid-year Population Estimates by District & Sex, 2014 - 2023 |website=statistics.gov.lk |access-date=26 Sep 2023 |archive-date=14 October 2020 |archive-url=https://web.archive.org/web/20201014212854/http://www.statistics.gov.lk/Population/StaticalInformation/VitalStatistics/ByDistrictandSex |url-status=live }}</ref> || |- | {{nts|62}} | {{flag|Kazakhstan}} | {{n+p|20118478|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 May 2024|abbr=on}} | Official estimate<ref>{{cite web |url=https://stat.gov.kz/en/ |title=Population |publisher=QAZSTAT |website=www.stat.gov.kz |access-date=23 Jun 2024}}</ref> || |- | {{nts|63}} | {{flag|Chile}} | {{n+p|19960889|{{worldpop}}|sigfig=1|disp=table}} || {{dts|30 Jun 2023|abbr=on}} | National annual projection<ref>{{cite web |url=https://www.ine.cl/estadisticas/sociales/demografia-y-vitales/proyecciones-de-poblacion |title = Proyecciones de población |website =www.ine.cl |access-date=30 Sep 2020 |archive-url=https://web.archive.org/web/20200828051041/https://www.ine.cl/estadisticas/sociales/demografia-y-vitales/proyecciones-de-poblacion |archive-date=28 Aug 2020 |url-status=live}}</ref> || |- | {{nts|64}} | {{flag|Zambia}} | {{n+p|19610769|{{worldpop}}|sigfig=1|disp=table}} || {{dts|14 Sep 2022|abbr=on}} | 2022 census result<ref>{{cite web |url=https://www.lusakatimes.com/2022/12/24/preliminary-data-shows-population-grew-from-13-1-million-in-2010-to-19-6-million-in-2022/ |title=Preliminary data shows Population grew from 13.1 million in 2010 to 19.6 million in 2022 |website=www.zamstats.gov.zm |date=24 Dec 2022 |access-date=27 Dec 2022 |archive-url=https://web.archive.org/web/20221227142047/https://www.zamstats.gov.zm/population-size-by-sex-and-rural-urban-zambia-2022/ |archive-date=27 Dec 2022 |url-status=live}}</ref> || |- | {{nts|65}} | {{flag|Romania}} | {{n+p|19051562|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2023|abbr=on}} | Official estimate<ref>{{cite web |url=https://insse.ro/cms/sites/default/files/com_presa/com_pdf/poprez_ian2023r.pdf |title=Populaţia rezidentă la 1 Ianuarie 2023 |language=ro |trans-title= |publisher=INSSE |website=insse.ro |access-date=30 August 2023 |archive-date=30 August 2023 |archive-url=https://web.archive.org/web/20230830090953/https://insse.ro/cms/sites/default/files/com_presa/com_pdf/poprez_ian2023r.pdf |url-status=live }}</ref> || |- | {{nts|66}} | {{flag|Chad}} | {{n+p|18675547|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2024|abbr=on}} | National annual projection<ref>{{cite web |url=https://www.inseed.td/ |website=www.inseed-td|publisher=Institut National de la Statistique, des Etudes Economiques et Démographiques |title=Population* 2024 |access-date=2 July 2024 |archive-date=14 October 2023 |archive-url=https://web.archive.org/web/20231014184620/https://www.inseed.td/ |url-status=live}}</ref> || |- | {{nts|67}} | {{flag|Somalia}} | {{n+p|18143379|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2023|abbr=on}} | UN projection<ref name="unpop" /> || {{efn|Includes [[Somaliland]].}} |- | {{nts|68}} | {{flag|Senegal}} | {{n+p|18126390|{{worldpop}}|sigfig=1|disp=table}} || {{dts|15 May 2023|abbr=on}} | 2023 census result<ref>{{cite web |url=https://www.ansd.sn/sites/default/files/recensements/rapport/rapport_national/RGPH-5_Rapport%20global-Prov-juillet2024_0.pdf |title=RAPPORT PROVISOIRE, PGPH-5, 2023 |access-date=16 July 2024 }}</ref> || |- | {{nts|69}} | {{flag|Netherlands}} | {{n+p|17977676|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 May 2024|abbr=on}} | Official estimate<ref>{{cite web |url=https://opendata.cbs.nl/statline/#/CBS/en/dataset/83474ENG/table?ts=1712804860730 |title=Population dynamics - month and year |website=www.opendata.cbs.nl |publisher=Centraal Bureau voor de Statistiek |date=28 June 2024 |access-date=29 June 2024}}</ref> || {{efn|Does not include the three constituent countries of the [[Kingdom of the Netherlands]] in the Caribbean Sea ([[Aruba]], [[Curaçao]], and [[Sint Maarten]]), but includes the three special municipalities of the [[Caribbean Netherlands]] ([[Bonaire]], [[Sint Eustatius]] and [[Saba (island)|Saba]]).}} |- | {{nts|70}} | {{flag|Guatemala}} | {{n+p|17843132|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2024|abbr=on}} | National annual projection<ref>{{cite web |date=2019 |publisher = Instituto Nacional de Estadística Guatemala |url= https://www.censopoblacion.gt/archivos/Estimaciones_y_proyecciones_nacionales_de_poblacion.pdf |title=Estimaciones y proyecciones nacionales de población |archive-url=https://web.archive.org/web/20191223192959/https://www.censopoblacion.gt/archivos/Estimaciones_y_proyecciones_nacionales_de_poblacion.pdf |archive-date=23 Dec 2019 |url-status=live }}</ref> || |- | {{nts|71}} | {{flag|Cambodia}} | {{n+p|17336307|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2024|abbr=on}} | Official projection<ref>{{cite web |date = Nov 2021|url = https://www.nis.gov.kh/nis/Census2019/Population%20Projection.pdf|publisher = National Institute of Statistics |title = Population Projection |access-date = 2 July 2024 |archive-date = 8 September 2023 |archive-url = https://web.archive.org/web/20230908123512/https://www.nis.gov.kh/nis/Census2019/Population%20Projection.pdf |url-status = live}}</ref> || |- | {{nts|72}} | {{flag|Ecuador}} | {{n+p|16938986|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Oct 2022|abbr=on}} | 2022 census result<ref>{{cite web |url= https://www.censoecuador.gob.ec/ |title= Inicio - Censo Ecuador |website= censoecuador.gob.ec |access-date= 15 June 2024}}</ref> || |- | {{nts|73}} | {{flag|Zimbabwe}} | {{n+p|15178979|{{worldpop}}|sigfig=1|disp=table}} || {{dts|20 Apr 2022|abbr=on}} | 2022 census result<ref>{{Cite report|url=http://www.zimstat.co.zw/wp-content/uploads/2022/07/Census2022_Preliminary_Report.pdf |title=2022 Population and housing census - Preliminary report on population figures |access-date=10 Aug 2022|publisher=Zimbabwe National Statistics Agency |archive-date=16 January 2023 |archive-url=https://web.archive.org/web/20230116224310/https://www.zimstat.co.zw/wp-content/uploads/2022/07/Census2022_Preliminary_Report.pdf |url-status=live}}</ref> || |- | {{nts|74}} | {{flag|South Sudan}} | {{n+p|14746494|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2023|abbr=on}} | National annual projection<ref>{{cite web |url=https://nbs.gov.ss/wp-content/uploads/2022/05/Population-projections-for-South-Sudan-2020-2040-1.pdf |title=South Sudan Population Projections, 2020-2040 (pg.12) |publisher=National Bureau of Statistics - South Sudan (NBSS) |website =www.nbs.gov.ss |date=January 2016 |access-date=22 January 2024}}</ref> || |- | {{nts|75}} | {{flag|Guinea}} | {{n+p|13261638|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2022|abbr=on}} | National annual projection<ref>{{cite web |url=http://www.stat-guinee.org/ |title=Institut National de la Statistique - stat-guinee.org |website=www.stat-guinee.org |access-date=25 May 2023 |archive-date=12 July 2012 |archive-url=https://web.archive.org/web/20120712050037/http://www.stat-guinee.org/ |url-status=live}}</ref> || |- | {{nts|76}} | {{flag|Rwanda}} | {{n+p|13246394|{{worldpop}}|sigfig=1|disp=table}} || {{dts|15 Aug 2022|abbr=on}} | 2022 census result<ref>{{cite web |title=MAIN INDICATORS: 5th Rwanda Population and Housing Census (PHC) |website=statistics.gov.rw |url=https://statistics.gov.rw/publication/main_indicators_2022 |access-date=17 Mar 2023 |archive-date=16 March 2023 |archive-url=https://web.archive.org/web/20230316110409/https://statistics.gov.rw/publication/main_indicators_2022 |url-status=live }}</ref> || |- | {{nts|77}} | {{flag|Benin}} | {{n+p|12910087|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2024|abbr=on}} | National annual projection<ref>{{cite web |url=https://instad.bj/images/docs/Actualit%C3%A9s/Projection%20r%C3%A9vis%C3%A9e.pdf |title=PROJECTIONS DEMOGRAPHIQUES DE 2014 A 2063 ET PERSPECTIVES DE LA DEMANDE SOCIALE DE 2014 A 2030 AU BENIN |website=www.insae-bj |access-date=2 July 2024 |archive-date=29 May 2023 |archive-url=https://web.archive.org/web/20230529104440/https://instad.bj/images/docs/Actualit%C3%A9s/Projection%20r%C3%A9vis%C3%A9e.pdf |url-status=live}}</ref> || |- | {{nts|78}} | {{flag|Burundi}} | {{n+p|12837740|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2022|abbr=on}} | National annual projection<ref>{{cite web |url=https://isteebu.bi/projection-de-la-population/ |title=Projections de la population au Burundi |website=www.isteebu.bi |access-date=25 May 2023 |archive-date=25 May 2023 |archive-url=https://web.archive.org/web/20230525150948/https://isteebu.bi/projection-de-la-population/ |url-status=live }}</ref> || |- | {{nts|79}} | {{flag|Bolivia}} | {{n+p|12006031|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2022|abbr=on}} | National annual projection<ref>{{cite web |url=https://www.ine.gob.bo/index.php/censos-y-proyecciones-de-poblacion-sociales/ |title=Proyecciones de Población, Revisión 2020|author=INE - Instituto Nacional de Estadística - Estadísticas Demográficas |website=www.ine.gob.bo |access-date=7 Dec 2022 |archive-date=6 June 2020 |archive-url=https://web.archive.org/web/20200606191939/https://www.ine.gob.bo/index.php/censos-y-proyecciones-de-poblacion-sociales/ |url-status=live}}</ref> || |- | {{nts|80}} | {{flag|Tunisia}} | {{n+p|11850232|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2023|abbr=on}} | National estimate<ref>{{cite web |url=https://www.ins.tn/en/statistiques/111 |title=Population |website=ins.tn |access-date=21 Sep 2023 |archive-date=3 October 2023 |archive-url=https://web.archive.org/web/20231003092313/https://www.ins.tn/en/statistiques/111 |url-status=live}}</ref> || |- | {{nts|81}} | {{flag|Belgium}} | {{n+p|11828948|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 May 2024|abbr=on}} | Official estimate<ref>{{cite web |url=https://www.ibz.rrn.fgov.be/fileadmin/user_upload/fr/pop/statistiques/stat-1-1_f.pdf |title=Chiffre global de la population par commune |website=www.ibz.rrn.fgov.be |access-date=23 Jun 2024 |archive-date=4 September 2020 |archive-url=https://web.archive.org/web/20200904173430/https://www.ibz.rrn.fgov.be/fileadmin/user_upload/fr/pop/statistiques/stat-1-1_f.pdf |url-status=live}}</ref> || |- | {{nts|82}} | {{flag|Papua New Guinea}} | {{n+p|11781559|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2021|abbr=on}} | National estimate<ref>{{cite web |url = https://www.nso.gov.pg/statistics/population/ |title = Population Estimates 2021 |website = nso.gov.pg |access-date = 20 Jul 2023 |archive-date = 20 July 2023 |archive-url = https://web.archive.org/web/20230720004458/https://www.nso.gov.pg/statistics/population/ |url-status = live }}</ref> || |- | {{nts|83}} | {{flag|Haiti}} | {{n+p|11743017|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2020|abbr=on}} | National annual projection<ref>{{cite web |url=http://www.ihsi.ht/pdf/projection/projectionspopulation_haiti_2007.pdf |title=Projections Population Haiti 2007 |website =www.ihsi.ht |access-date=18 Apr 2020 |archive-url=https://web.archive.org/web/20190608104339/http://www.ihsi.ht/pdf/projection/ProjectionsPopulation_Haiti_2007.pdf |archive-date=8 Jun 2019 |url-status=live}}</ref> || |- | {{nts|84}} | {{flag|Jordan}} | {{n+p|11516000|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 Dec 2023|abbr=on}} | National estimate<ref>{{cite web |url=https://dosweb.dos.gov.jo/DataBank/Population/Population_Estimares/PopulationEstimates.pdf |title= Estimated population of 2023 and some of selected data |publisher=Department of Statistics (DOS) |website=www.dosweb.dos.gov.jo |date=January 2024 |access-date=18 April 2024}}</ref> || |- | {{nts|85}} | {{flag|Cuba}} | {{n+p|11089511|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 Dec 2022|abbr=on}} | Official estimate<ref>{{cite web |url=http://www.onei.gob.cu/node/13815 |title=Indicadores Demográficos por provincias y municipios 2022 |website =www.onei.gob.cu |access-date=12 May 2023 |archive-url=https://web.archive.org/web/20201211142555/http://www.onei.gob.cu/node/13815 |archive-date=11 Dec 2020 |url-status=live }}</ref> || |- | {{nts|86}} | {{flag|Czech Republic}} | {{n+p|10859532|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 Mar 2024|abbr=on}} | Official estimate<ref>{{cite web |title=Population|url=https://www.czso.cz/csu/czso/population|publisher=[[Czech Statistical Office]] |date=28 June 2024 |access-date=28 June 2024 |archive-date=15 June 2018 |archive-url=https://web.archive.org/web/20180615234615/https://www.czso.cz/csu/czso/population |url-status=live}}</ref> || |- | {{nts|87}} | {{flag|Dominican Republic}} | {{n+p|10760028|{{worldpop}}|sigfig=1|disp=table}} || {{dts|10 Nov 2022|abbr=on}} | 2022 census result<ref>{{cite web |url=https://www.one.gob.do/noticias/2023/one-informa-los-primeros-resultados-preliminares-del-x-censo-nacional-de-poblacion-y-vivienda/ |title=ONE informa los primeros resultados preliminares del X Censo Nacional de Población y Vivienda |publisher=Oficina Nacional de Estadística (ONE) |website=www.one.gob.do |date=2023-08-11 |access-date=2023-10-04 |archive-date=13 October 2023 |archive-url=https://web.archive.org/web/20231013223345/https://www.one.gob.do/noticias/2023/one-informa-los-primeros-resultados-preliminares-del-x-censo-nacional-de-poblacion-y-vivienda/ |url-status=live }}</ref> || |- | {{nts|88}} | {{flag|Portugal}} | {{n+p|10,639,726|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 Dec 2023|abbr=on}} | 2023 estimate<ref>{{cite web|url=https://www.ine.pt/xportal/xmain?xpid=INE&xpgid=ine_destaques&DESTAQUESdest_boui=645507713&DESTAQUESmodo=2|title=População residente ultrapassa os 10,6 milhões - 2023|website=ine.pt|publisher=[[Instituto Nacional de Estatística (Portugal)|INE]]|access-date=18 June 2024}}</ref> || |- | {{nts|89}} | {{flag|Sweden}} | {{n+p|10548261 |{{worldpop}}|sigfig=1|disp=table}} || {{dts|30 April 2024|abbr=on}} | Monthly national estimate<ref>{{cite web |url=http://www.scb.se/en/finding-statistics/statistics-by-subject-area/population/population-composition/population-statistics |title=Population statistics |website=scb.se |access-date=11 April 2024 |archive-date=16 August 2019 |archive-url=https://web.archive.org/web/20190816175900/https://www.scb.se/en/finding-statistics/statistics-by-subject-area/population/population-composition/population-statistics/ |url-status=live}}</ref> || |- | {{nts|90}} | {{flag|Greece}} | {{n+p|10413982|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2023|abbr=on}} | Official estimate<ref>{{cite web |title=Estimated Population and Migration Flows, 2023 |url=https://www.statistics.gr/documents/20181/de3e26f6-9b77-d2e5-2ca3-e13bcafe482a |publisher=[[Hellenic Statistical Authority]] |access-date=8 January 2024 |location=Piraeus |date=29 December 2023}}</ref> || |- | {{nts|91}} | {{flag|United Arab Emirates}} | {{n+p|10288946|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 Dec 2022|abbr=on}} | Official estimate<ref>{{cite web |url=https://fcsa.gov.ae/ar-ae/Pages/Statistics/Statistics-by-Subject.aspx#/%3Fyear=&folder=%D8%A7%D9%84%D8%B3%D9%83%D8%A7%D9%86%D9%8A%D8%A9%20%D9%88%D8%A7%D9%84%D8%A7%D8%AC%D8%AA%D9%85%D8%A7%D8%B9%D9%8A%D8%A9/%D8%A7%D9%84%D8%B3%D9%83%D8%A7%D9%86/%D8%A7%D9%84%D8%B3%D9%83%D8%A7%D9%86%20%D9%88%D8%A7%D9%84%D8%AA%D8%B9%D8%AF%D8%A7%D8%AF%D8%A7%D8%AA%20%D8%A7%D9%84%D8%B3%D9%83%D8%A7%D9%86%D9%8A%D8%A9&subject=%D8%A7%D9%84%D8%B3%D9%83%D8%A7%D9%86%D9%8A%D8%A9%20%D9%88%D8%A7%D9%84%D8%A7%D8%AC%D8%AA%D9%85%D8%A7%D8%B9%D9%8A%D8%A9 |title=Statistics by Subject |website=fcsa.gov.ae |access-date=29 June 2024 |archive-url=https://web.archive.org/web/20200704064240/https://fcsa.gov.ae/ar-ae/Pages/Statistics/Statistics-by-Subject.aspx |archive-date=4 Jul 2020}}</ref> || |- | {{nts|92}} | {{flag|Tajikistan}} | {{n+p|10277100|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2024|abbr=on}} | Official estimate<ref>{{cite web |url=https://www.stat.tj/ |title=Агентии омори назди Президенти Ҷумҳурии Тоҷикистон / Агентии омори назди Президенти Ҷумҳурии Тоҷикистон |website=www.stat.tj |access-date=21 Jul 2024 |archive-date=13 December 2009 |archive-url=https://web.archive.org/web/20091213094447/https://www.stat.tj/ |url-status=live}}</ref> || |- | {{nts|93}} | {{flag|Azerbaijan}} | {{n+p|10151517|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2023|abbr=on}} | Monthly national estimate<ref>{{cite web |url=https://www.stat.gov.az/news/index.php?lang=az&id=5588 |website=Report |title=Azərbaycanda demoqrafik vəziyyət |date=16 Aug 2023 |access-date=3 Oct 2023 |archive-date=3 October 2023 |archive-url=https://web.archive.org/web/20231003092121/https://www.stat.gov.az/news/index.php?lang=az&id=5588 |url-status=live }}</ref> || |- | {{nts|94}} | {{flag|Israel}} | {{n+p|9907100|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 May 2024|abbr=on}} | National monthly estimate<ref>{{cite web |url = https://www.cbs.gov.il/en/Statistics/Pages/Generators/Time-Series-DataBank.aspx?level_1=2 |title = Time Series DataBank |date = 15 September 2023 |website = cbs.gov.il |access-date = 2 Jul 2024 |archive-date = 20 September 2023 |archive-url = https://web.archive.org/web/20230920122003/https://www.cbs.gov.il/en/Statistics/Pages/Generators/Time-Series-DataBank.aspx?level_1=2 |url-status = live }}</ref> || {{efn|Includes [[East Jerusalem]] and the [[Golan Heights]]. Also includes [[Israeli settlement|Israeli settlers]] within [[Area C (West Bank)|Area C]] of the [[West Bank]].}} |- | {{nts|95}} | {{flag|Honduras}} | {{n+p|9892632|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2024|abbr=on}} | National annual projection<ref>{{cite web |url=https://www.ine.gob.hn/publicaciones/Censos/Censo_2013/09Tomo-IX-Proyecciones-de-Poblacion/Cuadros%20xls/1.pdf |title=TOMO: HONDURAS: PROYECCIONES DE POBLACIÓN 2013-2050 |website=INE – Instituto Nacional de Estadística Honduras |access-date=2 July 2024 |archive-date=5 April 2023 |archive-url=https://web.archive.org/web/20230405210701/https://www.ine.gob.hn/publicaciones/Censos/Censo_2013/09Tomo-IX-Proyecciones-de-Poblacion/Cuadros%20xls/1.pdf |url-status=live }}</ref> || |- | {{nts|96}} | {{flag|Hungary}} | {{n+p|9584627|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2024|abbr=on}} | Official estimate<ref>{{cite web |url=https://www.ksh.hu/stadat_files/nep/en/nep0001.html |title=22.1.1.1. Main indicators of population and vital events|publisher=Hungarian Central Statistical Office (KSH) |website=www.ksh.hu |access-date=2 July 2024 |archive-date=15 February 2019 |archive-url=https://web.archive.org/web/20190215153106/http://www.ksh.hu/docs/eng/xstadat/xstadat_annual/i_wdsd001.html |url-status=live}}</ref> || |- | {{nts|97}} | {{flag|Austria}} | {{n+p|9170647|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 April 2024|abbr=on}} | National quarterly estimate<ref>{{cite web |url=https://www.statistik.at/en/statistics/population-and-society/population/population-stock/population-at-beginning-of-year/quarter |title=Population at beginning of year/quarter |website=www.statistik.at |access-date=10 Aug 2023 |archive-date=12 June 2015 |archive-url=https://web.archive.org/web/20150612161754/http://www.statistik.at/web_de/statistiken/menschen_und_gesellschaft/bevoelkerung/bevoelkerungsstand_und_veraenderung/bevoelkerung_zu_jahres-_quartalsanfang/023582.html |url-status=live}}</ref> || |- | {{nts|98}} | {{flag|Belarus}} | {{n+p|9155978|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2024|abbr=on}} | Official estimate<ref>{{cite web |url=http://dataportal.belstat.gov.by/Indicators/Preview?key=144299 |title=Population at the beginning of 2022 |website=belstat.gov.by |access-date=3 August 2022 |archive-date=5 January 2023 |archive-url=https://web.archive.org/web/20230105133324/http://dataportal.belstat.gov.by/Indicators/Preview?key=144299 |url-status=live}}</ref> || |- | {{nts|99}} | {{flag|Switzerland}} | {{n+p|8981565|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 March 2024|abbr=on}} | National quarterly estimate<ref>{{cite web |url=https://www.bfs.admin.ch/bfs/de/home/statistiken/kataloge-datenbanken/tabellen.assetdetail.14027789.html |title=Ständige Wohnbevölkerung nach Staatsangehörigkeitskategorie, Alter und Kanton, 2. Quartal 2020|language=de |date=22 Sep 2020|publisher=Bundesamt für Statistik |website=bfs.admin.ch |access-date=22 Nov 2019 |archive-date=30 June 2019 |archive-url=https://web.archive.org/web/20190630173058/https://www.bfs.admin.ch/bfs/de/home/statistiken/kataloge-datenbanken/tabellen.assetdetail.25673520.html |url-status=live}}</ref> || |- | {{nts|100}} | {{flag|Sierra Leone}} | {{n+p|8494260|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2022|abbr=on}} | National annual projection<ref>{{cite report|url=https://www.statistics.sl/images/StatisticsSL/Documents/Census/2015/sl_2015_phc_thematic_report_on_population_projections.pdf |title=Sierra Leone 2015 Population and Housing Census – Thematic Report on Population Projections|author=Gershon P. Y. Togoh|author2=Abu Bakarr Turay|author3=Allieu Komba |date=Oct 2017|publisher=Statistics Sierra Leone |website =www.statistics.sl |access-date=18 Apr 2020 |archive-url=https://web.archive.org/web/20191113162606/https://www.statistics.sl/images/StatisticsSL/Documents/Census/2015/sl_2015_phc_thematic_report_on_population_projections.pdf |archive-date=13 Nov 2019 |url-status=live}}</ref> || |- | {{nts|101}} | {{flag|Togo}} | {{n+p|8095498|{{worldpop}}|sigfig=1|disp=table}} || {{dts|8 Nov 2022|abbr=on}} | 2022 census result<ref>{{cite press release |url=https://inseed.tg/wp-content/uploads/2023/04/Communique_de_presse_Resultats_RGPH5_20230404.pdf |title=Resultats RGPH5 |language=fr |website=inseed.tg |access-date=20 Apr 2023 |archive-date=18 April 2023 |archive-url=https://web.archive.org/web/20230418035453/https://inseed.tg/wp-content/uploads/2023/04/Communique_de_presse_Resultats_RGPH5_20230404.pdf |url-status=live }}</ref> || |- | {{hs|end}} – | {{flag|Hong Kong}} (China) | {{n+p|7503100|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 Dec 2023|abbr=on}} | National estimate<ref>{{cite web |url=https://www.censtatd.gov.hk/hkstat/sub/so20.jsp |title=Population - Overview &#124; Census and Statistics Department |website=www.censtatd.gov.hk |access-date=21 Sep 2023 |archive-url=https://web.archive.org/web/20170425204930/https://www.censtatd.gov.hk/hkstat/sub/so20.jsp |archive-date=25 Apr 2017 |url-status=live}}</ref> || |- | {{nts|102}} | {{flag|Laos}} | {{n+p|7443000|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2022|abbr=on}} | National annual projection<ref>{{cite web |url = https://laosis.lsb.gov.la/majorIndicators.do?paramGrpId=all |title = Distribution of citizens by gender and age group |website = laosis.lsb.gov.la |access-date = 21 Sep 2023 |archive-date = 20 September 2023 |archive-url = https://web.archive.org/web/20230920122809/https://laosis.lsb.gov.la/majorIndicators.do?paramGrpId=all |url-status = live }}</ref> || |- | {{nts|103}} | {{flag|Kyrgyzstan}} | {{n+p|7100000|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Mar 2023|abbr=on}} | Monthly national estimate<ref>{{cite web |url=https://stat.gov.kg/ru/ |title=Социально-экономическое положение Кыргызской Республики |access-date=23 September 2016 |archive-date=18 September 2016 |archive-url=https://web.archive.org/web/20160918222911/http://www.stat.kg/ru/ |url-status=live }}</ref> || |- | {{nts|104}} | {{flag|Turkmenistan}} | {{n+p|7057841|{{worldpop}}|sigfig=1|disp=table}} || {{dts|17 Dec 2022|abbr=on}} | 2022 census result<ref>{{cite web |url=https://www.stat.gov.tm/ |title=Итоги сплошной переписи населения и жилищного фонда Туркменистана 2022 года |date=14 Jul 2023 |website=www.stat.gov.tm/ |access-date=21 Sep 2023 |archive-date=20 July 2023 |archive-url=https://web.archive.org/web/20230720125616/https://www.stat.gov.tm/ |url-status=live}}</ref> || |- | {{nts|105}} | {{flag|Libya}} | {{n+p|6931061|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2020|abbr=on}} | Official estimate<ref>{{cite web |url = https://www.citypopulation.de/en/libya/ |title = Libya |date = 2 Oct 2022 |website = City Population |access-date = 26 Jun 2023 |archive-date = 4 April 2023 |archive-url = https://web.archive.org/web/20230404191540/http://citypopulation.de/en/libya/ |url-status = live }}</ref> || |- | {{nts|106}} | {{flag|El Salvador}} | {{n+p|6884888|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2022|abbr=on}} | National annual projection<ref>{{cite web |url=http://www.digestyc.gob.sv/index.php/temas/des/ehpm/publicaciones-ehpm.html?download=488%3Aestimaciones-y-proyecciones-de-poblacion |title=El Salvador: Estimatciones y proyecciones de población nacional 2005–2050 departamental 2005-2025 |website=www.digestyc.gob.sv |access-date=29 Aug 2020 |archive-url=https://web.archive.org/web/20200422071030/http://www.digestyc.gob.sv/index.php/temas/des/ehpm/publicaciones-ehpm.html?download=488%3Aestimaciones-y-proyecciones-de-poblacion |archive-date=22 Apr 2020 |url-status=live}}</ref> || |- | {{nts|107}} | {{flag|Nicaragua}} | {{n+p|6803886|{{worldpop}}|sigfig=1|disp=table}} || {{dts|30 Jun 2023|abbr=on}} | Official estimate<ref>{{cite web |url=https://www.inide.gob.ni/docs/Anuarios/Anuario2022/ANUARIO_ESTADISTICO2022.pdf |title=Anuario Estadístico 2022 |website=www.inide.gob.ni |access-date=23 Jun 2024 }}</ref> || |- | {{nts|108}} | {{flag|Serbia}} | {{n+p|6605168|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2024|abbr=on}} | Official estimate<ref>{{cite web |url=https://stat.gov.rs/en-US |title=Population |publisher=Statistical Office of the Republic of Serbia (РЗС) |website=www.stat.gov.rs |access-date=2 July 2024 |language=en |archive-date=13 October 2023 |archive-url=https://web.archive.org/web/20231013224848/https://stat.gov.rs/en-US |url-status=live }}</ref>|| {{efn|Does not include [[Kosovo]].}} |- | {{nts|109}} | {{flag|Bulgaria}} | {{n+p|6445481|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 Dec 2023|abbr=on}} | Official annual estimate<ref>{{cite web |url=https://www.nsi.bg/bg/content/19806/%D0%BF%D1%80%D0%B5%D1%81%D1%81%D1%8A%D0%BE%D0%B1%D1%89%D0%B5%D0%BD%D0%B8%D0%B5/%D0%BD%D0%B0%D1%81%D0%B5%D0%BB%D0%B5%D0%BD%D0%B8%D0%B5-%D0%BA%D1%8A%D0%BC-7-%D1%81%D0%B5%D0%BF%D1%82%D0%B5%D0%BC%D0%B2%D1%80%D0%B8-2021-%D0%B3%D0%BE%D0%B4%D0%B8%D0%BD%D0%B0 |title=Население към 7 септември 2021 година |date=3 Oct 2022|language=Bulgarian |website=www.nsi.bg |access-date=21 Feb 2022 |archive-date=3 October 2022 |archive-url=https://web.archive.org/web/20221003102110/https://nsi.bg/bg/content/19806/%D0%BF%D1%80%D0%B5%D1%81%D1%81%D1%8A%D0%BE%D0%B1%D1%89%D0%B5%D0%BD%D0%B8%D0%B5/%D0%BD%D0%B0%D1%81%D0%B5%D0%BB%D0%B5%D0%BD%D0%B8%D0%B5-%D0%BA%D1%8A%D0%BC-7-%D1%81%D0%B5%D0%BF%D1%82%D0%B5%D0%BC%D0%B2%D1%80%D0%B8-2021-%D0%B3%D0%BE%D0%B4%D0%B8%D0%BD%D0%B0 |url-status=live}}</ref> || |- | {{nts|110}} | {{flag|Republic of the Congo}} | {{n+p|6142180|{{worldpop}}|sigfig=1|disp=table}} || {{dts|17 May 2023|abbr=on}} | 2023 census result<ref>{{cite web |url=https://ins-congo.cg/download/resultats-preliminaires/?wpdmdl=4068&amp;refresh=65af84d0d8a631706001616 |title= Le cinquième Recensement Général de la Population et de l'habitation du Congo (RGPH-5) / pg.VIII |publisher=INS Congo |website=www.ins-congo.cg |date=29 December 2023 |access-date=23 January 2024}}</ref> || |- | {{nts|111}} | {{flag|Paraguay}} | {{n+p|6109644|{{worldpop}}|sigfig=1|disp=table}} || {{dts|10 Nov 2022|abbr=on}} | Preliminary census result<ref>{{cite web |url=https://www.abc.com.py/nacionales/2023/08/31/paraguay-tiene-6109644-de-habitantes-segun-el-ultimo-censo-de-poblacion-y-viviendas/ |title=Paraguay has 6,109,644 inhabitants, according to the last Census |date=August 31, 2023 |website=ABC Color|language=ES |access-date=4 September 2023 |archive-date=5 September 2023 |archive-url=https://web.archive.org/web/20230905110622/https://www.abc.com.py/nacionales/2023/08/31/paraguay-tiene-6109644-de-habitantes-segun-el-ultimo-censo-de-poblacion-y-viviendas/ |url-status=live}}</ref> || |- | {{nts|112}} | {{flag|Denmark}} | {{n+p|5967824|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 April 2024|abbr=on}} | National quarterly estimate<ref>{{cite web |url=http://www.dst.dk/en/Statistik/emner/befolkning-og-valg/befolkning-og-befolkningsfremskrivning/folketal |title=Population figures |website=www.dst.dk |access-date=16 Feb 2024}}</ref> || {{efn|Does not include the [[Faroe Islands]] and [[Greenland]].}} |- | {{nts|113}} | {{flag|Singapore}} | {{n+p|5917600|{{worldpop}}|sigfig=1|disp=table}} || {{dts|30 Jun 2023|abbr=on}} | Official estimate<ref>{{cite web |url=https://www.singstat.gov.sg/find-data/search-by-theme/population/population-and-population-structure/latest-data |title=Singapore Department of Statistics (DOS) |website=www.singstat.gov.sg |access-date=30 Sep 2023 |archive-date=12 May 2018 |archive-url=https://web.archive.org/web/20180512142010/https://www.singstat.gov.sg/find-data/search-by-theme/population/population-and-population-structure/latest-data |url-status=live}}</ref> || |- | {{nts|114}} | {{flag|Central African Republic}} | {{n+p|5633412|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2020|abbr=on}} | National annual projection<ref>{{cite web |url=http://www.fao.org/3/a-ab588f.pdf |title=L'étude prospective du secteur forestier en Afrique (FOSA) République centrafricaine |website =www.fao.org |access-date=18 Apr 2020 |archive-url=https://web.archive.org/web/20190412033806/http://www.fao.org/3/a-ab588f.pdf |archive-date=12 Apr 2019 |url-status=live }}</ref> || |- | {{nts|115}} | {{flag|Finland}} | {{n+p|5583284|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 May 2024|abbr=on}} | Monthly national estimate<ref name="pxnet2.stat.fi">{{cite web |url=https://statfin.stat.fi/PxWeb/pxweb/en/StatFin/StatFin__vamuu/statfin_vamuu_pxt_11lj.px/table/tableViewLayout1/ |title=Statistics Finland - Population - Preliminary population statistics |website=statfin.stat.fi |access-date=29 June 2024 |archive-date=8 November 2023 |archive-url=https://web.archive.org/web/20231108055049/https://statfin.stat.fi/PxWeb/pxweb/en/StatFin/StatFin__vamuu/statfin_vamuu_pxt_11lj.px/ |url-status=live}}</ref> || {{efn|Does not include [[Åland]].}} |- | {{nts|116}} | {{flag|Norway}} | {{n+p|5562363|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 March 2024|abbr=on}} | National quarterly estimate<ref>{{cite web |url=https://www.ssb.no/en |title=2020-01-01 |website=ssb.no |access-date=22 May 2018 |archive-url=https://web.archive.org/web/20180526091639/https://www.ssb.no/en/ |archive-date=26 May 2018 |url-status=live}}</ref> || {{efn|Includes [[Svalbard]].}} |- | {{nts|117}} | {{flag|Lebanon}} | {{n+p|5490000|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2021|abbr=on}} | Official estimate<ref>{{cite web |url = https://www.citypopulation.de/en/lebanon/cities/ |title = Lebanese Republic |date = 2 Mar 2019 |website = City Population |access-date = 16 May 2023 |archive-date = 20 September 2023 |archive-url = https://web.archive.org/web/20230920133612/https://www.citypopulation.de/en/lebanon/cities/ |url-status = live }}</ref> || |- | {{nts|118}} | {{flag|Palestine}} | {{n+p|5483450|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2023|abbr=on}} | National annual projection<ref>{{cite web |url=https://pcbs.gov.ps/statisticsIndicatorsTables.aspx?lang=ar&table_id=676 |title=عدد السكان المقدر في فلسطين منتصف العام حسب المحافظة، 1997-2026 |website =www.pcbs.gov.ps |access-date=1 Feb 2023 |archive-url=https://web.archive.org/web/20221207112247/https://pcbs.gov.ps/statisticsIndicatorsTables.aspx?lang=ar&table_id=676 |archive-date=7 Dec 2022 |url-status=live}}</ref> || {{efn|Does not include [[East Jerusalem]] or [[Israeli settlement]]s in the [[West Bank]].}} |- | {{nts|119}} | {{flag|Slovakia}} | {{n+p|5422620|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 March 2024|abbr=on}} | National quarterly estimate<ref>{{cite web |url = https://slovak.statistics.sk/wps/portal/ext/products/informationmessages/inf_sprava_detail/!ut/p/z1/tVLNcpswGHyWHDgKfSABojfZqTExTWu7xLEuGSGErTr8BFPcvH1xxodO2pD2UF0kzexK--0uFvgei0r2Zic7U1fycbhvhf-wDGI2mTgcYJIQiG8W69v5dOZGqYc3L4BpxOc0SABYEnkQ83m6CpeEACdY_MpnkzsX4uB25SyWy2jh0Asf3lgc_o4_AhDj-u-wwEJVXdPt8bbOjnKPdIVMVSB56CwYDnVbDm70lUbHppX9swX9UXeHYc8YKShVFBWZqxAtCCDJchcp5QeMODRXOj8_3yiT4610HJUzT6JCexpRoAUKi0wjUEFIHOUz5mSvx_ldrxh3a3P-b8SQNLgAZusvLg9pNL1efZ5B_HXqsnXiuxD57wDAuQDGMn9P5g0WJivtkyptsAEIdUMWEsp88H0vJOfa8SojbIdFqwvd6tb-3g5t3Hddc_xggQW6sk_mYBqdG2nX7c6CP5H29bHD96-xeDu4HLzpshPgTW_0CafVOftHvP7HEOcvA47nOIg1356eBB_aV1ed_jHo_B_1a8o0TUtGntFh9bH41HmZtysfrhPOr65-AtbGoX4!/dz/d5/L2dBISEvZ0FBIS9nQSEh/ |title = Stock of population in the SR on 30st June 2023 |website = slovak.statistics.sk |access-date = 21 Sep 2023 |archive-date = 3 October 2023 |archive-url = https://web.archive.org/web/20231003092312/https://slovak.statistics.sk/wps/portal/ext/products/informationmessages/inf_sprava_detail/!ut/p/z1/tVLNcpswGHyWHDgKfSABojfZqTExTWu7xLEuGSGErTr8BFPcvH1xxodO2pD2UF0kzexK--0uFvgei0r2Zic7U1fycbhvhf-wDGI2mTgcYJIQiG8W69v5dOZGqYc3L4BpxOc0SABYEnkQ83m6CpeEACdY_MpnkzsX4uB25SyWy2jh0Asf3lgc_o4_AhDj-u-wwEJVXdPt8bbOjnKPdIVMVSB56CwYDnVbDm70lUbHppX9swX9UXeHYc8YKShVFBWZqxAtCCDJchcp5QeMODRXOj8_3yiT4610HJUzT6JCexpRoAUKi0wjUEFIHOUz5mSvx_ldrxh3a3P-b8SQNLgAZusvLg9pNL1efZ5B_HXqsnXiuxD57wDAuQDGMn9P5g0WJivtkyptsAEIdUMWEsp88H0vJOfa8SojbIdFqwvd6tb-3g5t3Hddc_xggQW6sk_mYBqdG2nX7c6CP5H29bHD96-xeDu4HLzpshPgTW_0CafVOftHvP7HEOcvA47nOIg1356eBB_aV1ed_jHo_B_1a8o0TUtGntFh9bH41HmZtysfrhPOr65-AtbGoX4!/dz/d5/L2dBISEvZ0FBIS9nQSEh/ |url-status = live }}</ref> || |- | {{nts|120}} | {{flag|New Zealand}} | {{n+p|5338900|{{worldpop}}|sigfig=1|disp=table}} || {{dts|30 March 2024|abbr=on}} | National quarterly estimate<ref>{{cite web |url = https://www.stats.govt.nz/indicators/population-of-nz/ |title = Estimated population of NZ |date = 16 Nov 2023 |website = stats.govt.nz |access-date = 5 Feb 2024 |archive-date = 26 April 2023 |archive-url = https://web.archive.org/web/20230426102243/https://www.stats.govt.nz/indicators/population-of-nz/ |url-status = live }}</ref> || {{efn|Does not include the two self-governing associated states of the [[Cook Islands]] and [[Niue]], and the dependent territory of [[Tokelau]].}} |- | {{nts|121}} | {{flag|Costa Rica}} | {{n+p|5309625|{{worldpop}}|sigfig=1|disp=table}} || {{dts|30 Jun 2024|abbr=on}} | National annual projection<ref>{{cite web |url=https://inec.cr/es/tematicas/listado?topics=91%252C646 |title=Estadísticas demográficas. 2011 – 2025. Proyecciones nacionales. Población total proyectada al 30 de junio por grupos de edades, según región de planificación y sexo |website=www.inec.go.cr |access-date=2 July 2024 |archive-date=20 September 2023 |archive-url=https://web.archive.org/web/20230920124541/https://inec.cr/es/tematicas/listado?topics=91%252C646 |url-status=live }}</ref> || |- | {{nts|122}} | {{flag|Ireland}} | {{n+p|5281600|{{worldpop}}|sigfig=1|disp=table}} || {{dts|01 Apr 2023|abbr=on}} | National estimate<ref>{{cite web |url=https://www.cso.ie/en/releasesandpublications/ep/p-pme/populationandmigrationestimatesapril2023/ |title=Population and Migration Estimates, April 2023 |website=www.cso.ie |date=25 September 2023 |access-date=26 Sep 2023 |archive-date=25 September 2023 |archive-url=https://web.archive.org/web/20230925143020/https://www.cso.ie/en/releasesandpublications/ep/p-pme/populationandmigrationestimatesapril2023/ |url-status=live }}</ref> || |- | {{nts|123}} | {{flag|Liberia}} | {{n+p|5248621|{{worldpop}}|sigfig=1|disp=table}} || {{dts|10 Nov 2022|abbr=on}} | 2022 census result<ref>{{cite web |url=https://liberia.un.org/en/220493-liberia-announces-provisional-results-its-5th-national-population-and-housing-census |title=Liberia announces provisional results of its 5th National Population and Housing Census |publisher=United Nations Liberia |website=www.liberia.un.org |date=23 February 2023 |access-date=14 October 2023 |archive-date=24 March 2023 |archive-url=https://web.archive.org/web/20230324043346/https://liberia.un.org/en/220493-liberia-announces-provisional-results-its-5th-national-population-and-housing-census |url-status=live }}</ref> || |- | {{nts|124}} | {{flag|Oman}} | {{n+p|5190332|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 May 2024|abbr=on}} | Monthly national estimate<ref>{{cite web |url=https://www.ncsi.gov.om/Elibrary/Pages/LibraryContentDetails.aspx?ItemID=T4YkMygGdMFA00vLVDvAEA%3d%3d |title=Monthly Statistical Bulletin: June 2024 |website=ncsi.gov.om |access-date=23 Jun 2024 }}</ref> || |- | {{nts|125}} | {{flag|Kuwait}} | {{n+p|4913271|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2024|abbr=on}} | Official estimate<ref>{{cite web |url=https://www.csb.gov.kw/Default_EN |title=Population Estimates at the beginning of 2024 |website=www.csb.gov.kw |access-date=23 Jun 2024 }}</ref> || |- | {{nts|126}} | {{flag|Mauritania}} | {{n+p|4475683|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2023|abbr=on}} | National annual projection<ref>{{cite web |url=https://ansade.mr/fr/ |title=Projections demographiques Mauritanie |website=www.ons.mr |date=17 August 2021 |access-date=21 Sep 2023 |archive-date=20 September 2023 |archive-url=https://web.archive.org/web/20230920133156/https://ansade.mr/fr/ |url-status=live }}</ref> || |- | {{nts|127}} | {{flag|Panama}} | {{n+p|4064780|{{worldpop}}|sigfig=1|disp=table}} || {{dts|8 Jan 2023|abbr=on}} | 2023 census result<ref>{{cite web |url=https://www.inec.gob.pa/archivos/P053342420231009161532Comentarios_Poblacion%20RFB%202023%20VF.pdf |title=Censos década - 2023 |publisher=[[:es:Instituto Nacional de Estadística y Censo (Panamá)|INEC de Panamá]] |website=www.inec.gob.pa |date=9 October 2023 |access-date=9 March 2023}}</ref> || |- | {{nts|128}} | {{flag|Croatia}} | {{n+p|3855641|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2022|abbr=on}} | Official estimate<ref>{{cite web |url = https://podaci.dzs.hr/2023/en/58064 |title = POPULATION ESTIMATE OF THE REPUBLIC OF CROATIA, 2022 |website = podaci.dzs.hr |date = 8 Sep 2023 |access-date = 21 Sep 2023 |archive-date = 2 October 2023 |archive-url = https://web.archive.org/web/20231002083017/https://podaci.dzs.hr/2023/en/58064 |url-status = live }}</ref> || |- | {{nts|129}} | {{flag|Eritrea}} | {{n+p|3748902|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2023|abbr=on}} | UN projection<ref name="unpop" /> || |- | {{nts|130}} | {{flag|Georgia}} | {{n+p|3694600|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2024|abbr=on}} | Official estimate<ref>{{cite web |url=http://geostat.ge/en |title=National Statistics Office of Georgia |website=www.geostat.ge |access-date=25 May 2023 |archive-date=28 May 2023 |archive-url=https://web.archive.org/web/20230528115010/https://www.geostat.ge/en |url-status=live}}</ref> || {{efn|Does not include [[Abkhazia]] (242,862, census 2011) and [[South Ossetia]] (53,559, census 2015).}} |- | {{nts|131}} | {{flag|Mongolia}} | {{n+p|3504741|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 Dec 2023|abbr=on}} | National annual projection<ref>{{cite web |url=http://www.nso.mn/ |title=Unified Database of Statistics |website=nso.mn |access-date=19 Jun 2023 |archive-date=27 September 2018 |archive-url=https://web.archive.org/web/20180927122656/http://www.nso.mn/ |url-status=live }}</ref> || |- | {{nts|132}} | {{flag|Uruguay}} | {{n+p|3444263|{{worldpop}}|sigfig=1|disp=table}} || {{dts|27 Nov 2023|abbr=on}} | 2023 National census<ref>{{cite web |url=http://www.ine.gub.uy/web/guest/estimaciones-y-proyecciones |title=Estimaciones y Proyecciones - Instituto Nacional de Estadística |website=www.ine.gub.uy |access-date=10 Aug 2023 |archive-url=https://web.archive.org/web/20190322200708/http://www.ine.gub.uy/web/guest/estimaciones-y-proyecciones |archive-date=22 Mar 2019 |url-status=live}}</ref> || |- | {{nts|133}} | {{flag|Bosnia and Herzegovina}} | {{n+p|3264873|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2023|abbr=on}} | Official estimate<ref>{{cite web |url=https://www.rzs.rs.ba/ |title=Republika Srpska Institute of Statistics |website=rzs.rs.ba |access-date=19 Jun 2023 |archive-date=16 September 2020 |archive-url=https://web.archive.org/web/20200916034658/https://rzs.rs.ba/ |url-status=live }}</ref><ref>{{cite web |url=https://fzs.ba/index.php/procjena-broja-stanovnika/ |title=Population estimation |website=fzs.ba |access-date=19 Jun 2023 |archive-date=19 June 2023 |archive-url=https://web.archive.org/web/20230619201152/https://fzs.ba/index.php/procjena-broja-stanovnika/ |url-status=live }}</ref> || |- | {{hs|end}} – | {{flag|Puerto Rico}} (US) | {{n+p|3205691|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2023|abbr=on}} | Annual projection<ref name="census">{{cite web |url=https://www.census.gov/quickfacts/PR |publisher=[[United States Census Bureau]] |title=Quick Facts Puerto Rico |date=18 May 2023 |access-date=19 Jun 2023 |archive-date=30 March 2021 |archive-url=https://web.archive.org/web/20210330054937/https://www.census.gov/quickfacts/PR |url-status=live }}</ref> || |- | {{nts|134}} | {{flag|Armenia}} | {{n+p|3039700|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2024|abbr=on}} | National quarterly estimate<ref>{{cite web |url=https://www.armstat.am/en/?nid=12&id=19001&submit=Search |title=Time series / Statistical Committee of the Republic of Armenia |website=www.armstat.am |access-date=23 Jun 2024 |archive-url=https://web.archive.org/web/20190403053506/https://www.armstat.am/en/?nid=12&id=19001&submit=Search |archive-date=3 Apr 2019 |url-status=live}}</ref> || |- | {{nts|135}} | {{flag|Namibia}} | {{n+p|3022401|{{worldpop}}|sigfig=1|disp=table}} || {{dts|24 Sept 2023|abbr=on}} | 2023 National census<ref>{{cite web |url=https://census.nsa.org.na/ |title=Namibia - Census (2023) |publisher=Namibia Statistics Agency |website=www.census.nsa.org.na |access-date=18 Apr 2024}}</ref> || |- | {{nts|136}} | {{flag|Lithuania}} | {{n+p|2891215|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 June 2024|abbr=on}} | Monthly national estimate<ref>{{cite web |url=https://osp.stat.gov.lt/ |title=Pradžia - Oficialiosios statistikos portalas |website=osp.stat.gov.lt |access-date=21 Sep 2023 |archive-date=20 March 2015 |archive-url=https://web.archive.org/web/20150320060836/http://osp.stat.gov.lt/ |url-status=live}}</ref> || |- | {{nts|137}} | {{flag|Qatar}} | {{n+p|2857822|{{worldpop}}|sigfig=1|disp=table}} || {{dts|30 Jun 2024|abbr=on}} | Monthly national estimate<ref>{{cite web |url=https://www.psa.gov.qa/en/statistics1/pages/topicslisting.aspx?parent=Population&child=Population |title=Population |website=psa.gov.qa |access-date=3 Jul 2024 |archive-date=19 June 2023 |archive-url=https://web.archive.org/web/20230619201210/https://www.psa.gov.qa/en/statistics1/pages/topicslisting.aspx?parent=Population&child=Population |url-status=live }}</ref> || |- | {{nts|138}} | {{flag|Jamaica}} | {{n+p|2825544|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2019|abbr=on}} | National projection<ref>{{cite web |url=https://statinja.gov.jm/Demo_SocialStats/PopulationStats.aspx |title=Population Statistics |access-date=19 Jun 2023 |website=statinja.gov.jm |archive-date=6 March 2023 |archive-url=https://web.archive.org/web/20230306205719/https://statinja.gov.jm/Demo_SocialStats/PopulationStats.aspx |url-status=live }}</ref> || |- | {{nts|139}} | {{flag|Albania}} | {{n+p|2761785|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2023|abbr=on}} | Official estimate<ref>{{cite web |url=https://www.instat.gov.al/en/themes/demography-and-social-indicators/population/publication/2023/population-of-albania-on-1st-january-2023/ |title=Population in Albania, 1 Jan 2023 |access-date=10 Aug 2023 |archive-date=20 September 2023 |archive-url=https://web.archive.org/web/20230920124542/https://www.instat.gov.al/en/themes/demography-and-social-indicators/population/publication/2023/population-of-albania-on-1st-january-2023/ |url-status=live}}</ref> || |- | {{nts|140}} | {{flag|Moldova}} | {{n+p|2423300|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2024|abbr=on}} | Official estimate<ref>{{cite web |url=https://statistica.gov.md/en/statistic_indicator_details/25 |title=Population |publisher=Biroului Național de Statistică (BNS) |website=statistica.gov.md |access-date=23 Jun 2024 }}</ref> || {{efn|Does not include [[Transnistria]].}} |- | {{nts|141}} | {{flag|Gambia}} | {{n+p|2417471|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2022|abbr=on}} | National projection<ref>{{cite web |url=https://www.gbosdata.org/data/40-population-and-household-characteristics/1653-population-total |title=Population - Total |access-date=25 May 2023 |archive-date=11 May 2023 |archive-url=https://web.archive.org/web/20230511134924/https://www.gbosdata.org/data/40-population-and-household-characteristics/1653-population-total |url-status=live}}</ref> || |- | {{nts|142}} | {{flag|Botswana}} | {{n+p|2410338|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2021|abbr=on}} | National annual projection<ref>{{cite web |url=http://www.statsbots.org.bw/sites/default/files/publications/Botswana%20Population%20Projections%202011_2026.pdf |title=Botswana Population Projections 2011-2026 |website=www.statsbots.org.bw |access-date=20 Jan 2019 |archive-url=https://web.archive.org/web/20190412033845/http://www.statsbots.org.bw/sites/default/files/publications/Botswana%20Population%20Projections%202011_2026.pdf |archive-date=12 Apr 2019 |url-status=live }}</ref> || |- | {{nts|143}} | {{flag|Gabon}} | {{n+p|2408586|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2024|abbr=on}} | National annual projection<ref>{{cite web |url=https://www.statgabon.ga/wp-content/uploads/2020/08/Rapport-Projection_Final.pdf |title=Projection de la Population Gabon 2019-2025 |website=www.statgabon.ga |access-date=3 Jul 2024 |archive-url=https://web.archive.org/web/20210429022734/https://www.statgabon.ga/wp-content/uploads/2020/08/Rapport-Projection_Final.pdf |archive-date=29 Apr 2021 |url-status=live}}</ref> || |- | {{nts|144}} | {{flag|Lesotho}} | {{n+p|2306000|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2023|abbr=on}} | Official projection<ref>{{cite web |url = https://www.citypopulation.de/en/lesotho/ |title = Lesotho |date = 8 Apr 2023 |website = City Population |access-date = 26 Jun 2023 |archive-date = 19 May 2023 |archive-url = https://web.archive.org/web/20230519091617/https://citypopulation.de/en/lesotho/ |url-status = live }}</ref> || |- | {{nts|145}} | {{flag|Slovenia}} | {{n+p|2123949|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2024|abbr=on}} | National quarterly estimate<ref>{{cite web |url=https://www.stat.si/statweb |title=Prebivalci Slovenije |website=www.stat.si |access-date=23 Jun 2024 }}</ref> || |- | {{nts|146}} | {{flag|Latvia}} | {{n+p|1862700|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 May 2024|abbr=on}} | Monthly national estimate<ref>{{cite web |url=https://data.stat.gov.lv/pxweb/en/OSP_PUB/START__POP__IR__IRS/IRS010m/ |title=IRS010m. Population and main dataSw of vital statistics 1995M01 - 2023M12 |website=Statistikas datubāzes |access-date=5 Feb 2024 |archive-date=12 May 2021 |archive-url=https://web.archive.org/web/20210512140058/http://data.stat.gov.lv/pxweb/en/OSP_PUB/START__POP__IR__IRS/IRS010m |url-status=live}}</ref> || |- | {{nts|147}} | {{flag|North Macedonia}} | {{n+p|1829954|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 Dec 2022|abbr=on}} | Official estimate<ref>{{cite web |url=https://www.stat.gov.mk/OblastOpsto_en.aspx?id=2 |title=Population |website= State Statistical Office |access-date=23 Jun 2024 }}</ref> || |- | {{nts|148}} | {{flag|Guinea-Bissau}} | {{n+p|1781308|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2023|abbr=on}} | National annual projection<ref>{{cite web |url = https://www.stat-guinebissau.com/ |title = National Institute of Statistics of Guinea-Bissau |website = stat-guinebissau |access-date = 4 Jul 2023 |archive-date = 21 July 2023 |archive-url = https://web.archive.org/web/20230721231946/https://www.stat-guinebissau.com/ |url-status = live }}</ref> || |- | {{hs|end}} – | {{flag|Kosovo}} | {{n+p|1762220|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 Dec 2022|abbr=on}} | Official estimate<ref>{{cite web |url=https://askdata.rks-gov.net/pxweb/sq/ASKdata/ASKdata__Population__Estimate,%20projection%20and%20structure%20of%20population__Population%20estimate/tab001.px/table/tableViewLayout1/?loadedQueryId=c54f1063-6282-4066-9578-89393f22477b&timeType=top&timeValue=1 |title=Popullsia e Kosovës e vlerësuar sipas indikatoret dhe komuna |website=ask.rks-gov.net |access-date=21 Sep 2023 |archive-date=8 November 2023 |archive-url=https://web.archive.org/web/20231108055001/https://askdata.rks-gov.net/pxweb/sq/ASKdata/ASKdata__Population__Estimate,%20projection%20and%20structure%20of%20population__Population%20estimate/tab001.px/ |url-status=live}}</ref> || |- | {{nts|149}} | {{flag|Bahrain}} | {{n+p|1577059|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2023|abbr=on}} | Official estimate<ref>{{cite web |url = https://bahrain.opendatasoft.com/explore/dataset/03-mid-year-population-by-nationality-and-sex/table/?disjunctive.sex&sort=-year |title = Mid-Year Population |date = 10 Jun 2023 |website = bahrain.opendatasoft.com |access-date = 12 Sep 2023 |archive-date = 1 September 2023 |archive-url = https://web.archive.org/web/20230901001907/https://bahrain.opendatasoft.com/explore/dataset/03-mid-year-population-by-nationality-and-sex/table/?disjunctive.sex&sort=-year |url-status = live }}</ref> || |- | {{nts|150}} | {{flag|Equatorial Guinea}} | {{n+p|1558160|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2022|abbr=on}} | Official estimate<ref>{{cite web |url=http://www.inege.gq/ |title=INEGE |website=www.inege.gq |access-date=25 May 2023 |archive-url=https://web.archive.org/web/20170420050225/http://www.inege.gq/ |archive-date=20 Apr 2017 |url-status=live}}</ref> || |- | {{nts|151}} | {{flag|Estonia}} | {{n+p|1374687|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2024|abbr=on}} | Official estimate<ref>{{cite web|last=|first= |date= |title=Population figure|url=https://www.stat.ee/et |access-date=20 April 2023 |archive-date=1 April 2023 |archive-url=https://web.archive.org/web/20230401052333/https://www.stat.ee/et/ |url-status=live}}</ref> || |- | {{nts|152}} | {{flag|East Timor}} | {{n+p|1373024|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2024|abbr=on}} | National annual projection<ref>{{cite web |url=https://timor-leste.unfpa.org/sites/default/files/pub-pdf/2015%20Population%20Projection%20Report_0.pdf |title=Analytical Report on Population Projection |website=timor-leste.unfpa.org |access-date=3 Jul 2024 |archive-url=https://web.archive.org/web/20200802031357/https://timor-leste.unfpa.org/sites/default/files/pub-pdf/2015%20Population%20Projection%20Report_0.pdf |archive-date=2 Aug 2020 |url-status=live}}</ref> || |- | {{nts|153}} | {{flag|Trinidad and Tobago}} | {{n+p|1367510|{{worldpop}}|sigfig=1|disp=table}} || {{dts|30 Jun 2023|abbr=on}} | Official estimate<ref>{{cite web |url=https://cso.gov.tt/subjects/population-and-vital-statistics/population/ |title=Population |website=cso.gov.tt |access-date=5 Feb 2024 |archive-url=https://web.archive.org/web/20191206192934/https://cso.gov.tt/subjects/population-and-vital-statistics/population/ |archive-date=6 Dec 2019 |url-status=live}}</ref> || |- | {{nts|154}} | {{flag|Mauritius}} | {{n+p|1260379|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 Dec 2023|abbr=on}} | National estimate<ref>{{cite web |url=https://statsmauritius.govmu.org/Pages/Statistics/By_Subject/Population/SB_Population.aspx|publisher=Statistics Mauritius |title=Population and Vital Statistics |date=2023 |website=statsmauritius.govmu.org |access-date=23 Jun 2024 }}</ref> || |- | {{nts|155}} | {{flag|Eswatini}} | {{n+p|1235549|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2024|abbr=on}} | Official projection<ref>{{cite web |url = http://ccm.org.sz/documents/reports/hiv/Population-Projections-Report-2007-2030.pdf |title =Swaziland Population Projections 2007-2030 |at = p20 |date = 1 Jul 2015 |website = ccm.org.sz |access-date = 3 Jul 2024 |archive-url = https://web.archive.org/web/20150723224521/http://ccm.org.sz/documents/reports/hiv/Population-Projections-Report-2007-2030.pdf |archive-date = 23 Jul 2015 |url-status = dead}}</ref> || |- | {{nts|156}} | {{flag|Djibouti}} | {{n+p|1001454|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2022|abbr=on}} | National projection <ref>{{cite web |title=Annuaire Statistique Edition 2022 |url=http://www.instad.dj/assets/doc/Annuaire_Statistique_2022.pdf |website=www.instad.dj |access-date=13 Feb 2023 |archive-url=https://web.archive.org/web/20221013191837/http://www.instad.dj/assets/doc/Annuaire_Statistique_2022.pdf |archive-date=13 Oct 2022 |url-status=live}}</ref> || |- | {{nts|157}} | {{flag|Cyprus}} | {{n+p|918100|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Oct 2021|abbr=on}} | 2021 census result<ref>{{cite web |url = https://www.cystat.gov.cy/el/PressRelease?id=66207 |title = Census of Population and Housing 2021 |date = 26 Nov 2021 |website = cystat.gov.cy |access-date = 29 Jun 2023 |archive-date = 6 August 2022 |archive-url = https://web.archive.org/web/20220806221137/https://www.cystat.gov.cy/el/PressRelease?id=66207 |url-status = live }}</ref> || {{efn|Des not include [[Northern Cyprus]].}} |- | {{nts|158}} | {{flag|Fiji}} | {{n+p|893468|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2021|abbr=on}} | National annual projection<ref>{{cite web |url=https://www.statsfiji.gov.fj/ |title=BULA and Welcome |website=www.statsfiji.gov.fj |access-date=27 Apr 2023 |archive-date=8 April 2020 |archive-url=https://web.archive.org/web/20200408072353/https://www.statsfiji.gov.fj/ |url-status=live }}</ref> || |- | {{nts|159}} | {{flag|Bhutan}} | {{n+p|777224|{{worldpop}}|sigfig=1|disp=table}} || {{dts|2024|abbr=on}} | National annual projection<ref>{{cite web |url=http://www.nsb.gov.bt/publication/files/pub2wc8278bd.pdf |title=Population Projections Bhutan 2017-2047 |website=www.nsb.gov.bt |access-date=23 Sep 2023 |archive-url=https://web.archive.org/web/20190128191433/http://www.nsb.gov.bt/publication/files/pub2wc8278bd.pdf |archive-date=28 Jan 2019 |url-status=live }}</ref> || |- | {{nts|160}} | {{flag|Guyana}} | {{n+p|772975|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2021|abbr=on}} | Official estimate<ref>{{cite web |url=https://statisticsguyana.gov.gy/subjects/demography-vital-and-social-statistics/year-end-and-mid-year-population-estimates-by-sex-guyana-1992-to-2021/ |title=Year-end and Mid-year Population Estimates by Sex, Guyana: 1992 to 2021 |access-date=23 Jun 2024 }}</ref> || |- | {{nts|161}} | {{flag|Comoros}} | {{n+p|758316|{{worldpop}}|sigfig=1|disp=table}} || {{dts|15 Dec 2017|abbr=on}} | 2017 census result<ref>{{cite web |url=http://www.inseed.km/index.php/publications/rapport1/rapports-rgph-2017 |title=Rapports RGPH 2017 |website=www.inseed.km |access-date=31 Aug 2020 |archive-url=https://web.archive.org/web/20191030105743/http://www.inseed.km/index.php/publications/rapport1/rapports-rgph-2017 |archive-date=30 Oct 2019 |url-status=live }}</ref> || |- | {{nts|162}} | {{flag|Solomon Islands}} | {{n+p|750325|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2024|abbr=on}} | National annual projection<ref>{{cite web |url = https://www.statistics.gov.sb/statistics/social-statistics/population |title = Population |website = www.statistics.gov.sb |access-date = 3 Jul 2024 |archive-date = 30 June 2022 |archive-url = https://web.archive.org/web/20220630191359/https://statistics.gov.sb/statistics/social-statistics/population |url-status = live }}</ref> || |- | {{hs|end}} – | {{flag|Macau}} (China) | {{n+p|686400|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 March 2024|abbr=on}} | National quarterly estimate<ref>{{cite web |url=https://www.dsec.gov.mo/ts/#!/step2/KeyIndicator/en-US/240 |title=DSEC - 統計數據庫 |website=www.dsec.gov.mo |access-date=21 Sep 2023 |archive-date=13 November 2020 |archive-url=https://web.archive.org/web/20201113211800/https://www.dsec.gov.mo/ts/#!/step2/KeyIndicator/en-US/240 |url-status=live}}</ref> || |- | {{nts|163}} | {{flag|Luxembourg}} | {{n+p|672020|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2024|abbr=on}} | Official estimate<ref>{{cite web |url=https://statistiques.public.lu/dam-assets/actualite/2024/stn16-pop-2024/stn16-2024-population-2024-v20.pdf |title=Une croissance démographique réduite en 2023 |date=2024-04-18 |website=statistiques.public.lu |language=fr}}</ref> || |- | {{nts|164}} | {{flag|Montenegro}} | {{n+p|616695|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2023|abbr=on}} | National annual estimate<ref>{{cite web |url=http://www.monstat.org/eng/page.php?id=234&pageid=48 |title=Statistical Office of Montenegro - MONSTAT |website=www.monstat.org |access-date=10 Aug 2023 |archive-url=https://web.archive.org/web/20170712071239/http://www.monstat.org/eng/page.php?id=234&pageid=48 |archive-date=12 Jul 2017 |url-status=live}}</ref> || |- | {{nts|165}} | {{flag|Suriname}} | {{n+p|616500|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2021|abbr=on}} | Official estimate<ref>{{cite web |url=https://statistics-suriname.org/en/population-statistics-2/ |title=Population Statistics |date=15 Aug 2019|publisher=Statistics-Suriname.org |access-date=20 Apr 2023 |archive-url=https://web.archive.org/web/20200328181226/https://statistics-suriname.org/en/population-statistics-2/ |archive-date=28 Mar 2020 |url-status=live}}</ref> || |- | {{hs|end}} – | {{flagicon image|Flag of the Sahrawi Arab Democratic Republic.svg}} {{Disputed inline |date=January 2024}} [[Western Sahara]] | {{n+p|587259|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2023|abbr=on}} | UN projection<ref name="unpop" /> || {{efn|{{Western Sahara-note}}}} |- | {{nts|166}} | {{flag|Malta}} | {{n+p|542051|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 Dec 2022|abbr=on}} | Official estimate<ref>{{cite web |url = https://nso.gov.mt/intercensal-population-revisions-2012-2021/ |title = Population and migration: 2012-2022 |date = 26 Jan 2024 |website = nso.gov.mt |access-date = 23 Jun 2024 }}</ref> || |- | {{nts|167}} | {{flag|Maldives}} | {{n+p|515132|{{worldpop}}|sigfig=1|disp=table}} || {{dts|13 Sep 2022|abbr=on}} | 2022 census result<ref>{{cite web |url = https://census.gov.mv/2022/wp-content/uploads/2023/07/WPD2023.pdf |title = Census 2022: Combined Summary |date = 11 Jul 2022 |website = census.gov.my |access-date = 14 Mar 2024}}</ref> || |- | {{nts|168}} | {{flag|Cape Verde}} | {{n+p|491233|{{worldpop}}|sigfig=1|disp=table}} || {{dts|16 Jun 2021|abbr=on}} | 2021 census result<ref>{{cite web |url=https://ine.cv/ine_censos_quadros_category/censo-2021/ |title=CENSO 2021 |publisher=ine.cv |access-date=23 Sep 2023 |archive-date=27 March 2023 |archive-url=https://web.archive.org/web/20230327051422/https://ine.cv/ine_censos_quadros_category/censo-2021/ |url-status=live }}</ref> || |- | {{nts|169}} | {{flag|Brunei}} | {{n+p|445400|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2022|abbr=on}} | Official estimate<ref>{{cite web |url=https://deps.mofe.gov.bn/SitePages/National%20Statistics.aspx |title=National Statistics |website=www.deps.mofe.gov.bn |access-date=23 Sep 2023 |archive-date=1 August 2023 |archive-url=https://web.archive.org/web/20230801041429/https://deps.mofe.gov.bn/SitePages/National%20Statistics.aspx |url-status=live}}</ref> || |- | {{nts|170}} | {{flag|Belize}} | {{n+p|397483|{{worldpop}}|sigfig=1|disp=table}} || {{dts|12 May 2022|abbr=on}} | 2022 census result<ref>{{cite web |title=Statistical Institute of Belize Presents Key Finding of the 2022 Population and Housing Census |website=Statistical Institute of Belize|date=8 April 2024 |url=https://sib.org.bz/press-release_census-launch/ |access-date=10 Apr 2024 |archive-url=https://web.archive.org/web/20240409065601/https://sib.org.bz/press-release_census-launch/ |archive-date=9 Apr 2024 |url-status=live}}</ref> || |- | {{nts|171}} | {{flag|Bahamas}} | {{n+p|397360|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2022|abbr=on}} | National annual projection| <ref>{{cite web |url=http://www.bahamas.gov.bs/wps/wcm/connect/22f9b2b0-68fa-4a26-8bd8-474952e42dc2/Population+Projection+Report+2010-2040.pdf?MOD=AJPERES |title=Population Projections 2010-2040 |website=www.bahamas.gov.bs |access-date=27 Apr 2023 |archive-url=https://web.archive.org/web/20160701011446/http://www.bahamas.gov.bs/wps/wcm/connect/22f9b2b0-68fa-4a26-8bd8-474952e42dc2/Populati396,960on+Projection+Report+2010-2040.pdf?MOD=AJPERES |archive-date=1 Jul 2016 |url-status=live }}</ref> || |- | {{nts|172}} | {{flag|Iceland}} | {{n+p|383726|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2024|abbr=on}} | National quarterly estimate<ref>{{cite web |url=https://statice.is/statistics/population/inhabitants/ |title=Statistics - Population |publisher=Statistics Iceland (Hagstofa Íslands) |website=www.statice.is |date=21 March 2024 |access-date=18 Apr 2024}}</ref> || |- | {{hs|end}} – | {{flag|Northern Cyprus}} | {{n+p|382836|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 Dec 2020|abbr=on}} | Official estimate<ref>{{cite web |url=https://www.devplan.org/belediyeler/Yerel-Yonetimler-2018-2020.pdf |title=2020 yılı: Yerel Yönetimler Raporu|language=tr|publisher=Devplan |date=2022 |access-date=15 October 2022 |archive-date=13 October 2022 |archive-url=https://web.archive.org/web/20221013104218/https://www.devplan.org/belediyeler/Yerel-Yonetimler-2018-2020.pdf |url-status=live}}</ref> || {{efn|''De facto'' independent, ''de jure'' part of [[Cyprus]].}} |- | {{hs|end}} – | {{flag|Transnistria}} | {{n+p|360938|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 Dec 2022|abbr=on}} | Official estimate<ref>{{cite news|url=https://gov.md/ro/content/352-de-mii-de-locuitori-din-regiunea-transnistreana-detin-cetatenia-republicii-moldova-si |title=352 de mii de locuitori din regiunea Transnistreană dețin cetățenia Republicii Moldova și 362 de mii figurează în registrul de stat al populației |website=gov.md |date=20 Jan 2023 |access-date=12 February 2023 |archive-date=6 March 2023 |archive-url=https://web.archive.org/web/20230306235144/https://gov.md/ro/content/352-de-mii-de-locuitori-din-regiunea-transnistreana-detin-cetatenia-republicii-moldova-si |url-status=live}}</ref> || {{efn|''De facto'' independent, ''de jure'' part of [[Moldova]].}} |- | {{nts|173}} | {{flag|Vanuatu}} | {{n+p|301295|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2021|abbr=on}} | National annual projection<ref name = "pacificdata">{{cite web |url = https://stats.pacificdata.org/ |title = Pacific Data Hub Explorer |website = stats.pacificdata.org |access-date = 20 Jul 2023 |archive-date = 21 July 2023 |archive-url = https://web.archive.org/web/20230721002349/https://stats.pacificdata.org/ |url-status = live }}</ref> || |- | {{hs|end}} – | {{flag|French Polynesia}} (France) | {{n+p|279890|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2021|abbr=on}} | National annual projection<ref name="pacificdata" /> || |- | {{hs|end}} – | {{flag|New Caledonia}} (France) | {{n+p|268510|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2023|abbr=on}} | National annual projection<ref>{{cite web |url=https://www.isee.nc/population/demographie |title=Bilan démographique 2022 : la Nouvelle-Calédonie perd 1 300 habitants |website=www.isee.nc/ |access-date=30 Sep 2023 |archive-date=13 November 2022 |archive-url=https://web.archive.org/web/20221113162110/https://www.isee.nc/population/demographie |url-status=live}}</ref> || |- | {{nts|174}} | {{flag|Barbados}} | {{n+p|267800|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 Dec 2022|abbr=on}} | Official estimate| <ref>{{cite web |url=https://stats.gov.bb/subjects/social-demographic-statistics/population-demography-statistics/vital-statistics-indicators-2022/ |title=Vital Statistic Indicators |access-date=15 Jun 2023 |archive-date=11 June 2023 |archive-url=https://web.archive.org/web/20230611031804/https://stats.gov.bb/subjects/social-demographic-statistics/population-demography-statistics/vital-statistics-indicators-2022/ |url-status=live}}</ref> || |- | {{hs|end}} – | {{flag|Abkhazia}} | {{n+p|244236|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2022|abbr=on}} | Official estimate<ref>{{cite web |url = https://ugsra.org/ofitsialnaya-statistika.php?ELEMENT_ID=600 |title = Управление государственной статистики Республики Абхазия |website = ugsra.org |access-date = 27 Oct 2023 |archive-date = 29 March 2023 |archive-url = https://web.archive.org/web/20230329170012/https://ugsra.org/ofitsialnaya-statistika.php?ELEMENT_ID=600 |url-status = live }}</ref> || {{efn|{{Abkhazia note}}}} |- | {{nts|175}} | {{flag|São Tomé and Príncipe}} | {{n+p|228319|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2024|abbr=on}} | National annual projection<ref>{{cite web |url=https://www.ine.st/ |title=Instituto Nacional de Estatística |website=ine.st |access-date=23 Jun 2024 }}</ref> || |- | {{nts|176}} | {{flag|Samoa}} | {{n+p|205557|{{worldpop}}|sigfig=1|disp=table}} || {{dts|6 Nov 2021|abbr=on}} | Census 2021<ref>{{cite web |url=https://www.sbs.gov.ws/ |title=POPULATION & HOUSING CENSUS 2016 - 2021 |website=www.sbs.gov.ws |access-date=27 Apr 2023 |archive-date=22 March 2020 |archive-url=https://web.archive.org/web/20200322144201/https://www.sbs.gov.ws/ |url-status=live}}</ref> || |- | {{nts|177}} | {{flag|Saint Lucia}} | {{n+p|184100|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2023|abbr=on}} | Official estimate<ref>{{cite web |url=https://www.stats.gov.lc/ |title=Home |website=The Central Statistical Office of Saint Lucia |access-date=23 Jun 2024 }}</ref> || |- | {{hs|end}} – | {{flag|Guam}} (US) | {{n+p|153836|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Apr 2020|abbr=on}} | 2020 census result<ref>{{cite web |url=https://www2.census.gov/programs-surveys/decennial/2020/data/island-areas/guam/population-and-housing-unit-counts/guam-phc-table01.pdf |title=Population of Guam: 2010 and 2020 |website=www2.census.gov |access-date=29 Jan 2022 |archive-date=3 November 2021 |archive-url=https://web.archive.org/web/20211103005501/https://www2.census.gov/programs-surveys/decennial/2020/data/island-areas/guam/population-and-housing-unit-counts/guam-phc-table01.pdf |url-status=live}}</ref> || |- | {{hs|end}} – | {{flag|Curacao}} (Netherlands) | {{n+p|155826 |{{worldpop}}|sigfig=1|disp=table}} || {{dts|2 Sep 2023|abbr=on}} | 2023 census result<ref>{{cite web |url=https://senso.cbs.cw/cbs-presents-the-first-results-of-the-2023-census?language_content_entity=pap-cw |title=CBS Presents the First Results of the 2023 Census |access-date=16 July 2024 }}</ref>|| |- | {{nts|178}} | {{flag|Kiribati}} | {{n+p|120740|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2021|abbr=on}} | National annual projection<ref name="pacificdata" /> || |- | {{nts|179}} | {{flag|Seychelles}} | {{n+p|120581|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 Dec 2023|abbr=on}} | Official estimate<ref>{{cite web |url=https://www.nbs.gov.sc/downloads/1547-end-of-year-2023-population/download |title=POPULATION AND VITAL STATISTICS DECEMBER 2023 |website=www.nbs.gov.sc |access-date=23 Jun 2024 }}</ref> || |- | {{nts|180}} | {{flag|Grenada}} | {{n+p|112579|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2019|abbr=on}} | Official estimate<ref>{{cite web |url = https://stats.gov.gd/subjects/population-2/annual-population-estimates-by-parish-and-sex-2011-to-2019/ |title = Annual Population Estimates by Parish and Sex, 2011 to 2019 |website = stats.gov.gd |access-date = Jun 24, 2023 | archive-url=https://web.archive.org/web/20230624211753/https://stats.gov.gd/subjects/population-2/annual-population-estimates-by-parish-and-sex-2011-to-2019/| archive-date=Jun 24, 2023}}</ref> || |- | {{nts|181}} | {{flag|Saint Vincent and the Grenadines}} | {{n+p|110872 |{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2022|abbr=on}} | Official estimate<ref>{{cite web |date= |title=Mid Year Total Population Estimates by Age and Sex, 2018 to 2022 |url=https://stats.gov.vc/subjects/population-and-demography/mid-year-total-population-estimates-by-age-and-sex/ |archive-url=https://web.archive.org/web/20230422031423/https://stats.gov.vc/subjects/population-and-demography/mid-year-total-population-estimates-by-age-and-sex/ |archive-date=22 Apr 2023 |access-date=2023-06-25 |website=stats.gov.vc |publisher=Statistical Office, Government of Saint Vincent and the Grenadines |language=en-US}}</ref>|| |- | {{hs|end}} – | {{flag|Aruba}} (Netherlands) | {{n+p|107151|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 Dec 2022|abbr=on}} | Official estimate<ref>{{cite web |url=https://cbs.aw/wp/index.php/2022/09/20/test-2/ |title=General characteristics of Aruba |website=cbs.aw |date=20 September 2022 |publisher=Central Bureau of Statistics Aruba |access-date=23 Jun 2024 }}</ref>|| |- | {{nts|182}} | {{flag|Micronesia}} | {{n+p|105754|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2021|abbr=on}} | National annual projection<ref name="pacificdata" /> || |- | {{nts|183}} | {{flag|Antigua and Barbuda}} | {{n+p|103603|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2024|abbr=on}} | National annual estimate<ref>{{cite web |title=National Annual estimate - Antigua and Barbuda; 2022 |url=https://statistics.gov.ag/ |url-status=live |archive-url=https://web.archive.org/web/20210111065316/https://statistics.gov.ag/wp-content/uploads/2017/12/Population-Projections-by-Age-Group-Annual-1991-2026.xlsx |archive-date=11 Jan 2021 |access-date=2 Jul 2024 |website=statistics.gov.ag}}</ref>|| |- | {{hs|end}} – | {{flag|Jersey}} (UK) | {{n+p|103267|{{worldpop}}|sigfig=1|disp=table}} || {{dts|21 Mar 2021|abbr=on}} | 2021 census result<ref>{{cite web |url=https://www.gov.je/SiteCollectionDocuments/Government%20and%20administration/R%20CensusBulletin1%2020220413%20SJ.pdf |title=Bulletin 1: Population characteristics|first=States of|last=Jersey |website=gov.je |access-date=23 Sep 2023 |archive-date=13 April 2022 |archive-url=https://web.archive.org/web/20220413165711/https://www.gov.je/SiteCollectionDocuments/Government%20and%20administration/R%20CensusBulletin1%2020220413%20SJ.pdf |url-status=live}}</ref> || |- | {{nts|184}} | {{flag|Tonga}} | {{n+p|100179|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2022|abbr=on}} | National annual estimate<ref>{{cite web |url=https://tongastats.gov.to/ |title=Tonga Statistics Department &#124; The official statistics provider for Tonga |access-date=16 November 2022 |archive-date=25 September 2021 |archive-url=https://web.archive.org/web/20210925121044/https://tongastats.gov.to/ |url-status=live}}</ref>|| |- | {{hs|end}} – | {{flag|U.S. Virgin Islands}} (US) | {{n+p|87146|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Apr 2020|abbr=on}} | 2020 census result<ref>{{cite web |url=https://www2.census.gov/programs-surveys/decennial/2020/data/island-areas/us-virgin-islands/population-and-housing-unit-counts/us-virgin-islands-phc-table01.pdf |title=Population of the United States Virgin Islands: 2010 and 2020 |website=www2.census.gov |access-date=29 Jan 2022 |archive-date=14 March 2023 |archive-url=https://web.archive.org/web/20230314154015/https://www2.census.gov/programs-surveys/decennial/2020/data/island-areas/us-virgin-islands/population-and-housing-unit-counts/us-virgin-islands-phc-table01.pdf |url-status=live}}</ref> || |- | {{nts|185}} | {{flag|Andorra}} | {{n+p|85101|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 Dec 2023|abbr=on}} | Official estimate<ref>{{cite web |url=https://www.estadistica.ad/serveiestudis/web/index.asp |title=Departament d'Estadística |website=www.estadistica.ad |access-date=5 Feb 2024 |archive-url=https://web.archive.org/web/20181113141726/https://www.estadistica.ad/serveiestudis/web/index.asp |archive-date=13 Nov 2018 |url-status=live}}</ref> || |- | {{hs|end}} – | {{flag|Isle of Man}} (UK) | {{n+p|84530|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 April 2023|abbr=on}} | 2021 census result<ref>{{cite web |url=https://www.gov.im/news/2022/feb/02/first-details-of-2021-census-report-published/ |title=First details of 2021 Census report published |website=www.gov.im |access-date=15 Feb 2022 |archive-date=1 May 2022 |archive-url=https://web.archive.org/web/20220501095106/https://www.gov.im/news/2022/feb/02/first-details-of-2021-census-report-published/ |url-status=live }}</ref> || |- | {{hs|end}} – | {{flag|Cayman Islands}} (UK) | {{n+p|71105|{{worldpop}}|sigfig=1|disp=table}} || {{dts|30 Sep 2020|abbr=on}} | 2021 Census<ref>{{cite web |title=Cayman Islands - 2021 Census results |url=https://www.eso.ky/user_images/census_2021/Section_1.1.pdf |url-status=live |archive-url=https://web.archive.org/web/20190326030427/https://www.eso.ky/ |archive-date=26 Mar 2019 |access-date=23 Mar 2019 |website=www.eso.ky}}</ref>|| |- | {{nts|186}} | {{flag|Dominica}} | {{n+p|67408|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 Dec 2017|abbr=on}} | Official estimate<ref>{{cite web |date=2011 |title=End of year Population by Sex 2005 to 2017 |url=https://stats.gov.dm/subjects/demographic-statistics/end-of-year-population-by-sex-2005-to-2017/ |archive-url=https://web.archive.org/web/20230331200204/https://stats.gov.dm/subjects/demographic-statistics/end-of-year-population-by-sex-2005-to-2017/ |archive-date=31 Mar 2023 |access-date=25 Jun 2023 |website=stats.gov.dm |publisher=Central Statistics Office of Dominica}}</ref>|| |- | {{hs|end}} – | {{flag|Guernsey}} (UK) | {{n+p|64421|{{worldpop}}|sigfig=1|disp=table}} || {{dts|30 Jun 2023|abbr=on}} | National quarterly estimate<ref>{{cite web |url=https://www.gov.gg/population |title=Population, Employment and Earnings |website=www.gov.gg |date=30 Apr 2024 |access-date=3 Jul 2024 |archive-date=8 October 2020 |archive-url=https://web.archive.org/web/20201008212617/https://www.gov.gg/population |url-status=live}}</ref> || |- | {{hs|end}} – | {{flag|Bermuda}} (UK) | {{n+p|64055|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2021|abbr=on}} | National annual projection<ref>{{cite web |url=https://www.gov.bm/sites/default/files/Bermuda-Population-Projections-2016-2026.pdf |title=Bermuda Population Projections 2016-2026 |website=www.gov.bm |access-date=11 Mar 2019 |archive-url=https://web.archive.org/web/20190412033845/https://www.gov.bm/sites/default/files/Bermuda-Population-Projections-2016-2026.pdf |archive-date=12 Apr 2019 |url-status=live}}</ref> || |- | {{hs|end}} – | {{flag|Greenland}} (Denmark) | {{n+p|56865|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 July 2023|abbr=on}} | National quarterly estimate<ref>{{cite web |url=http://bank.stat.gl/pxweb/en/Greenland/Greenland__BE__BE85/BEXSTK3.px/?rxid=BEXSAT101-08-2020%2023:21:09 |title=Quarterly Population |date=1 Aug 2020|publisher=Statistics Greenland |website=bank.stat.gl |access-date=1 Aug 2020 |archive-url=https://web.archive.org/web/20201028062810/https://bank.stat.gl/pxweb/en/Greenland/Greenland__BE__BE85/BEXSTK3.px/?rxid=BEXSAT101-08-2020%2023:21:09 |archive-date=28 Oct 2020 |url-status=live}}</ref> || |- | {{hs|end}} – | {{flag|South Ossetia}} | {{n+p|56520|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 Dec 2021|abbr=on}} | Official estimate<ref>{{cite web |url=https://ugosstat.ru/statisticheskij-sbornik-za-yanvar-dekabr-2021-g/ |title=В Цхинвале прошла пресс-конференция начальника Управления государственной статистики Южной Осетии Инала Тибилова |date=29 Mar 2022 |website=ugosstat.ru |access-date=11 Feb 2023 |archive-date=28 January 2023 |archive-url=https://web.archive.org/web/20230128005918/https://ugosstat.ru/statisticheskij-sbornik-za-yanvar-dekabr-2021-g/ |url-status=live}}</ref> || {{efn|{{South Ossetia-note}}}} |- | {{hs|end}} – | {{flag|Faroe Islands}} (Denmark) | {{n+p|54683|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 May 2024|abbr=on}} | Monthly national estimate<ref>{{cite web |url=https://hagstova.fo/en/population/population/population |title=POPULATION |date=10 January 2024 |website=[[Statistics Faroe Islands]] |access-date=5 February 2024}}</ref> || |- | {{nts|187}} | {{flag|Saint Kitts and Nevis}} | {{n+p|51320|{{worldpop}}|sigfig=1|disp=table}} || {{dts| 2022|abbr=on}} | 2022 census result<ref>{{cite web |url=https://nia.gov.kn/wp-content/uploads/2024/06/Census-Report-2021-2022.docx |title= THE POPULATION AND HOUSING CENSUS SUMMARY |access-date=16 July 2024 }}</ref> || |- | {{hs|end}} – | {{flag|American Samoa}} (US) | {{n+p|49710|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Apr 2020|abbr=on}} | 2020 census result<ref>{{cite web |url=https://www2.census.gov/programs-surveys/decennial/2020/data/island-areas/american-samoa/population-and-housing-unit-counts/american-samoa-phc-table01.pdf |title=Population of American Samoa: 2010 and 2020 |website=www2.census.gov |access-date=29 Jan 2022 |archive-date=6 February 2022 |archive-url=https://web.archive.org/web/20220206171554/https://www2.census.gov/programs-surveys/decennial/2020/data/island-areas/american-samoa/population-and-housing-unit-counts/american-samoa-phc-table01.pdf |url-status=live}}</ref> || |- | {{hs|end}} – | {{flag|Turks and Caicos Islands}} (UK) | {{n+p|49309|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2023|abbr=on}} | Official estimate<ref>{{cite web |url=https://www.gov.tc/stats/ |title=2023 INDICATORS |website=www.gov.tc |access-date=3 Jul 2024 |archive-date=27 February 2019 |archive-url=https://web.archive.org/web/20190227182013/https://www.gov.tc/stats/ |url-status=live}}</ref> || |- | {{hs|end}} – | {{flag|Northern Mariana Islands}} (US) | {{n+p|47329|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Apr 2020|abbr=on}} | 2020 census result<ref>{{cite web |url=https://www2.census.gov/programs-surveys/decennial/2020/data/island-areas/commonwealth-of-the-northern-mariana-islands/population-and-housing-unit-counts/commonwealth-northern-mariana-islands-phc-table01.pdf |title=Population of Commonwealth of the Northern Mariana Islands: 2010 and 2020 |website=www2.census.gov |access-date=29 Jan 2022 |archive-date=25 January 2022 |archive-url=https://web.archive.org/web/20220125125558/https://www2.census.gov/programs-surveys/decennial/2020/data/island-areas/commonwealth-of-the-northern-mariana-islands/population-and-housing-unit-counts/commonwealth-northern-mariana-islands-phc-table01.pdf |url-status=live}}</ref> || |- | {{hs|end}} – | {{flag|Sint Maarten}} (Netherlands) | {{n+p|42938|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2023|abbr=on}} | Official estimate<ref>{{cite web |url=http://stats.sintmaartengov.org/press_release/Population/2023/Press_release_est_Population_2023.pdf |title=Population Estimates and Vital Statistics 2023 |website=stat.gov.sx |access-date=12 May 2023 |archive-date=9 May 2023 |archive-url=https://web.archive.org/web/20230509131223/http://stats.sintmaartengov.org/press_release/Population/2023/Press_release_est_Population_2023.pdf |url-status=live}}</ref> || |- | {{nts|188}} | {{flag|Marshall Islands}} | {{n+p|42418|{{worldpop}}|sigfig=1|disp=table}} || {{dts|30 Sep 2021|abbr=on}} | 2021 Census<ref name=autogenerated1>{{cite web |url=https://spccfpstore1.blob.core.windows.net/digitallibrary-docs/files/60/605c69d76a40195baa447b5a558b0e02.pdf?sv=2015-12-11&sr=b&sig=GbQ42ZYNKrH3g389jZn7DvHUO8ObwYrFpnxIeXiRSU0%3D&se=2024-03-25T18%3A07%3A37Z&sp=r&rscc=public%2C%20max-age%3D864000%2C%20max-stale%3D86400&rsct=application%2Fpdf&rscd=inline%3B%20filename%3D%22Marshall_Islands_2021_Census_Vol1_Table_report.pdf%22 |title=Republic of the Marshall Islands 2021 Census Report, Volume 1: Basic Tables and Administrative Report |date=May 30, 2023 |website=Pacific Community (SPC): Statistics for Development Division |publisher=[[Pacific Community]] |access-date=September 27, 2023 |archive-date=27 September 2023 |archive-url=https://web.archive.org/web/20230927183142/https://spccfpstore1.blob.core.windows.net/digitallibrary-docs/files/60/605c69d76a40195baa447b5a558b0e02.pdf?sv=2015-12-11&sr=b&sig=GbQ42ZYNKrH3g389jZn7DvHUO8ObwYrFpnxIeXiRSU0%3D&se=2024-03-25T18:07:37Z&sp=r&rscc=public,%20max-age%3D864000,%20max-stale%3D86400&rsct=application/pdf&rscd=inline%3B%20filename%3D%22Marshall_Islands_2021_Census_Vol1_Table_report.pdf%22 |url-status=live }}</ref>|| |- | {{nts|189}} | {{flag|Liechtenstein}} | {{n+p|40023 |{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 December 2023|abbr=on}} | National estimate<ref>{{cite web |title=LLV |url=https://www.statistikportal.li/de/themen/bevoelkerung/bevoelkerungsstand#collapse-accordion-6374a91a53010709246463-1 }}</ref>|| |- | {{nts|190}} | {{flag|Monaco}} | {{n+p|38367|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 Dec 2023|abbr=on}} | Census<ref>{{cite web |url=https://www.monacostatistics.mc/Population-and-employment |title=Population and employment / IMSEE - Monaco IMSEE |website=www.monacostatistics.mc |access-date=27 Dec 2018 |archive-url=https://web.archive.org/web/20181227230517/https://www.monacostatistics.mc/Population-and-employment |archive-date=27 Dec 2018 |url-status=live}}</ref> || |- | {{hs|end}} – | {{flag|Gibraltar}} (UK) | {{n+p|34003|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2016|abbr=on}} | Official figure<ref>{{cite web |url = https://www.gibraltar.gov.gi/uploads/statistics/2019/Reports/Abstract%20of%20Statistics%202016%20whole%20report.pdf |title = Abstract of Statistics 2016 |website = gibraltar.gov.gi |access-date = 1 Jul 2023 |archive-date = 2 July 2023 |archive-url = https://web.archive.org/web/20230702070247/https://www.gibraltar.gov.gi/uploads/statistics/2019/Reports/Abstract%20of%20Statistics%202016%20whole%20report.pdf |url-status = live }}</ref> || |- | {{nts|191}} | {{flag|San Marino}} | {{n+p|33950|{{worldpop}}|sigfig=1|disp=table}} || {{dts|May 2024|abbr=on}} | Monthly national estimate<ref>{{cite web |url=https://www.statistica.sm/pub1/StatisticaSM/en/Dati-statistici/Popolazione/Struttura-Demografica.html |title=Demographic structure |access-date=8 July 2024 |archive-date=26 March 2020 |archive-url=https://web.archive.org/web/20200326182005/https://www.statistica.sm/on-line/en/home/statistics/population.html |url-status=live }}</ref> || |- | {{hs|end}} – | {{Flagicon image|Local flag of the Collectivity of Saint Martin.svg}} [[Collectivity of Saint Martin|Saint Martin]] (France) | {{n+p|32358|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2020|abbr=on}} | Official estimate<ref name="auto">{{cite web |url=https://www.insee.fr/fr/statistiques/6683019?sommaire=6683037#consulter |title=Populations légales des collectivités d'outre-mer en 2020 |website=www.insee.fr |access-date=27 Apr 2023 |archive-date=25 April 2023 |archive-url=https://web.archive.org/web/20230425205704/https://www.insee.fr/fr/statistiques/6683019?sommaire=6683037#consulter |url-status=live}}</ref> || |- | {{hs|end}} – | {{flag|British Virgin Islands}} (UK) | {{n+p|31538|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2023|abbr=on}} | UN projection<ref name="unpop" /> || |- | {{hs|end}} – | {{flag|Åland}} (Finland) | {{n+p|30604|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 May 2024|abbr=on}} | Monthly national estimate<ref name="pxnet2.stat.fi" /> || |- | {{nts|192}} | {{flag|Palau}} | {{n+p|16733|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2021|abbr=on}} | 2021 Census<ref>{{cite web |url=https://www.palaugov.pw/executive-branch/ministries/finance/budgetandplanning/health-statistics/ |title=Health Statistics – PalauGov.pw |access-date=16 November 2022 |archive-date=9 June 2023 |archive-url=https://web.archive.org/web/20230609150157/https://www.palaugov.pw/executive-branch/ministries/finance/budgetandplanning/health-statistics/ |url-status=live}}</ref>|| |- | {{hs|end}} – | {{flag|Anguilla}} (UK) | {{n+p|15780|{{worldpop}}|sigfig=1|disp=table}} || {{dts|31 Dec 2022|abbr=on}} | Official estimate<ref>{{cite web |url=http://statistics.gov.ai/ |title=Anguilla Statistics Department &#124; Home |website=statistics.gov.ai |access-date=16 November 2022 |archive-date=16 November 2022 |archive-url=https://web.archive.org/web/20221116101412/http://statistics.gov.ai/ |url-status=live}}</ref>|| |- | {{hs|end}} – | {{flag|Cook Islands}} | {{n+p|15040|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2021|abbr=on}} | 2021 Census<ref>{{cite web |url=https://www.mfem.gov.ck/statistics/census-and-surveys/census/267-census-2021 |title=Census 2021 - Cook Islands - Ministry of Finance and Economic Management |website=www.mfem.gov.ck |access-date=16 November 2022 |archive-date=30 June 2023 |archive-url=https://web.archive.org/web/20230630084401/https://www.mfem.gov.ck/statistics/census-and-surveys/census/267-census-2021 |url-status=dead}}</ref>|| |- | {{nts|193}} | {{flag|Nauru}} | {{n+p|11680|{{worldpop}}|sigfig=1|disp=table}} || {{dts|30 Oct 2021|abbr=on}} | 2021 Census<ref>{{cite web |title=NAURU 2021 POPULATION AND HOUSING CENSUS - ANALYTICAL REPORT |url=https://sdd.spc.int/news/2023/08/23/nauru-2021-population-and-housing-census-analytical-report |access-date=6 December 2023}}</ref> || |- | {{hs|end}} – | {{flag|Wallis and Futuna|variant=local}} (France) | {{n+p|11620|{{worldpop}}|sigfig=1|disp=table}} || {{dts|July 24 2023|abbr=on}} | 2023 census result<ref>{{cite web |title=Résultats du recensement de la population 2023 de Wallis-et-Futuna |url=https://www.wallis-et-futuna.gouv.fr/Actualites/Les-chiffres-INSEE-du-recensement-2023-authentifies-sont-parus |access-date=16 July 2024}}</ref> || |- | {{nts|194}} | {{flag|Tuvalu}} | {{n+p|10679|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2021|abbr=on}} | National annual projection<ref name="pacificdata" /> || |- | {{hs|end}} – | {{flag|Saint Barthélemy|variant=local}} (France) | {{n+p|10585|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2020|abbr=on}} | Official estimate<ref name="auto"/> || |- | {{hs|end}} – | {{flag|Saint Pierre and Miquelon|variant=local}} (France) | {{n+p|6092|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2020|abbr=on}} | Official estimate<ref name="auto"/> || |- | {{hs|end}} – | {{flag|Saint Helena, Ascension and Tristan da Cunha}} (UK) | {{n+p|5651|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2021|abbr=on}} | 2021 Census<ref>{{cite web |title=Saint Helena 2021 Population |url=https://www.tristandc.com/population.php |archive-url=https://web.archive.org/web/20211121070744/https://www.tristandc.com/population.php |archive-date=21 Nov 2021 }}</ref><ref>{{cite web |url=https://www.sainthelena.gov.sh/st-helena/statistics/ |title=Statistics |access-date=16 November 2022 |archive-date=26 July 2022 |archive-url=https://web.archive.org/web/20220726162858/https://www.sainthelena.gov.sh/st-helena/statistics/ |url-status=live}}</ref><ref>{{cite web |title=Ascension 2021 population |url=https://www.sainthelena.gov.sh/statistics |archive-url=https://web.archive.org/web/20191029005019/http://www.sainthelena.gov.sh/statistics/ |access-date=16 November 2022 |archive-date=29 October 2019 |url-status=live }}</ref>|| |- | {{hs|end}} – | {{flag|Montserrat}} (UK) | {{n+p|4386|{{worldpop}}|sigfig=1|disp=table}} || {{dts|23 Sep 2023|abbr=on}} | 2023 census result<ref>{{cite web |url=https://www.gov.ms/2024/04/19/key-findings-of-the-2023-population-and-housing-census/ |title=KEY FINDINGS OF THE 2023 POPULATION AND HOUSING CENSUS |access-date=16 July 2024 }}</ref> || |- | {{hs|end}} – | {{flag|Falkland Islands}} (UK) | {{n+p|3662|{{worldpop}}|sigfig=1|disp=table}} || {{dts|10 Oct 2021|abbr=on}} | 2021 census result<ref>{{cite web |url = https://www.falklands.gov.fk/policy/2021-census/census |title = 2021 Census - Preliminary Data Tables |date = 25 Oct 2022 |website = falklands.gov.fk |access-date = 1 Jul 2023 |archive-date = 11 June 2023 |archive-url = https://web.archive.org/web/20230611234237/https://www.falklands.gov.fk/policy/2021-census/census |url-status = live }}</ref> || |- | {{hs|end}} – | {{flag|Norfolk Island}} (Australia) | {{n+p|2188|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2021|abbr=on}} | 2021 Census<ref>{{cite web |url=https://www.abs.gov.au/census/find-census-data/quickstats/2021/SAL90004 |title=2021 Norfolk Island, Census All persons QuickStats &#124; Australian Bureau of Statistics |website=www.abs.gov.au |access-date=16 November 2022 |archive-date=1 November 2022 |archive-url=https://web.archive.org/web/20221101225958/https://www.abs.gov.au/census/find-census-data/quickstats/2021/SAL90004 |url-status=live}}</ref>|| |- | {{hs|end}} – | {{flag|Christmas Island}} (Australia) | {{n+p|1692|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2021|abbr=on}} | 2021 Census <ref>{{cite web |url=https://www.abs.gov.au/census/find-census-data/quickstats/2021/LGA51710 |title=2021 Christmas Island, Census All persons QuickStats &#124; Australian Bureau of Statistics |website=www.abs.gov.au |access-date=16 November 2022 |archive-date=24 December 2022 |archive-url=https://web.archive.org/web/20221224054405/https://abs.gov.au/census/find-census-data/quickstats/2021/LGA51710 |url-status=live}}</ref>|| |- | {{hs|end}} – | {{flag|Niue}} (New Zealand) | {{n+p|1681|{{worldpop}}|sigfig=1|disp=table}} || {{dts|11 Nov 2022|abbr=on}} | 2022 Census <ref>{{cite web |url=https://niuestatistics.nu/population/niue-census-of-population-and-housing-2022 |title=Niue Census of Population and Housing, 2022 Niue Statistics Office |date=30 August 2023 |website=niuestatistics.nu |access-date=6 December 2023}} </ref>|| |- | {{hs|end}} – | {{flag|Tokelau}} (New Zealand) | {{n+p|1647|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jan 2019|abbr=on}} | 2019 Census <ref>{{cite web |url=https://www.tokelau.org.nz/Stats.html |title=Statistics |website=www.tokelau.org.nz |access-date=16 November 2022 |archive-date=30 August 2019 |archive-url=https://web.archive.org/web/20190830214712/https://www.tokelau.org.nz/Stats.html |url-status=live}}</ref>|| |- | {{nts|195}} | {{flag|Vatican City}} | {{n+p|764|{{worldpop}}|sigfig=1|disp=table}} || {{dts|26 June 2023|abbr=on}} | Official figure<ref name=va>{{cite web |url=https://www.vaticanstate.va/it/stato-governo/note-generali/popolazione.html |title=Population |publisher=Vatican City State |date=26 June 2023 |website=www.vaticanstate.va |access-date=16 August 2023 |language=it |archive-date=17 May 2020 |archive-url=https://web.archive.org/web/20200517063650/https://www.vaticanstate.va/it/stato-governo/note-generali/popolazione.html |url-status=live }}</ref> || {{efn|764 residents regardless of citizenship, 618 citizens regardless of residence, 246 resident citizens.<ref name=va/>}} |- | {{hs|end}} – | {{flag|Cocos (Keeling) Islands}} (Australia) | {{n+p|593|{{worldpop}}|sigfig=1|disp=table}} || {{dts|30 Jun 2020|abbr=on}} | 2021 Census<ref>{{cite web |url=https://www.abs.gov.au/census/find-census-data/quickstats/2021/LGA51860 |title=2021 Cocos Islands, Census All persons QuickStats &#124; Australian Bureau of Statistics |website=www.abs.gov.au |access-date=16 November 2022 |archive-date=16 November 2022 |archive-url=https://web.archive.org/web/20221116101418/https://www.abs.gov.au/census/find-census-data/quickstats/2021/LGA51860 |url-status=live}}</ref>|| |- | {{hs|end}} – | {{flag|Pitcairn Islands}} (UK) | {{n+p|40|{{worldpop}}|sigfig=1|disp=table}} || {{dts|1 Jul 2021|abbr=on}} | Official estimate<ref>{{cite web |title=Pitcairn Islands Tourism |url=http://www.visitpitcairn.pn |url-status=live |archive-url=https://web.archive.org/web/20240510031311/https://visitpitcairn.pn |archive-date=10 May 2024 |access-date=6 July 2024 |website=visitpitcairn.pn}}</ref>|| |} ==Related pages== *[[List of countries by population (United Nations)]] ==References== {{Reflist}}e * CIA, 9 August, 2005. '''Notes''' <references group="lower-alpha"/> {{DEFAULTSORT:countries by population, List of}} [[Category:Lists of countries|*Population]]
This is a list of countries and dependencies by population. It includes sovereign states, inhabited dependent territories and, in some cases, constituent countries of sovereign states, with inclusion within the list being primarily based on the ISO standard ISO 3166-1. For instance, the United Kingdom is considered a single entity, while the constituent countries of the Kingdom of the Netherlands are considered separately. In addition, this list includes certain states with limited recognition not found in ISO 3166-1. Also given in a percentage is each country's population compared with the world population, which the United Nations estimates at 8.08 billion as of 2023. Figures used in this chart are based on the most up-to-date estimates or projections by the national census authority, where available, and are usually rounded off. Where updated national data are not available, figures are based on the estimates or projections for 2022 by the Population Division of the United Nations Department of Economic and Social Affairs. Because the compiled figures are not collected at the same time in every country, or at the same level of accuracy, the resulting numerical comparisons may create misleading conclusions. Furthermore, the addition of figures from all countries may not equal the world total. Areas that form integral parts of sovereign states, such as the countries of the United Kingdom, are counted as part of the sovereign states concerned. Not included are other entities that are not sovereign states, such as the European Union, and independent territories that do not have permanent populations, such as the Chagos Archipelago and various countries' claims to Antarctica. Note: A numbered rank is assigned to the 193 member states of the United Nations, plus the two observer states to the United Nations General Assembly. Dependent territories and constituent countries that are parts of sovereign states are not assigned a numbered rank. In addition, sovereign states with limited recognition are included, but not assigned a number rank.
{{selfref|For Wikipedia's policy on signatures, see [[Wikipedia:Signatures]]}} {{one source|date=April 2024}} [[Image:Autograf, Martin Luther, Nordisk familjebok.png|thumb|Signature of [[Martin Luther]]]] [[Image:Karldergrossesignatur.svg|thumb|Signature of [[Charlemagne]]. It contains the word KAROLVS.]] A '''signature''' is a special way that people write their name to let others know that they understand or approve of something that was written. A signature is often used to sign a [[contract]], a [[cheque]], a [[petition]] or a [[treaty]].<ref>{{Cite web|title=Definition of SIGNATURE|url=https://www.merriam-webster.com/dictionary/signature|access-date=2021-09-11|website=www.merriam-webster.com|language=en}}</ref> When it comes to signature's there are many ways to write a signature, and signatures may look different depending on your location. In the United States, many people have a signature which is made from writing their name in their own handwriting, often in [[cursive]]. Some signatures may be written in a different style than normal writing, which may make it difficult to read. Some people practice [[autograph]]s, or fancy signatures that are hard to copy. Hundreds of years ago some people used sealing wax to make a copy of their [[Seal (device)|signet rings]] next to or instead of a signature. This was called their "seal. This is where the term "seal of approval" comes from. Some formal documents still use an official wax seal, such as a [[coat of arms]]. == Types of signatures == '''Wet signature:-''' A wet signature, ink signature, or wet ink signature is the physical marking a person makes to sign their name on a physical document, usually with a pen. Wet signatures are the most common kind of signatures and have been used for centuries as a unique marking to indicate agreement and prevent fraud. Physical signatures have been used since 3500 B.C. in the form of seals, but it was in the 1600s when signatures written on paper became widely used and legally valid. Although wet signatures can be forged by skilled people, they were the norm until the beginning of the 21st century, when online signatures started to become legally binding. '''Electronic signature or E-signature:-''' Software programmes allow an electronic representation of a signature to be used that's been created by the software. '''Digital signature:-''' A digital signature is a signature created by the authoriser such as copying, pasting and embedding a copy of their signature or a signature created using a stylus pen. == References == <references /> [[Category:Writing]] [[Category:Authentication methods]] {{stub}}
A signature (/ˈsɪɡnətʃər/; from Latin: signare, "to sign") is a handwritten (and often stylized) depiction of someone's name, nickname, or even a simple "X" or other mark that a person writes on documents as a proof of identity and intent. The writer of a signature is a signatory or signer. Similar to a handwritten signature, a signature work describes the work as readily identifying its creator. A signature may be confused with an autograph, which is chiefly an artistic signature. This can lead to confusion when people have both an autograph and signature and as such some people in the public eye keep their signatures private whilst fully publishing their autograph. The traditional function of a signature is to permanently affix to a document a person's uniquely personal, undeniable self-identification as physical evidence of that person's personal witness and certification of the content of all, or a specified part, of the document. For example, the role of a signature in many consumer contracts is not solely to provide evidence of the identity of the contracting party, but also to provide evidence of deliberation and informed consent. In many countries, signatures may be witnessed and recorded in the presence of a notary public to carry additional legal force. In some jurisdictions, an illiterate signatory can make a "mark" (often an "X" but occasionally a personalized symbol) on legal documents, so long as the document is countersigned by a literate witness. In some countries, illiterate people place a thumbprint on legal documents in lieu of a written signature. In the United States, signatures encompass marks and actions of all sorts that are indicative of identity and intent. The legal rule is that unless a statute specifically prescribes a particular method of making a signature it may be made in any number of ways. These include by a mechanical or rubber stamp facsimile. A signature may be made by the purported signatory; alternatively someone else duly authorized by the signatory, acting in the signer's presence and at the signatory's direction, may make the signature. Many individuals have much more fanciful signatures than their normal cursive writing, including elaborate ascenders, descenders and exotic flourishes, much as one would find in calligraphic writing. As an example, the final "k" in John Hancock's famous signature on the US Declaration of Independence loops back to underline his name. This kind of flourish is also known as a paraph, a French term meaning flourish, initial or signature. The paraph is used in graphology analyses. Several cultures whose languages use writing systems other than alphabets do not share the Western notion of signatures per se: the "signing" of one's name results in a written product no different from the result of "writing" one's name in the standard way. For these languages, to write or to sign involves the same written characters. Also see Calligraphy. Special signature machines, called autopens, are capable of automatically reproducing an individual's signature. These are typically used by people required to sign a lot of printed matter, such as celebrities, heads of state or CEOs. More recently, Members of Congress in the United States have begun having their signature made into a TrueType font file. This allows staff members in the Congressman's office to easily reproduce it on correspondence, legislation, and official documents. In the East Asian languages of Chinese, Japanese, and Korean, people traditionally use stamp-like objects known as name-seals with the name carved in tensho script (seal script) in lieu of a handwritten signature. A wet signature is a person's name written in their own hand with ink. Some government agencies require that professional persons or official reviewers sign originals and all copies of originals to authenticate that they personally viewed the content. In the United States this is prevalent with architectural and construction plans. Its intent is to prevent mistakes or fraud but the practice is not known to be effective. Handwriting experts say "it is extremely difficult for anyone to be able to figure out if a signature or other very limited writing sample has been forged." High volume review of signatures, to decide if a signature is true or forged, occurs when election offices decide whether to accept absentee ballots arriving from voters, and possibly when banks decide whether to pay checks. The highest error rates in signature verification are found with lay people, higher than for computers, which in turn make more errors than experts. There have been concerns that signature reviews improperly reject ballots from young and minority voters at higher rates than others, with no or limited ability of voters to appeal the rejection. When errors are made with bank checks, the payer can ask the bank for corrections. In 2018, a fifth of adults in the United Kingdom said they sign so rarely they have no consistent signature, including 21% of people 18-24 and 16% of people over age 55. 55% of UK adults said they rarely sign anything. Researchers have published error rates for computerized signature verification. They compare different systems on a common database of true and false signatures. The best system falsely rejects 10% of true signatures, while it accepts 10% of forgeries. Another system has error rates on both of 14%, and the third-best has error rates of 17%. It is possible to be less stringent and reject fewer true signatures, at the cost of also rejecting fewer forgeries. Computer algorithms: look for a certain number of points of similarity between the compared signatures ... a wide range of algorithms and standards, each particular to that machine's manufacturer, are used to verify signatures. In addition, counties have discretion in managing the settings and implementing manufacturers' guidelines ... there are no statewide standards for automatic signature verification ... most counties do not have a publicly available, written explanation of the signature verification criteria and processes they use. In an experiment, experts rejected 5% of true signatures and 71% of forgeries. They were doubtful about another 57% of true signatures and 27% of forgeries. If computer verification is adjusted to reflect what experts are sure about, it will wrongly reject 5% of true signatures and wrongly accept 29% of forgeries. If computers were adjusted more strictly, rejecting all signatures which experts have doubts about, the computers would set aside 62% of true signatures, and still wrongly accept 2% of forgeries. Lay people made more mistakes and were doubtful less often, though the study does not report whether their mistakes were to accept more forgeries or reject more true signatures. Voters with short names are at a disadvantage, since experts make more mistakes on signatures with fewer "turning points and intersections." Participants in this study had 10 true signatures to compare to, which is more than most postal ballot verifications have. A more recent study for the US Department of Justice confirms the probabilistic nature of signature verification, though it does not provide numbers. In e-mail and newsgroup usage, another type of signature exists which is independent of one's language. Users can set one or more lines of custom text known as a signature block to be automatically appended to their messages. This text usually includes a name, contact information, and sometimes quotations and ASCII art. A shortened form of a signature block, only including one's name, often with some distinguishing prefix, can be used to simply indicate the end of a post or response. Some web sites also allow graphics to be used. Note, however, that this type of signature is not related to electronic signatures or digital signatures, which are more technical in nature and not directly understandable by humans. For guidance applicable in England and Wales on the use of pre-signed signature pages being subsequently attached to documents to effect a "virtual" signing, see Law of Property (Miscellaneous Provisions) Act 1989#Validity of execution under Mercury. The signature on a painting or other work of art has always been an important item in the assessment of art. Fake signatures are sometimes added to enhance the value of a painting, or are added to a fake painting to support its authenticity. A notorious case was the signature of Johannes Vermeer on the fake "Supper at Emmaus" made by the art-forger Han van Meegeren. However, the fact that painters' signatures often vary over time (particularly in the modern and contemporary periods) might complicate the issue. The signatures of some painters take on an artistic form that may be of less value in determining forgeries. If a painting is abstract or ambiguous, the signature can be the only clue to determine which side is the top. Under British law, the appearance of signatures (not the names themselves) may be protected under copyright law. Under United States copyright law, "titles, names [I c...]; mere variations of typographic ornamentation, lettering, or coloring" are not eligible for copyright; however, the appearance of signatures (not the names themselves) may be protected under copyright law. Uniform Commercial Code §1-201(37) of the United States generally defines signed as "using any symbol executed or adopted with present intention to adopt or accept a writing." The Uniform Commercial Code §3-401(b) for negotiable instruments states "A signature may be made (i) manually or by means of a device or machine, and (ii) by the use of any name, including a trade or assumed name, or by a word, mark, or symbol executed or adopted by a person with present intention to authenticate a writing."
{{Infobox country | conventional_long_name = Republic of Côte d'Ivoire | common_name = Côte d'Ivoire | native_name = ''République de Côte d'Ivoire'' {{resize|70%|([[French language|French]])}} | image_flag = Flag of Côte d'Ivoire.svg | image_coat = File:Coat of arms of Ivory Coast (2).svg | national_motto = {{native phrase|fr|"Union – Discipline – Travail"|italics=off}}<br>{{small|"Unity – Discipline – Work"}} | national_anthem = ''[[L'Abidjanaise]]''<br>{{small|''Song of [[Abidjan]]''}}<br> | image_map = Côte d'Ivoire (orthographic projection).svg | map_caption = | image_map2 = Ivory Coast - Location Map (2013) - CIV - UNOCHA.svg | capital = [[Yamoussoukro]] <small>(political)</small><br>[[Abidjan]] <small>(economic) </small> | coordinates = {{Coord|6|51|N|5|18|W|type:city}} | largest_city = [[Abidjan]] | official_languages = French | national_languages = Bété • Dioula • Baoulé • Abron • Agni • Cebaara Senufo • others | ethnic_groups = {{unbulleted list | 41.1% [[Akan people|Akan]] | {{nowrap|17.6% [[Ethnic groups in Ivory Coast|Voltaiques{{\}}Gur]]}} | 27.5% [[Mandé peoples|(Dyula, Maninka)]] | 11.0% Krous | 2.8% others<sup>a</sup> }} | ethnic_groups_year = 2018 | demonym = {{unbulleted list |Ivorian}} | government_type = [[Unitary state|Unitary]] [[Presidential system|presidential]] [[republic]] under a [[parliamentary system]] | leader_title1 = [[List of heads of state of Ivory Coast|President]] | leader_name1 = [[Alassane Ouattara]] | leader_title2 = [[Vice President of Ivory Coast|Vice President]] | leader_name2 = [[Tiémoko Meyliet Koné]] | leader_title3 = [[List of heads of government of Ivory Coast|Prime Minister]] | leader_name3 = [[Robert Beugré Mambé]] | legislature = [[Parliament of Ivory Coast]] | upper_house = [[Senate (Ivory Coast)|Senate]] | lower_house = [[National Assembly (Ivory Coast)|National Assembly]] | sovereignty_type = [[Independence]] | established_event1 = from [[France]] | established_date1 = 7 August 1960 | area_km2 = 322,463 | area_rank = 68th <!-- Should match [[List of countries and dependencies by area]] --> | area_sq_mi = 124,502 <!--Do not remove per [[WP:MOSNUM]]--> | percent_water = 1.4<ref name="CIA">{{cite web |url=https://www.cia.gov/library/publications/the-world-factbook/geos/iv.html |title=Côte d'Ivoire |website=The World Factbook |publisher=[[Central Intelligence Agency|CIA]] Directorate of Intelligence |accessdate=8 August 2008 |date=24 July 2008 |archive-date=31 August 2020 |archive-url=https://web.archive.org/web/20200831020553/https://www.cia.gov/library/publications/the-world-factbook/geos/iv.html |url-status=dead }}</ref> | population_estimate = 23,740,424<ref name="CIA World Factbook">{{cite web |url=https://www.cia.gov/library/publications/the-world-factbook/geos/iv.html |title=Côte d'Ivoire |website=The World Factbook |publisher=[[Central Intelligence Agency|CIA]] Directorate of Intelligence |accessdate=18 February 2017 |archive-date=31 August 2020 |archive-url=https://web.archive.org/web/20200831020553/https://www.cia.gov/library/publications/the-world-factbook/geos/iv.html |url-status=dead }}</ref> | population_census = 24,905,843 | population_estimate_year = 2018 | population_estimate_rank = 54th | population_census_year = 2015 | population_density_km2 = 63.9 | population_density_sq_mi = 191 <!--Do not remove per [[WP:MOSNUM]]--> | population_density_rank = 139th | GDP_PPP = $106.412&nbsp;billion<ref name=imf2>{{cite web |url=https://www.imf.org/external/pubs/ft/weo/2018/01/weodata/weorept.aspx?pr.x=83&pr.y=8&sy=2018&ey=2021&scsm=1&ssd=1&sort=country&ds=.&br=1&c=662&s=NGDPD%2CNGDPDPC%2CPPPGDP%2CPPPPC&grp=0&a=|title=Côte d'Ivoire|publisher=International Monetary Fund }}</ref> | GDP_PPP_year = 2018 | GDP_PPP_rank = | GDP_PPP_per_capita = $4,155<ref name=imf2/> | GDP_PPP_per_capita_rank = | GDP_nominal = $48.142&nbsp;billion<ref name=imf2/> | GDP_nominal_year = 2018 | GDP_nominal_rank = | GDP_nominal_per_capita = $1,879<ref name=imf2/> | GDP_nominal_per_capita_rank = | Gini = 41.5 <!--number only--> | Gini_year = 2008 | Gini_change = <!--increase/decrease/steady--> | Gini_ref = <ref name="wb-gini">{{cite web |url=http://data.worldbank.org/indicator/SI.POV.GINI/ |title=Gini Index |publisher=World Bank |accessdate=2 March 2011}}</ref> | Gini_rank = | HDI = 0.492 <!--number only--> | HDI_year = 2017<!-- Please use the year to which the data refers, not the publication year--> | HDI_change = increase<!--increase/decrease/steady--> | HDI_ref = <ref name="HDI">http://hdr.undp.org/sites/default/files/2018_human_development_statistical_update.pdf</ref> | HDI_rank = 170th | currency = [[West African CFA franc]] | currency_code = XOF | time_zone = [[Greenwich Mean Time|GMT]] | utc_offset = +0 | utc_offset_DST = | time_zone_DST = | drives_on = right | calling_code = [[+225]] | iso3166code = CI | cctld = [[.ci]] | footnote_a = Including approximately 130,000 [[Lebanese people|Lebanese]] and 14,000 [[French people|French]] people. | footnotes = }} '''Ivory Coast''' or '''Côte d'Ivoire''',{{efn|The latter being pronounced {{IPAc-en|ˌ|k|oʊ|t|_|d|iː|ˈ|v|w|ɑːr}} {{respell|KOHT|_|dee|VWAR}} in English and {{IPA-fr|kot divwaʁ||Fr-Côte-d'Ivoire-fr-Paris.ogg}} in French.<ref>{{Cite web |url=http://dictionary.reference.com/browse/cote%20d'ivoire |title=Cote d'Ivoire definition |publisher=Dictionary.com |accessdate=23 May 2014 }}</ref>}} officially the '''Republic of Côte d'Ivoire''', is a country in [[West Africa]]. The capital of Côte d'Ivoire is [[Yamoussoukro]] but its biggest city is [[Abidjan]]. Other cities can be found at [[List of cities in Côte d'Ivoire]]. ==Geography== It borders the [[Gulf of Guinea]] to the south and five other [[Africa]]n nations. [[Liberia]] is to the southwest, [[Guinea]] to the northwest, [[Mali]] to the north-northwest, [[Burkina Faso]] to the north-northeast, and [[Ghana]] to the east. == Districts == [[File:Districts of Côte d'Ivoire (Numbered).png|thumb|Districts of Ivory Coast]] Ivory Coast is divided into 12 districts and 2 district-level cities. The districts were created in 2011. {| class="sortable wikitable plainrowheaders" ! scope="col" width="15" |Map no. ! scope="col" width="140" |District ! scope="col" width="140" |District capital |- |1 |'''[[Abidjan]]''' <br />(''District Autonome d'Abidjan'') |- |2 |'''[[Bas-Sassandra District|Bas-Sassandra]]''' <br />(''District du Bas-Sassandra'') |[[San-Pédro, Ivory Coast|San-Pédro]] |- |3 |'''[[Comoé District|Comoé]]''' <br />(''District du Comoé'') |[[Abengourou]] |- |4 |'''[[Denguélé District|Denguélé]]''' <br />(''District du Denguélé'') |[[Odienné]] |- |5 |'''[[Gôh-Djiboua District|Gôh-Djiboua]]''' <br />(''District du Gôh-Djiboua'') |[[Gagnoa]] |- |6 |'''[[Lacs District|Lacs]]''' <br />(''District des Lacs'') |[[Dimbokro]] |- |7 |'''[[Lagunes District|Lagunes]]''' <br />(''District des Lagunes'') |[[Dabou]] |- |8 |'''[[Montagnes District|Montagnes]]''' <br />(''District des Montagnes'') |[[Man, Ivory Coast|Man]] |- |9 |'''[[Sassandra-Marahoué District|Sassandra-Marahoué]]''' <br />(''District du Sassandra-Marahoué'') |[[Daloa]] |- |10 |'''[[Savanes District|Savanes]]''' <br />(''District des Savanes'') |[[Korhogo]] |- |11 |'''[[Vallée du Bandama District|Vallée du Bandama]]''' <br />(''District de la Vallée du Bandama'') |[[Bouaké]] |- |12 |'''[[Woroba District|Woroba]]''' <br />(''District du Woroba'') |[[Séguéla]] |- |13 |'''[[Yamoussoukro]]''' <br />(''District Autonome du Yamoussoukro'') |- |14 |'''[[Zanzan District|Zanzan]]''' <br />(''District du Zanzan'') |[[Bondoukou]] |} ==Related pages== *[[Côte d'Ivoire at the Olympics]] *[[Côte d'Ivoire national football team]] *[[List of rivers of Côte d'Ivoire]] == References == {{reflist}} '''Notes''' {{Reflist|group=lower-alpha}} {{Africa}} {{DEFAULTSORT:Cote d'Ivoire}} [[Category:Ivory Coast| ]] [[Category:French-speaking countries]] [[Category:Members of the Organisation of Islamic Cooperation]] [[Category:1960 establishments in Africa]]
Ivory Coast, officially the Republic of Côte d'Ivoire, also known as Côte d'Ivoire, is a country on the southern coast of West Africa. Its capital is Yamoussoukro, in the centre of the country, while its largest city and economic centre is the port city of Abidjan. It borders Guinea to the northwest, Liberia to the west, Mali to the northwest, Burkina Faso to the northeast, Ghana to the east, and the Gulf of Guinea (Atlantic Ocean) to the south. Its official language is French, and indigenous languages are also widely used, including Bété, Baoulé, Dioula, Dan, Anyin, and Cebaara Senufo. In total, there are around 78 different languages spoken in Ivory Coast. The country has a religiously diverse population, including numerous followers of Islam, Christianity, and traditional faiths like Animism. Before its colonization, Ivory Coast was home to several states, including Gyaaman, the Kong Empire, and Baoulé. The area became a protectorate of France in 1843 and was consolidated as a French colony in 1893 amid the Scramble for Africa. It achieved independence in 1960, led by Félix Houphouët-Boigny, who ruled the country until 1993. Relatively stable by regional standards, Ivory Coast established close political-economic ties with its West African neighbours while maintaining close relations with the West, especially France. Its stability was diminished by a coup d'état in 1999 and two civil wars—first between 2002 and 2007 and again during 2010–2011. It adopted a new constitution in 2016. Ivory Coast is a republic with strong executive power vested in its president. Through the production of coffee and cocoa, it was an economic powerhouse in West Africa during the 1960s and 1970s, then experienced an economic crisis in the 1980s, contributing to a period of political and social turmoil that extended until 2011. Ivory Coast has experienced again high economic growth since the return of peace and political stability in 2011. From 2012 to 2021, the economy grew by an average of 7.4% per year in real terms, the second-fastest rate of economic growth in Africa and fourth-fastest rate in the world. In 2020, Ivory Coast was the world's largest exporter of cocoa beans and had high levels of income for its region. The economy still relies heavily on agriculture, with smallholder cash-crop production predominating. Originally, Portuguese and French merchant-explorers in the 15th and 16th centuries divided the west coast of Africa, very roughly, into four "coasts" reflecting resources available from each coast. The coast that the French named the Côte d'Ivoire and the Portuguese named the Costa do Marfim—both meaning "Coast of Ivory"—lay between what was known as the Guiné de Cabo Verde, so-called "Upper Guinea" at Cap-Vert, and Lower Guinea. There was also a Pepper Coast, also known as the "Grain Coast" (present-day Liberia), a "Gold Coast" (Ghana), and a "Slave Coast" (Togo, Benin and Nigeria). Like those, the name "Ivory Coast" reflected the major trade that occurred on that particular stretch of the coast: the export of ivory. Other names for the area included the Côte de Dents, literally "Coast of Teeth", again reflecting the ivory trade; the Côte de Quaqua, after the people whom the Dutch named the Quaqua (alternatively Kwa Kwa); the Coast of the Five and Six Stripes, after a type of cotton fabric also traded there; and the Côte du Vent, the Windward Coast, after perennial local off-shore weather conditions. In the 19th century, usage switched to Côte d'Ivoire. The coastline of the modern state is not quite coterminous with what the 15th- and 16th-century merchants knew as the "Teeth" or "Ivory" coast, which was considered to stretch from Cape Palmas to Cape Three Points and which is thus now divided between the modern states of Ghana and Ivory Coast (with a minute portion of Liberia). It retained the name through French rule and independence in 1960. The name had long since been translated literally into other languages, which the post-independence government considered increasingly troublesome whenever its international dealings extended beyond the Francophone sphere. Therefore, in April 1986, the government declared that Côte d'Ivoire (or, more fully, République de Côte d'Ivoire) would be its formal name for the purposes of diplomatic protocol and has since officially refused to recognize any translations from French to other languages in its international dealings. Despite the Ivorian government's request, the English translation "Ivory Coast" (often "the Ivory Coast") is still frequently used in English by various media outlets and publications. The first human presence in Ivory Coast has been difficult to determine because human remains have not been well preserved in the country's humid climate. However, newly found weapon and tool fragments (specifically, polished axes cut through shale and remnants of cooking and fishing) have been interpreted as a possible indication of a large human presence during the Upper Paleolithic period (15,000 to 10,000 BC), or at the minimum, the Neolithic period. The earliest known inhabitants of the Ivory Coast have left traces scattered throughout the territory. Historians believe that they were all either displaced or absorbed by the ancestors of the present indigenous inhabitants, who migrated south into the area before the 16th century. Such groups included the Ehotilé (Aboisso), Kotrowou (Fresco), Zéhiri (Grand-Lahou), Ega and Diès (Divo). The first recorded history appears in the chronicles of North African (Berber) traders, who, from early Roman times, conducted a caravan trade across the Sahara in salt, slaves, gold, and other goods. The southern termini of the trans-Saharan trade routes were located on the edge of the desert, and from there supplemental trade extended as far south as the edge of the rainforest. The most important terminals—Djenné, Gao, and Timbuctu—grew into major commercial centres around which the great Sudanic empires developed. By controlling the trade routes with their powerful military forces, these empires were able to dominate neighbouring states. The Sudanic empires also became centres of Islamic education. Islam had been introduced in the western Sudan by Muslim Berbers; it spread rapidly after the conversion of many important rulers. From the 11th century, by which time the rulers of the Sudanic empires had embraced Islam, it spread south into the northern areas of contemporary Ivory Coast. The Ghana Empire, the earliest of the Sudanic empires, flourished in the region encompassing present-day southeast Mauritania and southern Mali between the 4th and 13th centuries. At the peak of its power in the 11th century, its realms extended from the Atlantic Ocean to Timbuktu. After the decline of Ghana, the Mali Empire grew into a powerful Muslim state, which reached its apogee in the early part of the 14th century. The territory of the Mali Empire in the Ivory Coast was limited to the northwest corner around Odienné. Its slow decline starting at the end of the 14th century followed internal discord and revolts by vassal states, one of which, Songhai, flourished as an empire between the 14th and 16th centuries. Songhai was also weakened by internal discord, which led to factional warfare. This discord spurred most of the migrations southward toward the forest belt. The dense rainforest covering the southern half of the country created barriers to the large-scale political organizations that had arisen in the north. Inhabitants lived in villages or clusters of villages; their contacts with the outside world were filtered through long-distance traders. Villagers subsisted on agriculture and hunting. Five important states flourished in Ivory Coast during the pre-European early modern period. The Muslim Kong Empire was established by the Dyula in the early 18th century in the north-central region inhabited by the Sénoufo, who had fled Islamization under the Mali Empire. Although Kong became a prosperous centre of agriculture, trade, and crafts, ethnic diversity and religious discord gradually weakened the kingdom. In 1895 the city of Kong was sacked and conquered by Samori Ture of the Wassoulou Empire. The Abron kingdom of Gyaaman was established in the 17th century by an Akan group, the Abron, who had fled the developing Ashanti confederation of Asanteman in what is present-day Ghana. From their settlement south of Bondoukou, the Abron gradually extended their hegemony over the Dyula people in Bondoukou, who were recent arrivals from the market city of Begho. Bondoukou developed into a major centre of commerce and Islam. The kingdom's Quranic scholars attracted students from all parts of West Africa. In the mid-17th century in east-central Ivory Coast, other Akan groups fleeing the Asante established a Baoulé kingdom at Sakasso and two Agni kingdoms, Indénié and Sanwi. The Baoulé, like the Ashanti, developed a highly centralised political and administrative structure under three successive rulers. It finally split into smaller chiefdoms. Despite the breakup of their kingdom, the Baoulé strongly resisted French subjugation. The descendants of the rulers of the Agni kingdoms tried to retain their separate identity long after Ivory Coast's independence; as late as 1969, the Sanwi attempted to break away from Ivory Coast and form an independent kingdom. Compared to neighbouring Ghana, Ivory Coast, though practising slavery and slave raiding, suffered little from the slave trade. European slave and merchant ships preferred other areas along the coast. The earliest recorded European voyage to West Africa was made by the Portuguese in 1482. The first West African French settlement, Saint-Louis, was founded in the mid-17th century in Senegal, while at about the same time, the Dutch ceded to the French a settlement at Gorée Island, off Dakar. A French mission was established in 1687 at Assinie near the border with the Gold Coast (now Ghana). The Europeans suppressed the local practice of slavery at this time and forbade the trade to their merchants. Assinie's survival was precarious, however; the French were not firmly established in Ivory Coast until the mid-19th century. In 1843–44, French Admiral Louis Edouard Bouët-Willaumez signed treaties with the kings of the Grand-Bassam and Assinie regions, making their territories a French protectorate. French explorers, missionaries, trading companies, and soldiers gradually extended the area under French control inland from the lagoon region. Pacification was not accomplished until 1915. Activity along the coast stimulated European interest in the interior, especially along the two great rivers, the Senegal and the Niger. Concerted French exploration of West Africa began in the mid-19th century but moved slowly, based more on individual initiative than on government policy. In the 1840s, the French concluded a series of treaties with local West African chiefs that enabled the French to build fortified posts along the Gulf of Guinea to serve as permanent trading centres. The first posts in Ivory Coast included one at Assinie and another at Grand-Bassam, which became the colony's first capital. The treaties provided for French sovereignty within the posts and for trading privileges in exchange for fees or coutumes paid annually to the local chiefs for the use of the land. The arrangement was not entirely satisfactory to the French, because trade was limited and misunderstandings over treaty obligations often arose. Nevertheless, the French government maintained the treaties, hoping to expand trade. France also wanted to maintain a presence in the region to stem the increasing influence of the British along the Gulf of Guinea coast. The defeat of France in the Franco-Prussian War in 1871 and the subsequent annexation by Germany of the French province of Alsace–Lorraine caused the French government to abandon its colonial ambitions and withdraw its military garrisons from its West African trading posts, leaving them in the care of resident merchants. The trading post at Grand-Bassam was left in the care of a shipper from Marseille, Arthur Verdier, who in 1878 was named Resident of the Establishment of Ivory Coast. In 1886, to support its claims of effective occupation, France again assumed direct control of its West African coastal trading posts and embarked on an accelerated program of exploration in the interior. In 1887, Lieutenant Louis-Gustave Binger began a two-year journey that traversed parts of Ivory Coast's interior. By the end of the journey, he had concluded four treaties establishing French protectorates in Ivory Coast. Also in 1887, Verdier's agent, Marcel Treich-Laplène, negotiated five additional agreements that extended French influence from the headwaters of the Niger River Basin through Ivory Coast. By the end of the 1880s, France had established control over the coastal regions, and in 1889 Britain recognized French sovereignty in the area. That same year, France named Treich-Laplène the titular governor of the territory. In 1893, Ivory Coast became a French colony, with its capital in Grand-Bassam, and Captain Binger was appointed governor. Agreements with Liberia in 1892 and with Britain in 1893 determined the eastern and western boundaries of the colony, but the northern boundary was not fixed until 1947 because of efforts by the French government to attach parts of Upper Volta (present-day Burkina Faso) and French Sudan (present-day Mali) to Ivory Coast for economic and administrative reasons. France's main goal was to stimulate the production of exports. Coffee, cocoa, and palm oil crops were soon planted along the coast. Ivory Coast stood out as the only West African country with a sizeable population of European settlers; elsewhere in West and Central Africa, Europeans who emigrated to the colonies were largely bureaucrats. As a result, French citizens owned one-third of the cocoa, coffee, and banana plantations and adopted the local forced-labour system. Throughout the early years of French rule, French military contingents were sent inland to establish new posts. The African population resisted French penetration and settlement, even in areas where treaties of protection had been in force. Among those offering the greatest resistance was Samori Ture, who in the 1880s and 1890s was establishing the Wassoulou Empire, which extended over large parts of present-day Guinea, Mali, Burkina Faso, and Ivory Coast. Ture's large, well-equipped army, which could manufacture and repair its own firearms, attracted strong support throughout the region. The French responded to Ture's expansion and conquest with military pressure. French campaigns against Ture, which were met with fierce resistance, intensified in the mid-1890s until he was captured in 1898 and his empire dissolved. France's imposition of a head tax in 1900 to support the colony's public works program provoked protests. Many Ivorians saw the tax as a violation of the protectorate treaties because they felt that France was demanding the equivalent of a coutume from the local kings, rather than the reverse. Many, especially in the interior, also considered the tax a humiliating symbol of submission. In 1905, the French officially abolished slavery in most of French West Africa. From 1904 to 1958, Ivory Coast was part of the Federation of French West Africa. It was a colony and an overseas territory under the Third Republic. In World War I, France organized regiments from Ivory Coast to fight in France, and colony resources were rationed from 1917 to 1919. Until the period following World War II, governmental affairs in French West Africa were administered from Paris. France's policy in West Africa was reflected mainly in its philosophy of "association", meaning that all Africans in Ivory Coast were officially French "subjects" but without rights to representation in Africa or France. French colonial policy incorporated concepts of assimilation and association. Based on the assumed superiority of French culture, in practice the assimilation policy meant the extension of the French language, institutions, laws, and customs to the colonies. The policy of association also affirmed the superiority of the French in the colonies, but it entailed different institutions and systems of laws for the colonizer and the colonized. Under this policy, the Africans in Ivory Coast were allowed to preserve their own customs insofar as they were compatible with French interests. An indigenous elite trained in French administrative practice formed an intermediary group between French and Africans. After 1930, a small number of Westernized Ivorians were granted the right to apply for French citizenship. Most Ivorians, however, were classified as French subjects and were governed under the principle of association. As subjects of France, natives outside the civilized elite had no political rights. They were drafted for work in mines, on plantations, as porters, and on public projects as part of their tax responsibility. They were expected to serve in the military and were subject to the indigénat, a separate system of law. During World War II, the Vichy regime remained in control until 1943, when members of General Charles de Gaulle's provisional government assumed control of all French West Africa. The Brazzaville Conference of 1944, the first Constituent Assembly of the Fourth Republic in 1946, and France's gratitude for African loyalty during World War II, led to far-reaching governmental reforms in 1946. French citizenship was granted to all African "subjects", the right to organize politically was recognized, and various forms of forced labour were abolished. Between 1944 and 1946, many national conferences and constituent assemblies took place between France's government and provisional governments in Ivory Coast. Governmental reforms were established by late 1946, which granted French citizenship to all African "subjects" under the colonial control of the French. Until 1958, governors appointed in Paris administered the colony of Ivory Coast, using a system of direct, centralized administration that left little room for Ivorian participation in policy-making. The French colonial administration also adopted divide-and-rule policies, applying ideas of assimilation only to the educated elite. The French were also interested in ensuring that the small but influential Ivorian elite was sufficiently satisfied with the status quo to refrain from developing anti-French sentiments and calls for independence. Although strongly opposed to the practices of association, educated Ivorians believed that they would achieve equality in the French colonial system through assimilation rather than through complete independence from France. After the assimilation doctrine was implemented through the postwar reforms, though, Ivorian leaders realized that even assimilation implied the superiority of the French over the Ivorians and that discrimination and inequality would end only with independence. Félix Houphouët-Boigny, the son of a Baoulé chief, became Ivory Coast's father of independence. In 1944, he formed the country's first agricultural trade union for African cocoa farmers like himself. Angered that colonial policy favoured French plantation owners, the union members united to recruit migrant workers for their own farms. Houphouët-Boigny soon rose to prominence and was elected to the French Parliament in Paris within a year. A year later, the French abolished forced labour. Houphouët-Boigny established a strong relationship with the French government, expressing a belief that Ivory Coast would benefit from the relationship, which it did for many years. France appointed him as a minister, the first African to become a minister in a European government. A turning point in relations with France was reached with the 1956 Overseas Reform Act (Loi Cadre), which transferred several powers from Paris to elected territorial governments in French West Africa and also removed the remaining voting inequities. On 4 December 1958, Ivory Coast became an autonomous member of the French Community, which had replaced the French Union. By 1960, the country was easily French West Africa's most prosperous, contributing over 40% of the region's total exports. When Houphouët-Boigny became the first president, his government gave farmers good prices for their products to further stimulate production, which was further boosted by a significant immigration of workers from surrounding countries. Coffee production increased significantly, catapulting Ivory Coast into third place in world output, behind Brazil and Colombia. By 1979, the country was the world's leading producer of cocoa. It also became Africa's leading exporter of pineapples and palm oil. French technicians contributed to the "Ivorian miracle". In other African nations, the people drove out the Europeans following independence, but in Ivory Coast, they poured in. The French community grew from only 30,000 before independence to 60,000 in 1980, most of them teachers, managers, and advisors. For 20 years, the economy maintained an annual growth rate of nearly 10%—the highest of Africa's non-oil-exporting countries. Houphouët-Boigny's one-party rule was not amenable to political competition. Laurent Gbagbo, who would become the president of Ivory Coast in 2000, had to flee the country in the 1980s after he incurred the ire of Houphouët-Boigny by founding the Front Populaire Ivoirien. Houphouët-Boigny banked on his broad appeal to the population, who continued to elect him. He was criticized for his emphasis on developing large-scale projects. Many felt the millions of dollars spent transforming his home village, Yamoussoukro, into the new political capital were wasted; others supported his vision to develop a centre for peace, education, and religion in the heart of the country. In the early 1980s, the world recession and a local drought sent shock waves through the Ivorian economy. The overcutting of timber and collapsing sugar prices caused the country's external debt to increase three-fold. Crime rose dramatically in Abidjan as an influx of villagers exacerbated unemployment caused by the recession. In 1990, hundreds of civil servants went on strike, joined by students protesting institutional corruption. The unrest forced the government to support multi-party democracy. Houphouët-Boigny became increasingly feeble and died in 1993. He favoured Henri Konan Bédié as his successor. In October 1995, Bédié overwhelmingly won re-election against a fragmented and disorganised opposition. He tightened his hold over political life, jailing several hundred opposition supporters. In contrast, the economic outlook improved, at least superficially, with decreasing inflation and an attempt to remove foreign debt. Unlike Houphouët-Boigny, who was very careful to avoid any ethnic conflict and left access to administrative positions open to immigrants from neighbouring countries, Bedié emphasized the concept of Ivoirité to exclude his rival Alassane Ouattara, who had two northern Ivorian parents, from running for the future presidential election. As people originating from foreign countries are a large part of the Ivorian population, this policy excluded many people of Ivorian nationality. The relationship between various ethnic groups became strained, resulting in two civil wars in the following decades. Similarly, Bedié excluded many potential opponents from the army. In late 1999, a group of dissatisfied officers staged a military coup, putting General Robert Guéï in power. Bedié fled into exile in France. The new leadership reduced crime and corruption, and the generals pressed for austerity and campaigned in the streets for a less wasteful society. A presidential election was held in October 2000 in which Laurent Gbagbo vied with Guéï, but it was not peaceful. The lead-up to the election was marked by military and civil unrest. Following a public uprising that resulted in around 180 deaths, Guéï was swiftly replaced by Gbagbo. Ouattara was disqualified by the country's Supreme Court because of his alleged Burkinabé nationality. The constitution did not allow noncitizens to run for the presidency. This sparked violent protests in which his supporters, mainly from the country's north, battled riot police in the capital, Yamoussoukro. In the early hours of 19 September 2002, while the Gbagbo was in Italy, an armed uprising occurred. Troops who were to be demobilised mutinied, launching attacks in several cities. The battle for the main gendarmerie barracks in Abidjan lasted until mid-morning, but by lunchtime, the government forces had secured Abidjan. They had lost control of the north of the country, and rebel forces made their stronghold in the northern city of Bouaké. The rebels threatened to move on to Abidjan again, and France deployed troops from its base in the country to stop their advance. The French said they were protecting their citizens from danger, but their deployment also helped government forces. That the French were helping either side was not established as a fact, but each side accused the French of supporting the opposite side. Whether French actions improved or worsened the situation in the long term is disputed. What exactly happened that night is also disputed. The government claimed that former president Robert Guéï led a coup attempt, and state TV showed pictures of his dead body in the street; counter-claims stated that he and 15 others had been murdered at his home, and his body had been moved to the streets to incriminate him. Ouattara took refuge in the German embassy; his home had been burned down. President Gbagbo cut short his trip to Italy and on his return stated, in a television address, that some of the rebels were hiding in the shanty towns where foreign migrant workers lived. Gendarmes and vigilantes bulldozed and burned homes by the thousands, attacking residents. An early ceasefire with the rebels, which had the backing of much of the northern populace, proved short-lived and fighting over the prime cocoa-growing areas resumed. France sent in troops to maintain the cease-fire boundaries, and militias, including warlords and fighters from Liberia and Sierra Leone, took advantage of the crisis to seize parts of the west. In January 2003, Gbagbo and rebel leaders signed accords creating a "government of national unity". Curfews were lifted, and French troops patrolled the country's western border. The unity government was unstable, and central problems remained with neither side achieving its goals. In March 2004, 120 people were killed at an opposition rally, and subsequent mob violence led to the evacuation of foreign nationals. A report concluded the killings were planned. Though UN peacekeepers were deployed to maintain a "Zone of Confidence", relations between Gbagbo and the opposition continued to deteriorate. Early in November 2004, after the peace agreement had effectively collapsed because the rebels refused to disarm, Gbagbo ordered airstrikes against the rebels. During one of these airstrikes in Bouaké, on 6 November 2004, French soldiers were hit, and nine were killed; the Ivorian government said it was a mistake, but the French claimed it was deliberate. They responded by destroying most Ivorian military aircraft (two Su-25 planes and five helicopters), and violent retaliatory riots against the French broke out in Abidjan. Gbagbo's original term as president expired on 30 October 2005, but a peaceful election was deemed impossible, so his term in office was extended for a maximum of one year, according to a plan worked out by the African Union and endorsed by the United Nations Security Council. With the late-October deadline approaching in 2006, the election was regarded as very unlikely to be held by that point, and the opposition and the rebels rejected the possibility of another term extension for Gbagbo. The UN Security Council endorsed another one-year extension of Gbagbo's term on 1 November 2006; however, the resolution provided for strengthening of Prime Minister Charles Konan Banny's powers. Gbagbo said the next day that elements of the resolution deemed to be constitutional violations would not be applied. A peace accord between the government and the rebels, or New Forces, was signed on 4 March 2007, and subsequently Guillaume Soro, leader of the New Forces, became prime minister. These events were seen by some observers as substantially strengthening Gbagbo's position. According to UNICEF, at the end of the civil war, water and sanitation infrastructure had been greatly damaged. Communities across the country required repairs to their water supply. The presidential elections that should have been organized in 2005 were postponed until November 2010. The preliminary results showed a loss for Gbagbo in favour of former Prime Minister Ouattara. The ruling FPI contested the results before the Constitutional Council, charging massive fraud in the northern departments controlled by the rebels of the New Forces. These charges were contradicted by United Nations observers (unlike African Union observers). The report of the results led to severe tension and violent incidents. The Constitutional Council, which consisted of Gbagbo supporters, declared the results of seven northern departments unlawful and that Gbagbo had won the elections with 51% of the vote – instead of Ouattara winning with 54%, as reported by the Electoral Commission. After the inauguration of Gbagbo, Ouattara—who was recognized as the winner by most countries and the United Nations—organized an alternative inauguration. These events raised fears of a resurgence of the civil war; thousands of refugees fled the country. The African Union sent Thabo Mbeki, former president of South Africa, to mediate the conflict. The United Nations Security Council adopted a resolution recognising Ouattara as the winner of the elections, based on the position of the Economic Community of West African States, which suspended Ivory Coast from all its decision-making bodies while the African Union also suspended the country's membership. In 2010, a colonel of Ivory Coast armed forces, Nguessan Yao, was arrested in New York in a year-long U.S. Immigration and Customs Enforcement operation charged with procuring and illegal export of weapons and munitions: 4,000 handguns, 200,000 rounds of ammunition, and 50,000 tear-gas grenades, in violation of a UN embargo. Several other Ivory Coast officers were released because they had diplomatic passports. His accomplice, Michael Barry Shor, an international trader, was located in Virginia. The 2010 presidential election led to the 2010–2011 Ivorian crisis and the Second Ivorian Civil War. International organizations reported numerous human-rights violations by both sides. In Duékoué, hundreds of people were killed. In nearby Bloléquin, dozens were killed. UN and French forces took military action against Gbagbo. Gbagbo was taken into custody after a raid into his residence on 11 April 2011. The country was severely damaged by the war, and it was observed that Ouattara had inherited a formidable challenge to rebuild the economy and reunite Ivorians. Gbagbo was taken to the International Criminal Court in January 2016. He was declared acquitted by the court but given a conditional release in January 2019. Belgium has been designated as a host country. Ouattara has ruled the country since 2010. President Ouattara was re-elected in the 2015 presidential election. In November 2020, he won third term in office in elections boycotted by the opposition. His opponents argued it was illegal for Ouattara to run for a third term. Ivory Coast's Constitutional Council formally ratified President Ouattara's re-election to a third term in November 2020. In December 2022, Ivory Coast's electric production company, Compagnie ivoirienne d'électricité [fr] launched a commission to establish the country's first solar plant in Boundiali, with an installation of 37.5 million megawatts, backed by a 10-MW system. On 6 October 2023, Ouattara dissolved the government and removed Prime Minister Achi from his position. The government is divided into three branches: executive, legislative, and judicial. The Parliament of Ivory Coast, consists of the indirectly elected Senate and the National Assembly which has 255 members, elected for five-year terms. Since 1983, Ivory Coast's capital has been Yamoussoukro, while Abidjan was the administrative center. Most countries maintain their embassies in Abidjan. Although most of the fighting ended by late 2004, the country remained split in two, with the north controlled by the New Forces. A new presidential election was expected to be held in October 2005, and the rival parties agreed in March 2007 to proceed with this, but it continued to be postponed until November 2010 due to delays in its preparation. Elections were finally held in 2010. The first round of elections was held peacefully and widely hailed as free and fair. Runoffs were held on 28 November 2010, after being delayed one week from the original date of 21 November. Laurent Gbagbo as president ran against former Prime Minister Alassane Ouattara. On 2 December, the Electoral Commission declared that Ouattara had won the election by a margin of 54% to 46%. In response, the Gbagbo-aligned Constitutional Council rejected the declaration, and the government announced that country's borders had been sealed. An Ivorian military spokesman said, "The air, land, and sea border of the country are closed to all movement of people and goods." President Alassane Ouattara has led the country since 2010 and he was re-elected to a third term in November 2020 elections boycotted by two leading opposition figures former President Henri Konan Bedie and ex-Prime Minister Pascal Affi N'Guessan. The Achi II government has ruled the country since April 2022. In Africa, Ivorian diplomacy favors step-by-step economic and political cooperation. In 1959, Ivory Coast formed the Council of the Entente with Dahomey (Benin), Upper Volta (Burkina Faso), Niger, and Togo; in 1965, the African and Malagasy Common Organization (OCAM); in 1972, the Economic Community of West Africa (CEAO). The latter organization changed to the Economic Community of West African States (ECOWAS) in 1975. A founding member of the Organization of African Unity (OAU) in 1963 and then of the African Union in 2000, Ivory Coast defends respect for state sovereignty and peaceful cooperation between African countries. Worldwide, Ivorian diplomacy is committed to fair economic and trade relations, including the fair trade of agricultural products and the promotion of peaceful relations with all countries. Ivory Coast thus maintains diplomatic relations with international organizations and countries all around the world. In particular, it has signed United Nations treaties such as the Convention Relating to the Status of Refugees, the 1967 Protocol, and the 1969 Convention Governing Specific Aspects of Refugee Problems in Africa. Ivory Coast is a member of the Organisation of Islamic Cooperation, African Union, La Francophonie, Latin Union, Economic Community of West African States, and South Atlantic Peace and Cooperation Zone. Ivory Coast has partnered with nations of the Sub-Saharan region to strengthen water and sanitation infrastructure. This has been done mainly with the help of organizations such as UNICEF and corporations like Nestle. In 2015, the United Nations engineered the Sustainable Development Goals (replacing the Millennium Development Goals). They focus on health, education, poverty, hunger, climate change, water sanitation, and hygiene. A major focus was clean water and salinization. Experts working in these fields have designed the WASH concept. WASH focuses on safe drinkable water, hygiene, and proper sanitation. The group has had a major impact on the sub-Saharan region of Africa, particularly the Ivory Coast. By 2030, they plan to have universal and equal access to safe and affordable drinking water. As of 2012, major equipment items reported by the Ivory Coast Army included 10 T-55 tanks (marked as potentially unserviceable), five AMX-13 light tanks, 34 reconnaissance vehicles, 10 BMP-1/2 armoured infantry fighting vehicles, 41 wheeled APCs, and 36+ artillery pieces. In 2012, the Ivory Coast Air Force consisted of one Mil Mi-24 attack helicopter and three SA330L Puma transports (marked as potentially unserviceable). In 2017, Ivory Coast signed the UN treaty on the Prohibition of Nuclear Weapons. Since 2011, Ivory Coast has been administratively organised into 12 districts plus two district-level autonomous cities. The districts are sub-divided into 31 regions; the regions are divided into 108 departments; and the departments are divided into 510 sub-prefectures. In some instances, multiple villages are organised into communes. The autonomous districts are not divided into regions, but they do contain departments, sub-prefectures, and communes. Since 2011, governors for the 12 non-autonomous districts have not been appointed. As a result, these districts have not yet begun to function as governmental entities. The following is the list of districts, district capitals and each district's regions: Ivory Coast is a country in western sub-Saharan Africa. It borders Liberia and Guinea in the west, Mali and Burkina Faso in the north, Ghana in the east, and the Gulf of Guinea (Atlantic Ocean) in the south. The country lies between latitudes 4° and 11°N, and longitudes 2° and 9°W. Around 64.8% of the land is agricultural land; arable land amounted to 9.1%, permanent pasture 41.5%, and permanent crops 14.2%. Water pollution is one of the biggest issues that the country is currently facing. There are over 1,200 animal species including 223 mammals, 702 birds, 161 reptiles, 85 amphibians, and 111 species of fish, alongside 4,700 plant species. It is the most biodiverse country in West Africa, with the majority of its wildlife population living in the nation's rugged interior. The nation has nine national parks, the largest of which is Assgny National Park which occupies an area of around 17,000 hectares or 42,000 acres. The country contains six terrestrial ecoregions: Eastern Guinean forests, Guinean montane forests, Western Guinean lowland forests, Guinean forest–savanna mosaic, West Sudanian savanna, and Guinean mangroves. It had a 2018 Forest Landscape Integrity Index mean score of 3.64/10, ranking it 143rd globally out of 172 countries. Ivory Coast has, for the region, a relatively high income per capita (US$1,662 in 2017) and plays a key role in transit trade for neighbouring landlocked countries. The country is the largest economy in the West African Economic and Monetary Union, constituting 40% of the monetary union's total GDP. Ivory Coast is the fourth-largest exporter of general goods in sub-Saharan Africa (following South Africa, Nigeria, and Angola). The country is the world's largest exporter of cocoa beans. In 2009, cocoa-bean farmers earned $2.53 billion for cocoa exports and were projected to produce 630,000 metric tons in 2013. Ivory Coast also has 100,000 rubber farmers who earned a total of $105 million in 2012. Close ties to France since independence in 1960, diversification of agricultural exports, and encouragement of foreign investment have been factors in economic growth. In recent years, Ivory Coast has been subject to greater competition and falling prices in the global marketplace for its primary crops of coffee and cocoa. That, compounded with high internal corruption, makes life difficult for the grower, those exporting into foreign markets, and the labour force; instances of indentured labour have been reported in the country's cocoa and coffee production in every edition of the U.S. Department of Labor's List of Goods Produced by Child Labor or Forced Labor since 2009. Ivory Coast's economy has grown faster than that of most other African countries since independence. One possible reason for this might be taxes on exported agriculture. Ivory Coast, Nigeria, and Kenya were exceptions as their rulers were themselves large cash-crop producers, and the newly independent countries desisted from imposing penal rates of taxation on exported agriculture. As such, their economies did well. Around 7.5 million people made up the workforce in 2009. The workforce took a hit, especially in the private sector, during the early 2000s with numerous economic crises since 1999. Furthermore, these crises caused companies to close and move locations, especially in the tourism industry, and transit and banking companies. Decreasing job markets posed a huge issue as unemployment rates grew. Unemployment rates rose to 9.4% in 2012. Solutions proposed to decrease unemployment included diversifying jobs in small trade. This division of work encouraged farmers and the agricultural sector. Self-employment policy, established by the Ivorian government, allowed for very strong growth in the field with an increase of 142% in seven years from 1995. According to the December 14, 2021 census, the population was 29,389,150, up from 22,671,331 at the 2014 census. The first national census in 1975 counted 6.7 million inhabitants. According to a Demographic and Health Surveys nationwide survey, the total fertility rate stood at 4.3 children per woman in 2021 (with 3.6 in urban areas and 5.3 in rural areas), down from 5.0 children per woman in 2012. It is estimated that 78 languages are spoken in Ivory Coast. French, the official language, is taught in schools and serves as a lingua franca. A semi-creolized form of French, known as Nouchi, has emerged in Abidjan in recent years and spread among the younger generation. One of the most common indigenous languages is Dyula, which acts as a trade language in much of the country, particularly in the north, and is mutually intelligible with other Manding languages widely spoken in neighboring countries. Macroethnic groupings in the country include Akan (42.1%), Voltaiques or Gur (17.6%), Northern Mandés (16.5%), Kru-speaking peoples (11%), Southern Mandés (10%), and others (2.8%, including 100,000 Lebanese and 45,000 French; 2004). Each of these categories is subdivided into different ethnicities. For example, the Akan grouping includes the Baoulé, the Voltaique category includes the Senufo, the Northern Mande category includes the Dioula and the Maninka, the Kru category includes the Bété and the Kru, and the Southern Mande category includes the Yacouba. About 77% of the population is considered Ivorian. Since Ivory Coast has established itself as one of the most successful West African nations, about 20% of the population (about 3.4 million) consists of workers from neighbouring Liberia, Burkina Faso, and Guinea. About 4% of the population is of non-African ancestry. Many are French, Lebanese, Vietnamese and Spanish citizens, as well as evangelical missionaries from the United States and Canada. In November 2004, around 10,000 French and other foreign nationals evacuated Ivory Coast due to attacks from pro-government youth militias. Aside from French nationals, native-born descendants of French settlers who arrived during the country's colonial period are present. Ivory Coast is religiously diverse. According to the latest 2021 census data, adherents of Islam (mainly Sunni) represented 42.5% of the total population, while followers of Christianity (mainly Catholic and Evangelical) comprised 39.8% of the population. An additional 12.6% of the population identified as Irreligious, while 2.2% reported following Animism (traditional African religions). A 2020 estimate by the Pew Research Center, projected that Christians would represent 44% of the total population, while Muslims would represent 37.2% of the population. In addition, it estimated that 8.1% would be religiously unaffiliated, and 10.5% as followers of traditional African religions (animism). In 2009, according to U.S. Department of State estimates, Christians and Muslims each made up 35% to 40% of the population, while an estimated 25% of the population practised traditional (animist) religions. Yamoussoukro is home to the largest church building in the world, the Basilica of Our Lady of Peace. Life expectancy at birth was 42 for males in 2004; for females it was 47. Infant mortality was 118 of 1000 live births. Twelve physicians are available per 100,000 people. About a quarter of the population lives below the international poverty line of US$1.25 a day. About 36% of women have undergone female genital mutilation. According to 2010 estimates, Ivory Coast has the 27th-highest maternal mortality rate in the world. The HIV/AIDS rate was 19th-highest in the world, estimated in 2012 at 3.20% among adults aged 15–49 years. Among sub-Saharan African countries, Ivory Coast has one of the highest literacy rates. According to The World Factbook as of 2019, 89.9% of the population aged 15 and over can read and write. A large part of the adult population, in particular women, is illiterate. Many children between 6 and 10 years old are not enrolled in school. The majority of students in secondary education are male. At the end of secondary education, students can sit for the baccalauréat examination. Universities include Université Félix Houphouët-Boigny in Abidjan and the Université Alassane Ouattara in Bouaké. According to the Ministry of Higher Education and Scientific Research, Ivory Coast devotes about 0.13% of GDP to GERD. Apart from low investment, other challenges include inadequate scientific equipment, the fragmentation of research organizations and a failure to exploit and protect research results. Ivory Coast was ranked 112rd in the Global Innovation Index in 2023, down from 103rd in 2019. The share of the National Development Plan for 2012–2015 that is devoted to scientific research remains modest. Within the section on greater wealth creation and social equity (63.8% of the total budget for the Plan), just 1.2% is allocated to scientific research. Twenty-four national research programmes group public and private research and training institutions around a common research theme. These programmes correspond to eight priority sectors for 2012–2015, namely: health, raw materials, agriculture, culture, environment, governance, mining and energy; and technology. Each of the ethnic groups in the Ivory Coast has its own music genres, most showing strong vocal polyphony. Talking drums are common, especially among the Appolo, and polyrhythms, another African characteristic, are found throughout Ivory Coast and are especially common in the southwest. Popular music genres from Ivory Coast include zoblazo, zouglou, and Coupé-Décalé. A few Ivorian artists who have known international success are Magic Système, Alpha Blondy, Meiway, Dobet Gnahoré, Tiken Jah Fakoly, DJ Arafat, AfroB, Serge Beynaud and Christina Goh, of Ivorian descent. The most popular sport is association football. The men's national football team has played in the World Cup three times, in Germany 2006, in South Africa 2010, and Brazil in 2014. Côte d'Ivoire has produced many well-known footballers like Didier Drogba and Yaya Touré. The women's football team played in the 2015 Women's World Cup in Canada. The country has been the host of several major African sporting events, with the most recent being the 2013 African Basketball Championship. In the past, the country hosted the 1984 African Cup of Nations, in which the Ivory Coast football team finished fifth, and the 1985 African Basketball Championship, where the national basketball team won the gold medal. 400m metre runner Gabriel Tiacoh won the silver medal in the men's 400 metres at the 1984 Olympics. The country hosted the 8th edition of Jeux de la Francophonie in 2017. In the sport of athletics, well known participants include Marie-Josée Ta Lou and Murielle Ahouré. Rugby union is popular, and the national rugby union team qualified to play at the Rugby World Cup in South Africa in 1995. Ivory Coast has won two Africa Cups: one in 1992 and the other in 2015. Ivory Coast is known for Taekwondo with well-known competitors such as Cheick Cissé, Ruth Gbagbi, and Firmin Zokou. Traditional cuisine is very similar to that of neighbouring countries in West Africa in its reliance on grains and tubers. Cassava and plantains are significant parts of Ivorian cuisine. A type of corn paste called aitiu is used to prepare corn balls, and peanuts are widely used in many dishes. Attiéké is a popular side dish made with grated cassava, a vegetable-based couscous. Common street food is alloco, plantain fried in palm oil, spiced with steamed onions and chili, and eaten along with grilled fish or boiled eggs. Chicken is commonly consumed and has a unique flavor because of its lean, low-fat mass in this region. Seafood includes tuna, sardines, shrimp, and bonito, which is similar to tuna. Mafé is a common dish consisting of meat in peanut sauce. Slow-simmered stews with various ingredients are another common food staple. Kedjenou is a dish consisting of chicken and vegetables slow-cooked in a sealed pot with little or no added liquid, which concentrates the flavors of the chicken and vegetables and tenderizes the chicken. It is usually cooked in a pottery jar called a canary, over a slow fire, or cooked in an oven. Bangui is a local palm wine. Ivorians have a particular kind of small, open-air restaurant called a maquis, which is unique to the region. A maquis normally features braised chicken, and fish covered in onions and tomatoes served with attiéké or kedjenou. 8°N 5°W / 8°N 5°W / 8; -5
[[File:Crosby Garrett Helmet on auction at Christies.jpg|thumb|Auction of a [[helmet]] at [[Christie's]]. ]] [[File:Tea auction australia.jpg|thumb|Melburne [[tea]] auction, [[Printing|print]] from 1885. Also shows tea being weighed and tasted.]][[File:Tsukiji Fresh Tuna Auction.JPG|thumb|An auction of fresh [[tuna]], in [[Japan]]]] [[File:Awaiting Auction - geograph.org.uk - 580878.jpg|thumb|"Grasmere" on Ferry Road awaiting sale by auction. The auctioneers website says: "A detached Victorian property in need of complete renovation set in open countryside with surrounding grass paddock extending to 8.11 [[acre]]s. Guide Price: £150,000 - £200,000."]] An '''auction''' is a method by which goods or services can be [[Trade|sold and bought]]. In an auction, the [[price]] of an item is not fixed in advance. People wanting to buy the item say how much they are willing to pay for it. This is called their ''bid''. In each round, the bids are evaluated. If certain criteria are met, the auction is stopped and one bidder will buy the item at the specified price. Alternatively, the auction can continue for another round. If certain conditions are met, the auction will stop, and the item will not be sold. There are different kinds of auctions, with different rules. Auctions usually happen within a given timeframe. When the time expires, the bid that best matches will win, or the item will not be sold. There are companies that specialize in doing auctions. These companies will charge fees for doing the auction; they may also get a commission that depends on the price the item is sold at. Auctions can be done online, or they can be done offline. Sometimes, bids are placed by telephone, or over the internet. == Most common kinds of auction == The most common kinds of auction are: * ''English auction'': the highest bidder wins. The auction starts at the lowest price the seller is willing to sell at, and the price goes up as long as someone is willing to pay a higher price. * A ''Dutch auction'' goes the other way round: The auction starts at a high price, and goes down until someone is willing to buy. Before the auction, the seller sets the lowest price they'll sell the item for, and if the auction reaches that price without anyone buying, the auction is stopped without selling the item. * ''Sealed first-bid auction'': All the bids are submitted at the same time, without any of the bidders knowing each other's bids. The bidder with the highest bid wins. * ''Vickrey auction'' (or ''sealed-bid second-price auction)'' : This kind of auction works the same as the sealed first-bid auction, but the bidder only pays the second-highest price. This kind of auction is very important in economic theory, but isn't used as much in real life. == The mechanisms why auctions work == Auctions work because of [[information asymmetry]]: The seller wants to sell at the highest possible price, and the buyers want to pay as little as possible.But the seller does not know the prices the buyers are willing to pay: Setting a price that is too hight means that there will be no sale, setting a price that is too low means that there will be less profit. Each buyer knows how much he is willing to pay, but does not know the price the other buyers are willing to pay. So buyers influence each other. The theory that models these phenomena is called [[Action theory|auction theory]]. It is based on [[game theory]] and [[microeconomics]]. == Related pages == *[[Paul Milgrom]] [[Category:Commerce]]
An auction is usually a process of buying and selling goods or services by offering them up for bids, taking bids, and then selling the item to the highest bidder or buying the item from the lowest bidder. Some exceptions to this definition exist and are described in the section about different types. The branch of economic theory dealing with auction types and participants' behavior in auctions is called auction theory. The open ascending price auction is arguably the most common form of auction and has been used throughout history. Participants bid openly against one another, with each subsequent bid being higher than the previous bid. An auctioneer may announce prices, while bidders submit bids vocally or electronically. Auctions are applied for trade in diverse contexts. These contexts include antiques, paintings, rare collectibles, expensive wines, commodities, livestock, radio spectrum, used cars, real estate, online advertising, vacation packages, emission trading, and many more. The word "auction" is derived from the participle of the Latin word augeō, auctus ("I increase"). Auctions have been recorded as early as 500 BC. According to Herodotus, in Babylon, auctions of women for marriage were held annually. The auctions began with the woman the auctioneer considered to be the most beautiful and progressed to the least beautiful. It was considered illegal to allow a daughter to be sold outside of the auction method. Attractive maidens were offered in a forward auction to determine the price to be paid by a swain, while unattractive maidens required a reverse auction to determine the price to be paid to a swain. Auctions took place in Ancient Greece, other Hellenistic societies, and also in Rome. During the Roman Empire, after a military victory, Roman soldiers would often drive a spear into the ground around which the spoils of war were left, to be auctioned off. Slaves, often captured as the "spoils of war", were auctioned in the Forum under the sign of the spear, with the proceeds of sale going toward the war effort. The Romans also used auctions to liquidate the assets of debtors whose property had been confiscated. For example, Marcus Aurelius sold household furniture to pay off debts, the sales lasting for months. One of the most significant historical auctions occurred in 193 AD when the entire Roman Empire was put on the auction block by the Praetorian Guard. On 28 March 193, the Praetorian Guard first killed emperor Pertinax, then offered the empire to the highest bidder. Didius Julianus won the auction for the price of 6,250 drachmas per guard, an act that initiated a brief civil war. Didius was then beheaded two months later when Septimius Severus conquered Rome. From the end of the Roman Empire to the 18th century, auctions lost favor in Europe, while they had never been widespread in Asia. In China, the personal belongings of deceased Buddhist monks were sold at auction as early as seventh century AD. The first mention of "auction", according to the Oxford English Dictionary, appeared in 1595. In some parts of England during the 17th and 18th centuries, auctions by candle began to be used for the sale of goods and leaseholds. In a candle auction, the end of the auction was signaled by the expiration of a candle flame, which was intended to ensure that no one could know exactly when the auction would end and make a last-second bid. Sometimes, other unpredictable events, such as a footrace, were used instead of the expiration of a candle. This type of auction was first mentioned in 1641 in the records of the House of Lords. The practice rapidly became popular, and in 1660 Samuel Pepys' diary recorded two occasions when the Admiralty sold surplus ships "by an inch of candle". Pepys also relates a hint from a highly successful bidder who had observed that, just before expiring, a candle-wick always flares up slightly: on seeing this, he would shout his final – and winning – bid. The London Gazette began reporting on the auctioning of artwork in the coffeehouses and taverns of London in the late 17th century. The first known auction house in the world was the Stockholm Auction House, Sweden (Stockholms Auktionsverk), founded by Baron Claes Rålamb in 1674. Sotheby's, currently the world's second-largest auction house, was founded in London on 11 March 1744, when Samuel Baker presided over the disposal of "several hundred scarce and valuable" books from the library of an acquaintance. Christie's, now the world's largest auction house, was founded by James Christie in 1766 in London and published its first auction catalog that year, although newspaper advertisements of Christie's sales dating from 1759 have been found. Other early auction houses that are still in operation include Göteborgs Auktionsverk (1681), Dorotheum (1707), Uppsala auktionskammare (1731), Mallams (1788), Bonhams (1793), Phillips de Pury & Company (1796), Freeman's (1805) and Lyon & Turnbull (1826). By the end of the 18th century, auctions of art works were commonly held in taverns and coffeehouses. These auctions were held daily, and auction catalogs were printed to announce available items. In some cases, these catalogs were elaborate works of art themselves, containing considerable detail about the items being auctioned. At the time, Christie's established a reputation as a leading auction house, taking advantage of London's status as the major centre of the international art trade after the French Revolution. The Great Slave Auction took place in 1859 and is recorded as the largest single sale of enslaved people in U.S. history — with 436 men, women and children being sold. During the American Civil War, goods seized by armies were sold at auction by the Colonel of the division. Thus, some of today's auctioneers in the U.S. carry the unofficial title of "colonel". Tobacco auctioneers in the southern United States in the late 19th century had a style that mixed traditions of 17th century England with chants of slaves from Africa. The development of the internet has led to a significant rise in the use of auctions, as auctioneers can solicit bids via the internet from a wide range of buyers in a much larger variety of commodities than was previously practical. In the 1990s, the multi-attribute auction was invented to negotiate extensive conditions of construction and electricity contracts via auction. Also during this time, OnSale.com developed the Yankee auction as its trademark. In the early 2000s, the Brazilian auction was invented as a new type of auction to trade gas through electronic auctions for Linde plc in Brazil. With the emergence of the internet, online auctions have developed, with eBay being the most typical example. For example, if someone owns a rare item, they can display the item through an online auction platform. Interested parties may place bids, with the highest bidder winning the opportunity to purchase the item. Online auctions allow more people to participate and also make traditional auction theory more complex. By increasing visibility of an item and therefore demand, auctions can make an extremely rare item more likely to sell for a higher price. In 2008, the US National Auctioneers Association reported that the gross revenue of the auction industry for that year was approximately $268.4 billion, with the fastest growing sectors being agricultural, machinery, equipment, and residential real estate auctions. The auctions with the largest revenue for the government are often spectrum auctions (typical revenue is estimated in billions of euros) and quota auctions. In 2019, Russia's crab quota was auctioned for €2 billion. Between 1999 and 2002, the British government auctioned off their gold reserves, raising approximately $3.5 billion. The most expensive item to ever be sold in an auction is Leonardo da Vinci's Salvator Mundi in 2017 ($450.3 million). In 2018, the yearly revenues of the two biggest auction houses were $5 billion (Christie's) and $4 billion (Sotheby's). Auctions come in a variety of types and categories, which are sometimes not mutually exclusive. Typification of auctions is considered to be a part of Auction theory. The economists Paul Milgrom and Robert B. Wilson were awarded the 2020 Nobel Prize for the introduction of new auction types (or formats). Auction types share features, which can be summarized into the following list. Auctions can differ in the number and type of participants. There are two types of participants: a buyer and a seller. A buyer pays to acquire a certain good or service, while a seller offers goods or services for money or barter exchange. There can be single or multiple buyers and single or multiple sellers in an auction. If just one seller and one buyer are participating, the process is not considered to be an auction. The forward auction is the most common type of auction — a seller offers item(s) for sale and expects the highest price. A reverse auction is a type of auction in which the roles of the buyer and the seller are reversed, with the primary objective to drive purchase prices downward. While ordinary auctions provide suppliers the opportunity to find the best price among interested buyers, reverse auctions and buyer-determined auctions give buyers a chance to find the lowest-price supplier. During a reverse auction, suppliers may submit multiple offers, usually as a response to competing suppliers' offers, bidding down the price of a good or service to the lowest price they are willing to receive. A reverse price auction is not necessarily 'descending-price' — the reverse Dutch auction is an ascending-price auction because forward Dutch auctions are descending. By revealing the competing bids in real-time to every participating supplier, reverse auctions promote "information transparency". This, coupled with the dynamic bidding process, improves the chances of reaching the fair market value of the item. A double auction is a combination of both forward and reverse auctions. A Walrasian auction or Walrasian tâtonnement is a double auction in which the auctioneer takes bids from both buyers and sellers in a market of multiple goods. The auctioneer progressively either raises or drops the current proposed price depending on the bids of both buyers and sellers, the auction concluding when supply and demand exactly balance. As a high price tends to dampen demand while a low price tends to increase demand, in theory there is a particular price somewhere in the middle where supply and demand will match. A Barter double auction is an auction where every participant has a demand and an offer consisting of multiple attributes and no money is involved. For the mathematical modelling of satisfaction level, Euclidean distance is used, where the offer and demand are treated as vectors. Auctions can be categorized into three types of procedures for auctions depending on the occurrence of a price development during an auction run and its causes. Multiunit auctions sell more than one identical item at a time, rather than having separate auctions for each. This type can be further classified as either a uniform price auction or a discriminatory price auction. An example for them is spectrum auctions. A combinatorial auction is any auction for the simultaneous sale of more than one item where bidders can place bids on an "all-or-nothing" basis on "packages" rather than just individual items. That is, a bidder can specify that they will pay for items A and B, but only if they get both. In combinatorial auctions, determining the winning bidder(s) can be a complex process where even the bidder with the highest individual bid is not guaranteed to win. For example, in an auction with four items (W, X, Y and Z), if Bidder A offers $50 for items W & Y, Bidder B offers $30 for items W & X, Bidder C offers $5 for items X & Z and Bidder D offers $30 for items Y & Z, the winners will be Bidders B & D while Bidder A misses out because the combined bids of Bidders B & D is higher ($60) than for Bidders A and C ($55). Deferred-acceptance auction is a special case of a combinatorial auction. Another special case of a combinatorial auction is the combinatorial clock auction (CCA), which combines a clock auction, during which bidders may provide their confirmations in response to the rising prices, with a subsequantial sealed bid auction, in which bidders submit sealed package bids. The auctioneer uses the final bids to compute the best value allocation and the Vickrey payments. Generalized first-price auctions and Generalized second-price auctions offer slots for multiple bidders instead of making a single deal. The bidders get the slots according to the ranking of their bids. The second-price ruling is derived from the Vickrey auction and means the final deal sealing for the number one bidder is based on the second bidder's price. A No-reserve auction (NR), also known as an absolute auction, is an auction in which the item for sale will be sold regardless of price. From the seller's perspective, advertising an auction as having no reserve price can be desirable because it potentially attracts a greater number of bidders due to the possibility of a bargain. If more bidders attend the auction, a higher price might ultimately be achieved because of heightened competition from bidders. This contrasts with a reserve auction, where the item for sale may not be sold if the final bid is not high enough to satisfy the seller. In practice, an auction advertised as "absolute" or "no-reserve" may nonetheless still not sell to the highest bidder on the day, for example, if the seller withdraws the item from the auction or extends the auction period indefinitely, although these practices may be restricted by law in some jurisdictions or under the terms of sale available from the auctioneer. A reserve auction is an auction where the item for sale may not be sold if the final bid is not high enough to satisfy the seller; that is, the seller reserves the right to accept or reject the highest bid. In these cases, a set 'reserve' price known to the auctioneer, but not necessarily to the bidders, may have been set, below which the item may not be sold. If the seller announces to the bidders the reserve price, it is a public reserve price auction. In contrast, if the seller does not announce the reserve price before the sale, it is a secret reserve price auction. However, potential bidders may be able to deduce an approximate reserve price, if one exists at all, from any estimate given in advance by the auction house. The reserve price may be fixed or discretionary. In the latter case, the decision to accept a bid is deferred to the auctioneer, who may accept a bid that is marginally below it. A reserve auction is safer for the seller than a no-reserve auction as they are not required to accept a low bid, but this could result in a lower final price if less interest is generated in the sale. An all-pay auction is an auction in which all bidders must pay their bids regardless of whether they win. The highest bidder wins the item. All-pay auctions are primarily of academic interest, and may be used to model lobbying or bribery (bids are political contributions) or competitions such as a running race. Bidding fee auction, a variation of all-pay auction, also known as a penny auction, often requires that each participant must pay a fixed price to place each bid, typically one penny (hence the name) higher than the current bid. When an auction's time period expires, the highest bidder wins the item and must pay a final bid price. Unlike in a conventional auction, the final price is typically much lower than the value of the item, but all bidders (not just the winner) will have paid for each bid placed; the winner will buy the item at a very low price (plus price of rights-to-bid used), all the losers will have paid, and the seller will typically receive significantly more than the value of the item. A senior auction is a variation on the all-pay auction, and has a defined loser in addition to the winner. The top two bidders must pay their full final bid amounts, and only the highest wins the auction. The intent is to make the high bidders bid above their upper limits. In the final rounds of bidding, when the current losing party has hit their maximum bid, they are encouraged to bid over their maximum (seen as a small loss) to avoid losing their maximum bid with no return (a very large loss). Another variation of all-pay auction, the top-up auction is primarily used for charity events. Losing bidders must pay the difference between their bid and the next lowest bid. The winning bidder pays the amount bid for the item, without top-up. In a Chinese auction, bidders make sealed bids in advance and their probability of winning grows with the relative size of their bids. In usual auctions like the English one, bids are prices. In Dutch and Japanese auctions, the bids are confirmations. In a version of the Brazilian auction, bids are numbers of units being traded. Structure elements of a bid are called attributes. If a bid is one number like price, it is a single-attribute auction. If bids consists of multiple-attributes, it is a multi-attribute auction. A Yankee auction is a single-attribute multiunit auction running like a Dutch auction, where the bids are the portions of a total amount of identical units. The amount of auctioned items is firm in a Yankee auction unlike a Brazilian auction. The portions of the total amount, bidders can bid, are limited to lower numbers than the total amount. Therefore, only a portion of the total amount will be traded for the best price and the rest to the suboptimal prices. In an English auction, all current bids are visible to all bidders and in a sealed-bid auction, bidders only get to know if their bid was the best. Best/not best auctions are sealed-bid auctions with multiple bids, where the bidders submit their prices like in English auction and get responses about the leadership of their bid. Rank auction is an extension of best/not best auction, where the bidders also see the rank of their bids. Traffic-light auction shows traffic lights to bidders as a response to their bids. These traffic lights depend on the position of the last bid in the distribution of all bids. A buyout auction is an auction with an additional set price (the 'buyout' price) that any bidder can accept at any time during the auction, thereby immediately ending the auction and winning the item. This means that if an item offers its buyout price at the beginning, one participant can stop all other potential participants from bidding at all, or stop the bidding process before the bid price has reached the buyout price. If no bidder chooses to utilize the buyout option before the end of bidding, the highest bidder wins and pays their bid. Buyout options can be either temporary or permanent. In a temporary-buyout auction the option to buy out the auction is not available after the first bid is placed. In a permanent-buyout auction the buyout option remains available throughout the entire auction until the close of bidding. The buyout price can either remain the same throughout the entire auction, or vary throughout according to rules or simply as decided by the seller. The winner selection in most auctions selects the best bid. Unique bid auctions offer a special winner selection. The winner is the bidder with the lowest unique bid. The Chinese auction selects a winner partially based on random. The final price for the selected winner is not always conducted according to their final bid. In the case of the second-price ruling as in a Vickrey auction, the final price for the winner is based on the second bidder's price. A Proxy bid is a special case of second-price ruling used by eBay, where a predefined increment is added to the second highest bid in response to a yet higher bid. Auctions with more than one winner are called multi-winner auctions. Multiunit auction, Combinatorial auction, Generalized first-price auction and Generalized second-price auction are multi-winner auctions. Auctions can be cascaded, one after the other. For instance, an Amsterdam auction is a type of premium auction which begins as an English auction. Once only two bidders remain, each submits a sealed bid. The higher bidder wins, paying either the first or second price. Both finalists receive a premium: a proportion of the excess of the second price over the third price (at which English auction ended). An Anglo-Dutch auction starts as an English or Japanese auction and then continues as a Dutch auction with a reduced number of bidders. A French auction is a preliminary sealed-bid auction before the actual auction, whose reserve price it determines. A sequential auction is an auction where the bidders can participate in a sequence of auctions. A Calcutta auction is a subtype of sequential auction, where the ordering in the sequence is determined by random. A simultaneous ascending auction is an opposite of a sequential auction, where the auctions are run in parallel. The Silent auction is a variant of the English auction in which bids are written on a sheet of paper. At the predetermined end of the auction, the highest listed bidder wins the item. This auction is often used in charity events, with many items auctioned simultaneously and "closed" at a common finish time. The auction is "silent" in that there is no auctioneer selling individual items, the bidders writing their bids on a bidding sheet often left on a table near the item. At charity auctions, bid sheets usually have a fixed starting amount, predetermined bid increments, and a "guaranteed bid" amount which works the same as a "buy now" amount. Other variations of this type of auction may include sealed bids. The highest bidder pays the price they submitted. In private value auctions, every bidder has their own valuation of the auctioned good. A common value auction is opposite, where the valuation of the auctioned good is identical among the bidders. The range of auctions' contexts is extremely wide and one can buy almost anything, from a house to an endowment policy and everything in between. Some of the recent developments have been the use of the Internet both as a means of disseminating information about various auctions and as a vehicle for hosting auctions themselves. As already mentioned in the history section, auctions have been used to trade commodified people from the very first. Auctions have been used in slave markets throughout history until modern times in the post-Gaddafi era Libya. The word for slave auction in the Atlantic slave trade was scramble. A child auction is a Swedish and Finnish historical practice of selling children into slavery-like conditions by authorities using a descending English auction. Fattigauktion is a similar Swedish practice involving poor people being auctioned to church organizations. Trade of wives by auctions was also a common practice throughout history. For instance, in the old English custom of wife selling, a wife was divorced by selling her in a public auction for the highest bid. ISIS conducted slave auctions to sell up to 7,000 Yazidi women as reported in 2020. A virginity auction is the voluntary practice of individuals seeking to sell their own virginity to the highest bid. Cricket players are routinely put up for auction, whereby cricket teams can bid for their services. Indian Premier League (IPL) started annual public auctioning of cricket players in 2008 as an entertainment for mass consumption. Also, Bangladesh Premier League conducts cricket player auctions, starting in 2012. In some countries, such as Australia, auctioning is a common method for the sale of real estate. Auctions were traditionally used as an alternative to the private sale/treaty method to sell property that, due to their unique characteristics, were difficult to determine a price for. The law does not require a vendor to disclose their reserve price prior to the auction. During the 1990s and 2000s, auctions became the primary method for the sale of real estate in the two largest cities, Melbourne and Sydney. This was largely due to the fact that in a private sale the vendor has disclosed the price that they want, and potential purchasers would attempt to low-ball the price, whereas in an auction purchasers do not know what the vendor wants, and thus need to keep lifting the price until the reserve price is reached. The method has been the subject of increased controversy during the twenty-first century as house prices sky-rocketed. The rapidly rising housing market saw many homes, especially in Victoria and New South Wales, selling for significantly more than both the vendors' reserve price and the advertised price range. Subsequently, the auction systems' lack of transparency about the value of the property was brought into question, with estate agents and their vendor clients being accused of "under-quoting". Significant attention was given to the matter by the Australian media, with the government in Victoria eventually bowing to pressure and implementing changes to legislation in an effort to increase transparency. In the UK, historically, auction houses were perceived to sell properties which may have been repossessed - where a home owner fails to make regular mortgage payments - or were probate sales i.e. a family home being sold by the heirs. However, more recently, selling at auction has become an alternative to a normal property sale, due to the speedy nature of the entire process. In China, land auctions are under the sole control of local government officials. Because some developers may use bribes to please government officials to obtain the right to purchase the land, the central government requires that future land auctions be conducted using a spectrum auction in order to prevent the spread of corruption. Although this method cannot completely solve the problem of corruption, it is still a significant contribution to the auction. A government auction is simply an auction held on behalf of a government body generally at a general sale. Items for sale are often surplus needed to be liquidated. Auctions ordered by estate executors enter the assets of individuals who have perhaps died intestate (those who have died without leaving a will), or in debt. In legal contexts where forced auctions occur, as when one's farm or house is sold at auction on the courthouse steps. Property seized for non-payment of property taxes, or under foreclosure, is sold in this manner. Police auctions are generally held at general auctions, although some forces use online sites including eBay, to dispose of lost and found and seized goods. Debt auctions, in which governments issue and sell debt obligations, such as bonds, to investors. The auction is usually sealed and the uniform price paid by the investors is typically the best non-winning bid. In most cases, investors can also place so-called non-competitive bids, which indicates interest to purchase the debt obligation at the resulting price, whatever it may be. Some states use courts to run such auctions. In spectrum auctions conducted by the government, companies purchase licenses to use portions of the electromagnetic spectrum for communications (e.g., mobile phone networks). In certain jurisdictions, if a storage facility's tenant fails to pay rent, the contents of their locker(s) may be sold at a public auction. Several television shows focus on such auctions, including Storage Wars and Auction Hunters. Auctions are used to trade commodities; for example, fish wholesale auctions. In wool auctions, wool is traded in the international market. The wine auction business offers serious collectors an opportunity to gain access to rare bottles and mature vintages, which are not typically available through retail channels. In livestock auctions, sheep, cattle, pigs and other livestock are sold. Sometimes very large numbers of stock are auctioned, such as the regular sales of 50,000 or more sheep during a day in New South Wales. In timber auctions, companies purchase licenses to log on government land. In timber allocation auctions, companies purchase timber directly from the government. In electricity auctions, large-scale generators and distributors of electricity bid on generating contracts. Produce auctions link growers to localized wholesale buyers (buyers who are interested in acquiring large quantities of locally grown produce). Online auctions are a form of E-commerce that relies on the advantages of a digital platform's ability to overcome geographical constraints, provide real-time information and reduce transaction costs, bringing greater convenience to people and allowing more people to participate as bidders, as well as being able to view a greater selection of auctions. Websites like eBay provide a potential market of millions of bidders to sellers. Established auction houses, as well as specialist internet auctions, sell many things online, from antiques and collectibles to holidays, air travel, brand new computers, and household equipment. Private electronic markets use combinatorial auction techniques to continuously sell commodities (coal, iron ore, grain, water, etc.) online to a pre-qualified group of buyers (based on price and non-price factors). Furthermore, online auctions facilitate the process for prospective bidders to discover and evaluate items by enabling searches across numerous auctions and employing filters to refine their selections. On the other hand, an alternative perspective suggests that the format of online auctions could also give rise to collusive conduct and other types of market manipulation, potentially skewing the market and diminishing its efficiency. Firstly, online auctions might enable bidders to obscure their identities, such as utilizing pseudonyms or multiple accounts to maintain anonymity. This concealment could simplify collusion without detection. Secondly, online auctions might ease the implementation of collusive arrangements among bidders. The accessibility of bidding data in online auctions, for instance, allows colluding bidders to monitor each other's bids, guarantee adherence to their agreements, and penalize non-compliance. This enhanced oversight capacity strengthens the stability of collusive agreements. Katehakis and Puranam provided the first model for the problem of optimal bidding for a firm that in each period procures items to meet a random demand by participating in a finite sequence of auctions. In this model an item valuation derives from the sale of the acquired items via their demand distribution, sale price, acquisition cost, salvage value and lost sales. They established monotonicity properties for the value function and the optimal dynamic bid policy. They also provided a model for the case in which the buyer must acquire a fixed number of items either at a fixed buy-it-now price in the open market or by participating in a sequence of auctions. The objective of the buyer is to minimize their expected total cost for acquiring the fixed number of items. During an auction, the seller might possess more comprehensive knowledge regarding the item on offer compared to the buyer, creating an information asymmetry. This lack of information could lead the bidder to overvalue the item and consequently pay a higher price, resulting in the winner's curse. Nevertheless, bidders may also choose to employ bid shading as a strategy to circumvent this predicament. Bid shading is placing a bid which is below the bidder's actual value for the item. Such a strategy risks losing the auction but has the possibility of winning at a low price. Bid shading can also be a strategy to avoid the winner's curse. In either case, the allocation of resources may be inefficient, as the product will not ultimately be acquired by the individual who values it the most. Instead, it will go to the person who either overvalues it the most or effectively employs bid shading. Auction cancellation hunters bid minimal amounts on multiple auctions and expect them to be cancelled. If an auction is cancelled by the seller, they will claim for damages in the amount of the difference between the maximum bid at the time of the auction cancellation and the price of a replacement purchase of the offered item in the auction, when the market is in equilibrium, even if the seller has not sold any of the items, the shadow of bidding still exists. This is the self-protection instinct of the auction market. In order to make this transaction fairer. Auction sniping is the practice of placing a bid at the last moment of the auction. According to the analysis of auction data from eBay, in general, experienced bidders are more likely to snipe in auctions, and those who snipe in auctions are more likely to win. Jump bidding is an aggressive tactic of increasing every bid by high amounts. Calor licitantis is also known as "auction fever" and describes the irrational behavior of bidders at auctions. Suicide bidding is practice in reverse auctions, whereby a bidder submits a bid, which ends up in a loss for this bidder. Whenever bidders at an auction are aware of the identity of the other bidders there is a risk that they will form a "ring" or "pool" and thus manipulate the auction result, a practice known as collusion or more specially bid-rigging. By agreeing to bid only against outsiders, never against members of the "ring", competition becomes weaker, which may dramatically affect the final price level. After the end of the official auction, an unofficial auction may take place among the "ring" members. The difference in price between the two auctions could then be split among the members. This form of a ring was used as a central plot device in the opening episode of the 1979 British television series The House of Caradus, 'For Love or Money', uncovered by Helena Caradus on her return from Paris. In the UK, this auction practice is illegal. It jeopardizes competition on the auction and can demotivate other bidders from participating. It robs the seller of the true value of their good and reduces the auctioneer's commission. Beyond explicit collusion, a tacit coordination of bidders to keep bids low is at least theoretically possible. In case of spectrum auctions, the Federal Communications Commission (FCC) experimented with precautions restricting visibility of bids, limiting the number of bids, click-box bidding, and anonymous bidding in order to prevent bidders from signalling bidding information by embedding it into digits of the bids. Actions within the auction mechanism serve as a communication channel for collusive behavior, once no other channels are legal. This is the practice, especially by high-end art auctioneers, of raising false bids at crucial times in the bidding in order to create the appearance of greater demand or to extend bidding momentum for a work on offer. To call out these nonexistent bids auctioneers might fix their gaze at a point in the auction room that is difficult for the audience to pin down. The practice is frowned upon in the industry. In the United States, chandelier bidding is not illegal. In fact, an auctioneer may bid up the price of an item to the reserve price, which is a threshold below which the consignor may refuse to sell the item. However, the auction house is required to disclose this information. In the United Kingdom this practice is legal on property auctions up to but not including the reserve price, and is also known as off-the-wall bidding. A ring can also be used to increase the price of an auction lot, in which the owner of the object being auctioned may increase competition by taking part in the bidding themself, but drop out of the bidding just before the final bid. This form of a ring was used as a central plot device in an episode of the British television series Lovejoy (series 4, episode 3), in which the price of a watercolour by the (fictional) Jessie Webb is inflated so that others by the same artist could be sold for more than their purchase price. In an English auction, a dummy bid is a bid made by a dummy bidder acting in collusion with the auctioneer or vendor, designed to deceive genuine bidders into paying more. In a first-price auction, a dummy bid is an unfavourable bid designed so as not to become the winning bid. (The bidder does not want to win this auction, but they want to make sure to be invited to the next auction). In Britain and many other countries, rings and other forms of bidding on one's own object are illegal. In Australia, a dummy bid or also a shill is a criminal offence, but a vendor bid or a co-owner bid below the reserve price is permitted if clearly declared as such by the auctioneer. These are all official legal terms in Australia but may have other meanings elsewhere. A co-owner is one of two or several owners (who disagree among themselves). In Sweden and many other countries, there are no legal restrictions, but it will severely hurt the reputation of an auction house that knowingly permits any other bids except genuine bids. If the reserve is not reached this should be clearly declared. In South Africa, auctioneers can use their staff or any bidder to raise the price as long as it is disclosed before the auction sale. Rael Levitt's companies The Auction Alliance controversy focused on vendor bidding and led to its downfall in 2012. There will usually be an estimate of what price the lot will fetch. In an ascending open auction, it is considered important to get at least a 50-percent increase in the bids from start to finish. To accomplish this, the auctioneer must start the auction by announcing a suggested opening bid (SOB) that is low enough to be immediately accepted by one of the bidders. Once there is an opening bid, there will quickly be several other, higher bids submitted. Experienced auctioneers will often select an SOB that is about 45 percent of the (lowest) estimate. Thus there is a certain margin of safety to ensure that there will indeed be a lively auction with many bids submitted. Several observations indicate that the lower the SOB, the higher the final winning bid. This is due to the increase in the number of bidders attracted by the low SOB. A chi-squared distribution shows many low bids but few high bids. Bids "show up together"; without several low bids there will not be any high bids. Another approach to choosing a SOB: The auctioneer may achieve good success by asking the expected final sales price for the item, as this method suggests to the potential buyers the item's particular value. For instance, an auctioneer is about to sell a $1,000 car at a sale. Instead of asking $100, hoping to entice wide interest, the auctioneer may suggest an opening bid of $1,000; although the first bidder may begin bidding at a mere $100, the final bid may more likely approach $1,000. The Journal of Economic Literature (JEL) classification code for auctions is D44.
#redirect [[trade]]
Buy may refer to a trade, i.e., an exchange of goods and services via bartering or a monetary purchase. The term may also refer to:
{{update|date=August 2020}} {{Infobox political division | name = Hong Kong | native_name = | native_name_lang = | settlement_type = [[Special administrative regions of China|Special administrative region]] | official_name = {{raise|0.2em|Hong Kong Special Administrative Region of the People's Republic of China}}{{Infobox |subbox=yes |bodystyle=font-size:90%;font-weight:normal; | rowclass1 = mergedrow |label1=[[Traditional Chinese characters|Chinese]]:|data1={{lang|zh-hant|中華人民共和國香港特別行政區}} | rowclass2 = mergedrow |label2=[[Hong Kong Cantonese|Cantonese]] [[Yale romanisation of Cantonese|Yale romanisation]]: |data2=''Jūng'wàh Yàhnmàhn Guhng'wòhgwok Hēunggóng Dahkbiht Hàhngjingkēui'' }} | image_flag = Flag of Hong Kong.svg | flag_size = 125px | flag_alt = A flag with a white 5-petalled flower design on solid red background | flag_link = Flag of Hong Kong | image_seal = Regional Emblem of Hong Kong.svg | seal_size = 85px | seal_alt = A red circular emblem, with a white 5-petalled flower design in the centre, and surrounded by the words "Hong Kong" and "{{lang|zh-hant|中華人民共和國香港特別行政區}}" | seal_type = Emblem | seal_link = Emblem of Hong Kong | image_map = Hong Kong in China (zoomed) (+all claims hatched).svg | map_alt = Location of Hong Kong | map_caption = Location of Hong Kong within China | mapsize = 250px | subdivision_type = [[Sovereign state]] | subdivision_name = [[China]] | established_title = [[Convention of Chuenpi|British possession]] | established_date = 26 January 1841 | established_title1 = [[Treaty of Nanking]] | established_date1 = 29 August 1842 | established_title2 = [[Convention of Peking]] | established_date2 = 24 October 1860 | established_title3 = [[Convention for the Extension of Hong Kong Territory|New Territories lease]] | established_date3 = 9 June 1898 | established_title4 = [[Japanese occupation of Hong Kong|Imperial Japanese occupation]] | established_date4 = 25 December 1941 to 30 August 1945 | established_title5= [[Sino-British Joint Declaration]] | established_date5 = 19 December 1984 | established_title6 = [[Handover of Hong Kong|Handover to China]] | established_date6 = 1 July 1997 | official_languages = {{hlist|Chinese{{efn|name=chinese-varieties|No specific variety of Chinese is official in the territory. Residents predominantly speak [[Hong Kong Cantonese|Cantonese]], the ''de facto'' regional standard.<ref>{{harvnb|Leung|2016}}.</ref><ref name="OfficialLanguagesOrd">{{harvnb|Official Languages Ordinance}}.</ref><ref name="2016By-CensusLanguages">{{harvnb|Population By-Census|2016|pp=31, 51–52}}</ref>}}|[[Hong Kong English|English]]{{efn|name=language-status|For all government use, documents written using [[Traditional Chinese characters]] are authoritative over ones inscribed with [[Simplified Chinese characters]].<ref>{{harvnb|Legislative Council Disclaimer and Copyright Notice}}</ref> English shares equal status with Chinese in all official proceedings.<ref name="ChineseInCourt">{{harvnb|Use of Chinese in Court Proceedings|2011}}</ref>}}}} | capital_type = [[Administrative centre]] | capital = [[Tamar, Hong Kong|Tamar]] | largest_settlement_type = [[Districts of Hong Kong|district]]<br />{{nobold|by population}} | largest_settlement = [[Sha Tin District|Sha Tin]] | languages_type = [[Regional language]] | languages_sub = yes | languages = [[Hong Kong Cantonese|Cantonese]]{{efn|name=chinese-varieties}} | languages2_type = [[Official script]]s | languages2_sub = yes | languages2 = [[Traditional Chinese]]{{efn|name=language-status}}<br />[[English alphabet]] | demonym = [[Hongkongers|Hongkonger]] | ethnic_groups = 92.0% [[Chinese people|Chinese]]<br />2.5% [[Filipinos in Hong Kong|Filipino]]<br />2.1% [[Indonesians in Hong Kong|Indonesian]]<br />1.1% [[South Asians in Hong Kong|Indian]]<br/>0.8% [[White people|White]]<br />0.3% [[South Asians in Hong Kong|Nepalese]]<br />1.6% Others<ref name="demographics">{{harvnb|Population By-Census|2016|p=46}}.</ref> | ethnic_groups_year = 2016 | government_type = [[Devolution|Devolved]] [[Executive (government)|executive-led]] government within a [[Unitary state|unitary]] [[one-party state]]<ref>{{cite web|title=China (People's Republic of) 1982 (rev. 2004)|url=https://www.constituteproject.org/constitution/China_2004?lang=en|publisher=Constitute project|access-date=25 August 2019}}</ref> | leader_title1 = [[Chief Executive of Hong Kong|Chief Executive]]<!--- DO NOT insert Head of State of PRC here; HK is not a sovereign state and the Hong Kong Government's website (http://www.gov.hk) indicates there is no such position as Head of State of Hong Kong. ---> | leader_name1 = [[John Lee Ka-chiu|John Lee]] | leader_title2 = [[Chief Secretary for Administration|Chief Secretary]] | leader_name2 = [[Eric Chan]] | leader_title3 = [[President of the Legislative Council of Hong Kong|Council President]] | leader_name3 = [[Andrew Leung]] | leader_title4 = [[Chief Justice of the Court of Final Appeal|Chief Justice]] | leader_name4 = [[Andrew Cheung]] | legislature = [[Legislative Council of Hong Kong|Legislative Council]] | national_representation_type1 = [[National People's Congress]] | national_representation1 = [[2022 National People's Congress election in Hong Kong|36 deputies]] | national_representation_type2 = [[Chinese People's Political Consultative Conference|Chinese People's<br />{{nowrap|Political Consultative}}<br />Conference]] | national_representation2 = 203 delegates<ref name="NationalReps">{{harvnb|Cheung|2017}}.</ref> | area_km2 = 2754.97<ref name="landsd area">{{cite web | title=Survey and Mapping Office – Circulars and Publications | url=https://www.landsd.gov.hk/en/resources/mapping-information/hk-geographic-data.html | publisher=Survey and Mapping Office | access-date=20 October 2020}}</ref> | area_sq_mi = 1063.6 <!-- Do not remove as per WP:MOSNUM --> | area_rank = 168th | area_label2 = Land | area_data2 = 1,110.18 km{{smallsup|2}}<br />(428.64 sq mi)<ref name="landsd area"/> <!-- Do not remove as per WP:MOSNUM --> | percent_water = 59.70%<br />(1644.79 km{{smallsup|2}};<br />635.05 sq mi)<ref name="landsd area"/> | elevation_max_m = 957 | elevation_max_point = [[Tai Mo Shan]] | elevation_min_m = 0 | elevation_min_point = [[South China Sea]] | population_estimate = {{decreaseNeutral}} 7,291,600<ref>{{cite press release |title=Mid-year population for 2022 |date=11 August 2022 |publisher=[[Government of Hong Kong]] |url=https://www.info.gov.hk/gia/general/202208/11/P2022081100393.htm |access-date=4 September 2022 }}</ref> | population_estimate_year = 2022 | population_estimate_rank = | population_census = {{increaseNeutral}} 7,413,070<ref>{{cite web|url=https://www.census2021.gov.hk/doc/media/Table(EN).pdf |title= Key statistics of the 2021 and 2011 Population Census |publisher=census2021.gov.hk |access-date=15 March 2022}}</ref> | population_census_year = 2021 | population_density_km2 = 6,801<ref>{{cite web|url=https://www.census2021.gov.hk/en/main_tables.html |title= Main Tables – 2021 Population Census |publisher=census2021.gov.hk |access-date=15 March 2022}}</ref> | population_density_sq_mi = 17,614 <!-- Do not remove as per WP:MOSNUM --> | population_density_rank = 4th | GDP_PPP = {{increase}} $518.743&nbsp;billion<ref name="IMFWEOHK">{{cite web |url=https://www.imf.org/en/Publications/WEO/weo-database/2022/October |title=World Economic Outlook Database, October 2022 |website=IMF.org |publisher=[[International Monetary Fund]] |access-date=21 October 2022}}</ref> | GDP_PPP_year = 2022 | GDP_PPP_rank = 48th | GDP_PPP_per_capita = {{increase}} $69,987<ref name="IMFWEOHK"/> | GDP_PPP_per_capita_rank = 11th | GDP_nominal = {{decrease}} $368.373&nbsp;billion<ref name="IMFWEOHK"/> | GDP_nominal_year = 2022 | GDP_nominal_rank = 43rd | GDP_nominal_per_capita = {{decrease}}$49,700<ref name="IMFWEOHK"/> | GDP_nominal_per_capita_rank = 18th | Gini = 53.9 <!-- number only --><!--- DO NOT USE CIA World Factbook. The Gini index is a parameter in calculating the HDI so the Gini index and the HDI should be from the same source, i.e. Human Development Report 2009. ---> | Gini_year = 2016 | Gini_change = increase <!-- increase/decrease/steady --> | Gini_ref =<ref name="GiniRef">{{harvnb|Household Income Distribution|2016|p=7}}</ref> | HDI = 0.952 <!-- number only --> | HDI_year = 2021<!-- Please use the year to which the data refers, not the publication year--> | HDI_change = increase<!-- increase/decrease/steady --> | HDI_ref = <ref name="HDI">{{cite web|url=https://hdr.undp.org/system/files/documents/global-report-document/hdr2021-22pdf_1.pdf|title=Human Development Report 2021/2022|language=en|publisher=[[United Nations Development Programme]]|date=8 September 2022|access-date=8 September 2022}}</ref> | HDI_rank = 4th | currency = [[Hong Kong dollar]] (HK$) | currency_code = HKD | timezone = [[Hong Kong Time|HKT]] | utc_offset = +08:00 | date_format = dd/mm/yyyy<br />yyyy年mm月dd日 | electricity = 220 V–50 Hz | drives_on = left{{efn|Except for the [[Hong Kong–Zhuhai–Macau Bridge#Hong Kong Link Road|Hong Kong-Zhuhai-Macao Bridge Hong Kong Link Road]], which drives on the right.<ref>{{harvnb|Technical Legislative Amendments on Traffic Arrangements for the Hong Kong-Zhuhai-Macao Bridge|2017}}</ref>}} | calling_code = [[Telephone numbers in Hong Kong|+852]] | blank_name_sec1 = [[Vehicle registration plates of China#Cross-border with Hong Kong and Macau|License plate prefixes]] | blank_info_sec1 = None for local vehicles, {{lang|zh-cn|粤Z}} for cross-boundary vehicles | iso_code = {{hlist|[[ISO 3166-2:HK|HK]]|[[ISO 3166-2:CN|CN-HK]]}} | cctld = {{hlist|[[.hk]]|[[.香港]]}} }} '''Hong Kong''' ({{IPAc-en|audio=En-ca-Hong Kong.oga|ˌ|h|ɒ|ŋ|ˈ|k|ɒ|ŋ|}}; {{zh|香港}}, {{small|[[Hong Kong Cantonese]]:}} {{IPA-yue|hœ́ːŋ.kɔ̌ːŋ||Yue-heung1gong2.ogg}}, literally "Fragrant Port"), officially '''The Hong Kong Special Administrative Region of The People's Republic of China''',<ref>{{Cite web|title=Basic Law|url=https://www.basiclaw.gov.hk/en/basiclaw/basiclaw.html|access-date=2021-12-12|work=www.basiclaw.gov.hk}}</ref> is one of two [[Special Administrative Region]]s (SARs) of the [[People's Republic of China]] (the other is [[Macau]]). It is one of the richest and most developed parts in the world, and one where the [[cost of living]] is one of the highest. Hong Kong grew quickly in the decades after [[World War II]], becoming a famous world-class financial centre. Hong Kong was one of the last territories of the [[British Empire]] until 1997, when United Kingdom handed it over to China. China then promised to grant it special status for 50 years. The population of Hong Kong is over seven million. The economy has rapidly grown from a trading port to a very rich city. Hong Kong has the most skyscrapers in the world. Hong Kong is divided into 3 main parts: * [[Hong Kong Island]] * [[Kowloon]] * [[New Territories]] (including 235 outlying [[island]]s) Hong Kong was a British colony from 1842 to 1997 as China surrendered the city after losing the [[Second Opium War]]. After the [[Transfer of sovereignty over Hong Kong|handover]] in 1997, Hong Kong became under Chinese control under special status. Hong Kong has its own [[constitution]] that is different from that of the People's Republic of China (PRC). == Climate == Hong Kong is in a sub-tropical area, and has [[monsoon]] winds. It is cool and wet in [[winter]] (January-March), hot and rainy from [[Spring (season)|spring]] through [[summer]] (April-September), and warm, sunny and dry in the [[autumn]] (October-December). The rainy [[season]] is from May until September. In summer and early autumn, there is a frequent threat of [[typhoons]]. == Population and language == The population of Hong Kong reached 7.5 million (7,503,100) in 2023. Most of the people in Hong Kong are [[China|Chinese]]. Hong Kong is one of the most densely populated countries in the world. It has an overall [[population density]] of 6,300 people per square kilometre. Hong Kong has one of the world’s lowest birth rates — 1.05 per woman at an age of being capable to give birth as of 2019. This is far below the rate needed to replace each person, 2.1. People from Hong Kong mainly speak [[Cantonese language|Cantonese]]. Students are required to learn [[English language|English]] at school. Ever since Hong Kong became a part of [[China]], the number of people who speak [[Mandarin language|Mandarin]] has increased because Mandarin is the official language of China. == Public holidays == In Hong Kong, 17 days of the year are [[public holiday]]s: * 1 January - [[New Year's Day]], the beginning of the year * [[Chinese New Year]] (3 public holidays) * Ching Ming Festival, also known as "Tomb Sweeping Day" * [[Easter]] (3 public holidays) * 1 May - [[Labour Day]] * [[Vesak|The Buddha's birthday]] * [[Dragon Boat Festival]] * 1 July - Hong Kong SAR Establishment Day * The day after [[Mid-Autumn Festival]] * 1 Oct - [[National Day of the People's Republic of China|Chinese National Day]] * Chung Yeung Festival - a day when people honour their ancestors, similar to "Tomb Sweeping Day", and people often do something for the olds who is still living * 25 December - [[Christmas]] * 26 December - [[Boxing Day]] == Currency == There are [[coin]]s from 10 cents to 10 Hong Kong Dollars and bank-notes (paper money bills) from $10 to $1000. One American [[dollar]] equals to HK$7.80±0.05 Hong Kong Dollars, according to the official pegged exchange rate. == Public transport == * [[MTR|Mass Transit Railway]] (MTR) which was established in 1979, owns seven lines (10 lines after KCR merger), ** Kwun Tong Line (running between Whampoa and Tiu Keng Leng) ** Tsuen Wan Line (running between Tsuen Wan and Central) ** Island Line (running between Kennedy Town and Chai Wan) ** South Island Line (running between Admiralty and South Horizons) ** Tseung Kwan O Line (running between North Point and Po Lam/LOHAS Park) ** Tung Chung Line (running between Tung Chung and Hong Kong) ** Airport Express (running between AsiaWorld-Expo and Hong Kong) ** Disneyland Resort Line (running between Sunny Bay and Disneyland Resort) ** East Rail Line (running between Lo Wu/Lok Ma Chau and Admiralty) ** Tuen Ma Line (running between Tuen Mun and Wu Kai Sha) * Bus: there are four major bus companies in Hong Kong, namely KMB which mainly serves [[Kowloon]] and [[New Territories]], [[New World First Bus]], which mainly serves [[Hong Kong Island]], [[New Lantau Bus]], which <nowiki/>mainly serves [[Lantau Island]], and [[Citybus]], which mainly serves [[Hong Kong International Airport|airport]] routes. * Public Light Bus: in the 1960s, it was an [[illegal]] [[transportation]], but later on, the [[government]] noticed that if there were only buses in Hong Kong, then some [[villages]] in N.T. will not have a [[public transport]]. Therefore, the [[Legislative Council]] made it legal and under government regulation. * Tram (running between Kennedy Town, Happy Valley, and Shau Kei Wan) * [[Peak Tram]] (running between Garden Road and Victoria Peak) * Taxi [Red ([[Kowloon]] and [[Hong Kong Island]][[Kowloon|)]], Green ([[New Territories]]), and Blue ([[Lantau Island]])] * Ferry (Lots of different companies, including the [[Star Ferry]], First Ferry and Fortune Ferry) * [[Hong Kong International Airport]] is the main airport. It is the [[List of busiest airports in the world|busiest airport in the world]] in terms of [[Cargo|cargo traffic]]. ==Education== Hong Kong's education system is mostly based on the [[Education in England|English]] system. Children are required to be in school from age 6 until completing [[secondary school]] (high school) at age 18. ===Universities=== Hong Kong has 11 Universities: *[[University of Hong Kong]] *[[Chinese University of Hong Kong]] *[[Hong Kong University of Science and Technology]] *[[Hong Kong Polytechnic University]] *[[City University of Hong Kong]] *[[Hong Kong Baptist University]] *[[Education University of Hong Kong]] *[[Lingnan University]] *[[Hong Kong Metropolitan University]] *[[Hong Kong Shue Yan University]] *[[Hang Seng University of Hong Kong]] ==Housing in Hong Kong== According to the International Housing Affordability survey, Hong Kong has the most [[affordable housing|unaffordable housing market]] in the world since 2010.<ref>{{Cite web|url=https://www.fortunebuilders.com/worlds-unaffordable-housing-markets-34656/|title=The World's Most Unaffordable Housing Markets|date=2014-01-24|website=FortuneBuilders|access-date=2020-06-24|archive-date=2020-06-25|archive-url=https://web.archive.org/web/20200625070054/https://www.fortunebuilders.com/worlds-unaffordable-housing-markets-34656/|url-status=dead}}</ref><ref>{{Cite web|url=https://www.bloomberg.com/news/articles/2019-01-21/hong-kong-housing-ranked-world-s-least-affordable-for-9th-year|title=Hong Kong Housing Is World's Least Affordable for 9th Year|last=Kwan|first=Shawna|date=21 January 2019|website=[[Bloomberg News]]|access-date=21 June 2020}}</ref> == Timeline of Hong Kong == Here is a brief history of Hong Kong: [[File:Rocks at old age.JPG|thumb|Some very old rocks found in Hong Kong]] '''Around 4000 BC''' * Sea levels rose above 100&nbsp;meters '''Around 3500 BC''' * Ceramic forms decorated with a wide range of patterns '''Around 2000 BC''' * Bronze weapons, knives, arrowheads & tools. * Metal worked locally '''Around 500 BC''' * Ancient Chinese writing developed '''221 BC''' * People from [[Mainland China]] came to Hong Kong '''220 BC''' * Coins from the Chinese Han period were used in Hong Kong '''1555''' * A Portuguese named Jorge Álvares was the first European to reach Hong Kong '''1799''' * China banned the drug trade in Hong Kong '''1800''' * Opium became a huge business [[File:Opium.jpg|thumb|Opium flower]] '''1839''' * Lin Zexu was appointed Special Commissioner * The First Opium war began '''1841''' * Hong Kong was given to the British and became a [[British Hong Kong|dependent territory of United Kingdom]] * Lord Palmerston wrote that Hong Kong was a barren island with only a few houses on it ''' January 26, 1841''' * The British flag was raised at Possession Point, on Hong Kong Island '''August 1841''' * Sir Henry Pottinger became Hong Kong's first governor * The Treaty of Nanjing was signed, ending the First Opium War '''1860''' * China was defeated in the Second Opium War. Boundary Street and Stonecutter's Island were leased to Britain '''1888''' * The Peak Tram started operating on Hong Kong Island '''1898''' * Lantau Island and the New Territories were leased to the British for 99 years '''1900s''' * Hong Kong became a refuge for exiles from China '''1920s-1930s''' * Western dress began to come in fashion for the locals '''1933-1934''' * Father Daniel Finn began excavations on Lamma Island '''1941''' * Refugees fleeing the Chinese Communist Party came to Hong Kong [[File:HKUrbanCouncil.svg|thumb|[[Bauhinia × blakeana]] was adopted as the [[National emblem|floral emblem]] of Hong Kong by the [[Urban Council, Hong Kong|Urban Council]] in 1965.]] [[File:Urban Council of Hong Kong Armorial Bearings.png|thumb|The armorial bearings granted by the [[College of Arms]], appointed by the [[British Sovereign]], a part of the [[Royal Household of the United Kingdom]].]] [[File:Regional Emblem of Hong Kong.svg|thumb|Hong Kong Special Administrative Region of the People's Republic of China's emblem after it was given to China since 1997. The Chinese name of [[Bauhinia × blakeana]] has also been frequently shortened as 紫荊/紫荆 (洋 ''yáng'' means "foreign" in Chinese, it's also a bit unfriendly, and this would be deemed inappropriate by the PRC government), although 紫荊/紫荆 refers to another genus called ''[[Cercis]]''. A [[statue]] of the plant has been erected in [[Golden Bauhinia Square]] in Hong Kong. Although the flowers are bright pinkish purple in colour, they are depicted in white on the Flag of Hong Kong.]] '''December 8, 1941''' * The [[Empire of Japan]] invaded Hong Kong '''December 25, 1941''' * The British authorities surrendered Hong Kong to the Japanese Army '''August 1945''' * Britain reclaimed its territory after Japan's surrender '''1949''' * Double-decker buses were introduced to Hong Kong '''1950''' * Hong Kong became a free port '''1953''' * The Shek Kip Mei Estate was built, establishing the program of public housing '''1955''' * A Han period tomb was discovered near Lei Cheng Uk '''1983''' * The Hong Kong dollar was tied to the US dollar '''1984''' * China and Britain signed the Sino-British Joint Declaration '''1990''' * The Hong Kong Basic Law was confirmed '''1997''' * Asia's financial crisis * Archaeologists discovered 20 graves on the island of Ma Wa * Tung Chee Wa elected as the first [[Chief Executive]] of Hong Kong. Voting was conducted by 400 committees of an electoral college whose members are appointed by the Chinese Government. '''July 1, 1997''' * Hong Kong became a [[Special Administrative Region]] of China for 50 years '''1998''' * [[Hong Kong International Airport]] opened, replacing [[Kai Tak Airport]] in Kowloon '''June 2002''' * Tung Chee Hwa was elected as [[Chief Executive]] for a second term. '''2003''' * Citizens wanted a more democratic and republican system * The [[SARS]] epidemic began '''March 10, 2005''' * Tung Chee Hwa resigned as chief executive because of health problems. '''June 16, 2005''' * Donald Tsang Yam-kuen was elected unopposed as [[Chief Executive]]. '''March 2012''' * Leung Chun Ying was elected as [[Chief Executive]]. '''2014''' * People occupied the Central region to demand [[universal suffrage]] for the next chief executive election, to take place in 2017. '''2015''' * The government voted against the universal suffrage demanded by the people. '''2016''' * There were more protests in [[Mong Kok]] because the government voted against universal suffrage. Police had to use [[pepper spray]] and [[Tear gas]] on the protesters to get them to leave. '''2019''' * [[Protestors]] in Hong Kong demonstrated against a new [[extradition]] law proposed by the Hong Kong government.<ref>{{Cite news|url=https://nationalpost.com/news/world/black-clad-anti-extradition-protesters-flood-streets-of-hong-kong|title=Black-clad, anti-extradition protesters flood streets of Hong Kong|website=National Post|date=21 June 2019|language=en-CA|access-date=2020-06-24|last1=Master|first1=Farah}}</ref><ref>{{Cite news|url=https://www.amnesty.org/en/latest/news/2019/09/hong-kong-protests-explained/|title=HONG KONG'S PROTESTS EXPLAINED|date=9 September 2019|work=[[Amnesty International]]}}</ref>It was the largest [[2019 Hong Kong protests|protest]] in Hong Kong's history.<ref>{{Cite web|url=https://www.theepochtimes.com/by-involving-north-korea-and-iran-beijing-tries-to-divert-public-attention-from-mass-protest-in-hong-kong_2968897.html|title=By Using North Korea and Iran, Beijing Seeks to Divert Attention From Hong Kong Mass Protests|date=2019-06-18|website=www.theepochtimes.com|language=en-US|access-date=2020-06-24}}</ref> This law would allow Hong Kong to send people to other jurisdictions where they have committed a crime, including mainland China.<ref>*{{cite news |url = https://www.aljazeera.com/news/2019/06/hong-kong-rallies-govt-fails-act-protester-demands-190621015118139.html |title = Thousands of protesters again hit Hong Kong's streets |author = Euan McKirdy |publisher = Al Jazeera |date = June 21, 2019 }}</ref> People opposed this because of the poor reputation of China due to allegations of [[torture]], forced confessions and arbitrary detentions. There were fears that the bill would just bring Hong Kong closer under China's control, even though Hong Kong have their own judicial system.<ref>{{Cite news|url=https://www.bbc.com/news/world-asia-china-48715221|title=Hong Kong protests: Thousands surround police headquarters|date=21 June 2019|work=BBC News|access-date=2020-06-24|language=en-GB}}</ref> '''2020''' * The [[Hong Kong National Security Law]] is created. This law will give [[Beijing]], more power to control Hong Kong. Beijing also put a security office in Hong Kong. The law was created at 11 P.M. on 30 June 2020, Hong Kong Time.<ref> {{Cite news|date=2020-06-30|title=Hong Kong security law: What is it and is it worrying?|language=en-GB|work=BBC News|url=https://www.bbc.com/news/world-asia-china-52765838|access-date=2020-11-10}}</ref> == Media == Hong Kong has a few media companies, notably [[TVB|Television Broadcasts Limited]] (TVB), [[HK Television Entertainment]] (HKTVE), and [[Fantastic Television]]. Access to television is not affected by Chinese regulations, such as the [[Great Firewall]], which filters and blocks certain programs. == Places in Hong Kong == * [[Disneyland Hong Kong|Hong Kong Disneyland]] * [[Victoria Peak]] * [[Victoria Park (Hong Kong)|Victoria Park]] * [[Ocean Park]] * [[Man Mo Temple]] * [[Repulse Bay]] * [[Hong Kong Park]] * [[Yuen Po Street Bird Garden]] * [[Hong Kong Museum of History]] * [[Hong Kong Space Museum]] * [[Hong Kong Museum of Science & Technology]] * [[Wong Tai Sin Temple (Hong Kong)|Wong Tai Sin Temple]] * [[Fung Ying Sin Koon]] * [[Kowloon Walled City]] * [[Kowloon Mosque and Islamic Centre]] * [[Avenue of Stars, Hong Kong]] * [[Sham Tung Uk]] * [[Po Lin Monastery]] and the [[Big Buddha]] on [[Lantau Island]] (currently the largest bronze Buddha in world) * [[Cheung Chau]] ([[Cheung Chau|Long Island]]) * [[Chek Lap Kok]] * [[Clock Tower, Hong Kong]] ==Notes== {{notelist}} == References == {{reflist}} ==Other websites== {{Commonscat|Hong Kong}} * {{cite web |url=https://www.cia.gov/library/publications/the-world-factbook/geos/hk.html |title=Hong Kong |work=The World Factbook |publisher=CIA |date=23 August 2010 |access-date=17 September 2010 |archive-date=14 May 2009 |archive-url=https://www.webcitation.org/5glwctzKV?url=https://www.cia.gov/library/publications/the-world-factbook/geos/hk.html |url-status=dead }} {{Province-level divisions of China}} {{Major cities of Greater China}} {{Asia}} {{authority control}} [[Category:Hong Kong| ]]
Hong Kong (US: /ˈhɒŋkɒŋ/ or UK: /hɒŋˈkɒŋ/; Chinese: 香港; Jyutping: hoeng1 gong2, Cantonese: [hœ́ːŋ.kɔ̌ːŋ] ), officially the Hong Kong Special Administrative Region of the People's Republic of China (abbr. Hong Kong SAR or HKSAR), is a city and a special administrative region in China. With 7.4 million residents of various nationalities in a 1,104-square-kilometre (426 sq mi) territory, Hong Kong is one of the most densely populated territories in the world. Hong Kong was established as a colony of the British Empire after the Qing Empire ceded Hong Kong Island in 1841–1842 as a consequence of losing the First Opium War. The colony expanded to the Kowloon Peninsula in 1860 and was further extended when the United Kingdom obtained a 99-year lease of the New Territories in 1898. Hong Kong was occupied by Japan from 1941 to 1945 during World War II. The whole territory was transferred from the United Kingdom to China in 1997. Hong Kong maintains separate governing and economic systems from that of mainland China under the principle of "one country, two systems". Originally a sparsely populated area of farming and fishing villages, the territory is now one of the world's most significant financial centres and commercial ports. Hong Kong is the world's fourth-ranked global financial centre, ninth-largest exporter, and eighth-largest importer. Its currency, the Hong Kong dollar, is the ninth most traded currency in the world. Home to the second-highest number of billionaires of any city in the world, Hong Kong has the largest number of ultra high-net-worth individuals. Although the city has one of the highest per capita incomes in the world, severe income inequality exists among the population. Despite having the largest number of skyscrapers of any city in the world, housing in Hong Kong has been well-documented to experience a chronic persistent shortage. Hong Kong is a highly developed territory and has a Human Development Index (HDI) of 0.952, ranking fourth in the world. The city has the highest life expectancy in the world, and a public transport rate exceeding 90%. The name of the territory, first romanised as "He-Ong-Kong" in 1780, originally referred to a small inlet located between Aberdeen Island and the southern coast of Hong Kong Island. Aberdeen was an initial point of contact between British sailors and local fishermen. Although the source of the romanised name is unknown, it is generally believed to be an early phonetic rendering of the Cantonese (or Tanka Cantonese) phrase hēung góng. The name translates as "fragrant harbour" or "incense harbour". "Fragrant" may refer to the sweet taste of the harbour's freshwater influx from the Pearl River or to the odour from incense factories lining the coast of northern Kowloon. The incense was stored near Aberdeen Harbour for export before Victoria Harbour was developed. Sir John Davis (the second colonial governor) offered an alternative origin; Davis said that the name derived from "Hoong-keang" ("red torrent"), reflecting the colour of soil over which a waterfall on the island flowed. The simplified name Hong Kong was frequently used by 1810. The name was also commonly written as the single word Hongkong until 1926, when the government officially adopted the two-word name. Some corporations founded during the early colonial era still keep this name, including Hongkong Land, Hongkong Electric Company, Hongkong and Shanghai Hotels and the Hongkong and Shanghai Banking Corporation (HSBC). Earliest known human traces in what is now Hong Kong are dated by some to 35,000 and 39,000 years ago during the Paleolithic period. The claim is based on an archaeological investigation in Wong Tei Tung, Sai Kung in 2003. The archaeological works revealed knapped stone tools from deposits that were dated using optical luminescence dating. During the Middle Neolithic period, about 6,000 years ago, the region had been widely occupied by humans. Neolithic to Bronze Age Hong Kong settlers were semi-coastal people. Early inhabitants are believed to be Austronesians in the Middle Neolithic period and later the Yueh people. As hinted by the archaeological works in Sha Ha, Sai Kung, rice cultivation had been introduced since Late Neolithic period. Bronze Age Hong Kong featured coarse pottery, hard pottery, quartz and stone jewelry, as well as small bronze implements. The Qin dynasty incorporated the Hong Kong area into China for the first time in 214 BCE, after conquering the indigenous Baiyue. The region was consolidated under the Nanyue kingdom (a predecessor state of Vietnam) after the Qin collapse and recaptured by China after the Han conquest. During the Mongol conquest of China in the 13th century, the Southern Song court was briefly located in modern-day Kowloon City (the Sung Wong Toi site) before its final defeat in the 1279 Battle of Yamen by the Yuan Dynasty. By the end of the Yuan dynasty, seven large families had settled in the region and owned most of the land. Settlers from nearby provinces migrated to Kowloon throughout the Ming dynasty. The earliest European visitor was Portuguese explorer Jorge Álvares, who arrived in 1513. Portuguese merchants established a trading post called Tamão in Hong Kong waters and began regular trade with southern China. Although the traders were expelled after military clashes in the 1520s, Portuguese-Chinese trade relations were re-established by 1549. Portugal acquired a permanent lease for Macau in 1557. After the Qing conquest, maritime trade was banned under the Haijin policies. From 1661 to 1683, the population of most of the area forming present day Hong Kong was cleared under the Great Clearance, turning the region into a wasteland. The Kangxi Emperor lifted the maritime trade prohibition, allowing foreigners to enter Chinese ports in 1684. Qing authorities established the Canton System in 1757 to regulate trade more strictly, restricting non-Russian ships to the port of Canton. Although European demand for Chinese commodities like tea, silk, and porcelain was high, Chinese interest in European manufactured goods was insignificant, so that Chinese goods could only be bought with precious metals. To reduce the trade imbalance, the British sold large amounts of Indian opium to China. Faced with a drug crisis, Qing officials pursued ever more aggressive actions to halt the opium trade. In 1839, the Daoguang Emperor rejected proposals to legalise and tax opium and ordered imperial commissioner Lin Zexu to eradicate the opium trade. The commissioner destroyed opium stockpiles and halted all foreign trade, triggering a British military response and the First Opium War. The Qing surrendered early in the war and ceded Hong Kong Island in the Convention of Chuenpi. British forces began controlling Hong Kong shortly after the signing of the convention, from 26 January 1841. However, both countries were dissatisfied and did not ratify the agreement. After more than a year of further hostilities, Hong Kong Island was formally ceded to the United Kingdom in the 1842 Treaty of Nanking. Administrative infrastructure was quickly built by early 1842, but piracy, disease, and hostile Qing policies initially prevented the government from attracting commerce. Conditions on the island improved during the Taiping Rebellion in the 1850s, when many Chinese refugees, including wealthy merchants, fled mainland turbulence and settled in the colony. Further tensions between the British and Qing over the opium trade escalated into the Second Opium War. The Qing were again defeated and forced to give up Kowloon Peninsula and Stonecutters Island in the Convention of Peking. By the end of this war, Hong Kong had evolved from a transient colonial outpost into a major entrepôt. Rapid economic improvement during the 1850s attracted foreign investment, as potential stakeholders became more confident in Hong Kong's future. The colony was further expanded in 1898 when the United Kingdom obtained a 99-year lease of the New Territories. The University of Hong Kong was established in 1911 as the territory's first institution of higher education. Kai Tak Airport began operation in 1924, and the colony avoided a prolonged economic downturn after the 1925–26 Canton–Hong Kong strike. At the start of the Second Sino-Japanese War in 1937, Governor Geoffry Northcote declared Hong Kong a neutral zone to safeguard its status as a free port. The colonial government prepared for a possible attack, evacuating all British women and children in 1940. The Imperial Japanese Army attacked Hong Kong on 8 December 1941, the same morning as its attack on Pearl Harbor. Hong Kong was occupied by Japan for almost four years before the British resumed control on 30 August 1945. Its population rebounded quickly after the war, as skilled Chinese migrants fled from the Chinese Civil War and more refugees crossed the border when the Chinese Communist Party took control of mainland China in 1949. Hong Kong became the first of the Four Asian Tiger economies to industrialise during the 1950s. With a rapidly increasing population, the colonial government attempted reforms to improve infrastructure and public services. The public-housing estate programme, Independent Commission Against Corruption, and Mass Transit Railway were all established during the post-war decades to provide safer housing, integrity in the civil service, and more reliable transportation. Nevertheless, widespread public discontent resulted in multiple protests from the 1950s to 1980s, including pro-Republic of China and pro-Chinese Communist Party protests. In the 1967 Hong Kong riots, pro-PRC protestors clashed with the British colonial government. As many as 51 were killed and 802 were injured in the violence, including dozens killed by the Royal Hong Kong Police via beatings and shootings. Although the territory's competitiveness in manufacturing gradually declined because of rising labour and property costs, it transitioned to a service-based economy. By the early 1990s, Hong Kong had established itself as a global financial centre and shipping hub. The colony faced an uncertain future as the end of the New Territories lease approached, and Governor Murray MacLehose raised the question of Hong Kong's status with Deng Xiaoping in 1979. Diplomatic negotiations with China resulted in the 1984 Sino-British Joint Declaration, in which the United Kingdom agreed to transfer the colony in 1997 and China would guarantee Hong Kong's economic and political systems for 50 years after the transfer. The impending transfer triggered a wave of mass emigration as residents feared an erosion of civil rights, the rule of law, and quality of life. Over half a million people left the territory during the peak migration period, from 1987 to 1996. The Legislative Council became a fully elected legislature for the first time in 1995 and extensively expanded its functions and organisations throughout the last years of the colonial rule. Hong Kong was transferred to China on 1 July 1997, after 156 years of British rule. Immediately after the transfer, Hong Kong was severely affected by several crises. The Hong Kong government was forced to use substantial foreign exchange reserves to maintain the Hong Kong dollar's currency peg during the 1997 Asian financial crisis, and the recovery from this was muted by an H5N1 avian-flu outbreak and a housing surplus. This was followed by the 2003 SARS epidemic, during which the territory experienced its most serious economic downturn. Political debates after the transfer of sovereignty have centred around the region's democratic development and the Chinese central government's adherence to the "one country, two systems" principle. After reversal of the last colonial era Legislative Council democratic reforms following the handover, the regional government unsuccessfully attempted to enact national security legislation pursuant to Article 23 of the Basic Law. The central government decision to implement nominee pre-screening before allowing chief executive elections triggered a series of protests in 2014 which became known as the Umbrella Revolution. Discrepancies in the electoral registry and disqualification of elected legislators after the 2016 Legislative Council elections and enforcement of national law in the West Kowloon high-speed railway station raised further concerns about the region's autonomy. In June 2019, mass protests erupted in response to a proposed extradition amendment bill permitting the extradition of fugitives to mainland China. The protests are the largest in Hong Kong's history, with organisers claiming to have attracted more than three million Hong Kong residents. The Hong Kong regional government and Chinese central government responded to the protests with a number of administrative measures to quell dissent. In June 2020, the Legislative Council passed the National Anthem Ordinance, which criminalised "insults to the national anthem of China". The Chinese central government meanwhile enacted the Hong Kong national security law to help quell protests in the region. Nine months later, in March 2021, the Chinese central government introduced amendments to Hong Kong's electoral system, which included the reduction of directly elected seats in the Legislative Council and the requirement that all candidates be vetted and approved by a Beijing-appointed Candidate Eligibility Review Committee. In May 2023, the Legislative Council introduced legislation to reduce the number of directly elected seats in the district councils as well, and a District Council Eligibility Review Committee was similarly established to vet candidates. Hong Kong is a special administrative region of China, with executive, legislative, and judicial powers devolved from the national government. The Sino-British Joint Declaration provided for economic and administrative continuity through the transfer of sovereignty, resulting in an executive-led governing system largely inherited from the territory's history as a British colony. Under these terms and the "one country, two systems" principle, the Basic Law of Hong Kong is the regional constitution. The regional government is composed of three branches: The chief executive is the head of government and serves for a maximum of two five-year terms. The State Council (led by the Premier of China) appoints the chief executive after nomination by the Election Committee, which is composed of 1500 business, community, and government leaders. The Legislative Council has 90 members, each serving a four-year term. Twenty are directly elected from geographical constituencies, thirty-five represent functional constituencies (FC), and forty are chosen by an election committee consisting of representatives appointed by the Chinese central government. Thirty FC councillors are selected from limited electorates representing sectors of the economy or special interest groups, and the remaining five members are nominated from sitting district council members and selected in region-wide double direct elections. All popularly elected members are chosen by proportional representation. The 30 limited electorate functional constituencies fill their seats using first-past-the-post or instant-runoff voting. Twenty-two political parties had representatives elected to the Legislative Council in the 2016 election. These parties have aligned themselves into three ideological groups: the pro-Beijing camp (the current government), the pro-democracy camp, and localist groups. The Chinese Communist Party does not have an official political presence in Hong Kong, and its members do not run in local elections. Hong Kong is represented in the National People's Congress by 36 deputies chosen through an electoral college and 203 delegates in the Chinese People's Political Consultative Conference appointed by the central government. Chinese national law does not generally apply in the region, and Hong Kong is treated as a separate jurisdiction. Its judicial system is based on common law, continuing the legal tradition established during British rule. Local courts may refer to precedents set in English law and overseas jurisprudence. However, mainland criminal procedure law applies to cases investigated by the Office for Safeguarding National Security of the CPG in the HKSAR. Interpretative and amending power over the Basic Law and jurisdiction over acts of state lie with the central authority, making regional courts ultimately subordinate to the mainland's socialist civil law system. Decisions made by the Standing Committee of the National People's Congress override any territorial judicial process. Furthermore, in circumstances where the Standing Committee declares a state of emergency in Hong Kong, the State Council may enforce national law in the region. The territory's jurisdictional independence is most apparent in its immigration and taxation policies. The Immigration Department issues passports for permanent residents which differ from those of the mainland or Macau, and the region maintains a regulated border with the rest of the country. All travellers between Hong Kong and China and Macau must pass through border controls, regardless of nationality. Mainland Chinese citizens do not have right of abode in Hong Kong and are subject to immigration controls. Public finances are handled separately from the national government; taxes levied in Hong Kong do not fund the central authority. The Hong Kong Garrison of the People's Liberation Army is responsible for the region's defence. Although the Chairman of the Central Military Commission is supreme commander of the armed forces, the regional government may request assistance from the garrison. Hong Kong residents are not required to perform military service, and current law has no provision for local enlistment, so its defence is composed entirely of non-Hongkongers. The central government and Ministry of Foreign Affairs handle diplomatic matters, but Hong Kong retains the ability to maintain separate economic and cultural relations with foreign nations. The territory actively participates in the World Trade Organization, the Asia-Pacific Economic Cooperation forum, the International Olympic Committee, and many United Nations agencies. The regional government maintains trade offices in Greater China and other nations. The imposition of the Hong Kong national security law by the central government in Beijing in June 2020 resulted in the suspension of bilateral extradition treaties by the United Kingdom, Canada, Australia, New Zealand, Finland, and Ireland. The United States ended its preferential economic and trade treatment of Hong Kong in July 2020 because it was no longer able to distinguish Hong Kong as a separate entity from the People's Republic of China. The territory is divided into 18 districts, each represented by a district council. These advise the government on local issues such as public facility provisioning, community programme maintenance, cultural promotion, and environmental policy. As of 2019, there are a total of 479 district council seats, 452 of which are directly elected. Rural committee chairmen, representing outlying villages and towns, fill the 27 non-elected seats. In May 2023, the government proposed reforms to the District Council electoral system which further cut the number of directly elected seats from 452 to 88, and total seats from 479 to 470. A requirement that district council candidates be vetted and approved by the District Council Eligibility Review Committee was also proposed. The Legislative Council approved the reforms in July 2023. Hong Kong is governed by a hybrid regime that is not fully representative of the population. Legislative Council members elected by functional constituencies composed of professional and special interest groups are accountable to these narrow corporate electorates and not the general public. This electoral arrangement has guaranteed a pro-establishment majority in the legislature since the transfer of sovereignty. Similarly, the chief executive is selected by establishment politicians and corporate members of the Election Committee rather than directly elected. Although universal suffrage for the chief executive and all Legislative Council elections are defined goals of Basic Law Articles 45 and 68, the legislature is only partially directly elected, and the executive continues to be nominated by an unrepresentative body. The government has been repeatedly petitioned to introduce direct elections for these positions. Ethnic minorities (except those of European ancestry) have marginal representation in government and often experience discrimination in housing, education, and employment. Employment vacancies and public service appointments frequently have language requirements which minority job seekers do not meet, and language education resources remain inadequate for Chinese learners. Foreign domestic helpers, predominantly women from the Philippines and Indonesia, have little protection under regional law. Although they live and work in Hong Kong, these workers are not treated as ordinary residents and do not have the right of abode in the territory. Sex trafficking in Hong Kong is an issue. Local and foreign women and girls are often forced into prostitution in brothels, homes, and businesses in the city. The Joint Declaration guarantees the Basic Law of Hong Kong for 50 years after the transfer of sovereignty. It does not specify how Hong Kong will be governed after 2047, and the central government's role in determining the territory's future system of government is the subject of political debate and speculation. Hong Kong's political and judicial systems may be integrated with China's at that time, or the territory may continue to be administered separately. However, in response to large-scale protests in 2019 and 2020, the Standing Committee of the National People's Congress passed the controversial Hong Kong national security law. The law criminalises secession, subversion, terrorism and collusion with foreign elements and establishes the Office for Safeguarding National Security of the CPG in the HKSAR, an investigative office under Central People's Government authority immune from HKSAR jurisdiction. Some of the aforementioned acts were previously considered protected speech under Hong Kong law. The United Kingdom considers the law to be a serious violation of the Joint Declaration. In October 2020, Hong Kong police arrested seven pro-democracy politicians over tussles with pro-Beijing politicians in the Legislative Council in May. They were charged with contempt and interfering with members of the council, while none of the pro-Beijing lawmakers were detained. Annual commemorations of the 1989 Tiananmen Square protests and massacre were also cancelled amidst fears of violating the national security law. In March 2021, the Chinese central government unilaterally changed Hong Kong's electoral system and established the Candidate Eligibility Review Committee, which would be tasked with screening and evaluating political candidates for their "patriotism". Hong Kong is on China's southern coast, 60 km (37 mi) east of Macau, on the east side of the mouth of the Pearl River estuary. It is surrounded by the South China Sea on all sides except the north, which neighbours the Guangdong city of Shenzhen along the Sham Chun River. The territory's 1,110.18 km (428.64 sq mi) area (2,754.97 km if the maritime area is included) consists of Hong Kong Island, the Kowloon Peninsula, the New Territories, Lantau Island, and over 200 other islands. Of the total area, 1,073 km (414 sq mi) is land and 35 km (14 sq mi) is water. The territory's highest point is Tai Mo Shan, 957 metres (3,140 ft) above sea level. Urban development is concentrated on the Kowloon Peninsula, Hong Kong Island, and in new towns throughout the New Territories. Much of this is built on reclaimed land; 70 km (27 sq mi) (6% of the total land or about 25% of developed space in the territory) is reclaimed from the sea. Undeveloped terrain is hilly to mountainous, with very little flat land, and consists mostly of grassland, woodland, shrubland, or farmland. About 40% of the remaining land area is country parks and nature reserves. The territory has a diverse ecosystem; over 3,000 species of vascular plants occur in the region (300 of which are native to Hong Kong), and thousands of insect, avian, and marine species. Hong Kong has a humid subtropical climate (Köppen Cwa), characteristic of southern China, despite being located south of the Tropic of Cancer. Summers are long, hot and humid, with occasional showers and thunderstorms and warm air from the southwest. The humid nature of Hong Kong exacerbates the warmth of summer. Typhoons occur most often then, sometimes resulting in floods or landslides. Winters are short, mild and usually sunny at the beginning, becoming cloudy towards February. Frequent cold fronts bring strong, cooling winds from the north and occasionally result in chilly weather. Autumn is the sunniest season, whilst spring is generally cloudy. Snowfall has been extremely rare in Hong Kong; the last reported instance was on Tai Mo Shan in 1975. Hong Kong averages 1,709 hours of sunshine per year. Historic temperature extremes at the Hong Kong Observatory are 36.6 °C (97.9 °F) on 22 August 2017 and 0.0 °C (32.0 °F) on 18 January 1893. The highest and lowest recorded temperatures in all of Hong Kong are 39.0 °C (102 °F) at Wetland Park on 22 August 2017, and −6.0 °C (21.2 °F) at Tai Mo Shan on 24 January 2016. Hong Kong has the world's largest number of skyscrapers, with 482 towers taller than 150 metres (490 ft), and the third-largest number of high-rise buildings in the world. The lack of available space restricted development to high-density residential tenements and commercial complexes packed closely together on buildable land. Single-family detached homes are uncommon and generally only found in outlying areas. The International Commerce Centre and Two International Finance Centre are the tallest buildings in Hong Kong and are among the tallest in the Asia-Pacific region. Other distinctive buildings lining the Hong Kong Island skyline include the HSBC Main Building, the anemometer-topped triangular Central Plaza, the circular Hopewell Centre, and the sharp-edged Bank of China Tower. Demand for new construction has contributed to frequent demolition of older buildings, freeing space for modern high-rises. However, many examples of European and Lingnan architecture are still found throughout the territory. Older government buildings are examples of colonial architecture. The 1846 Flagstaff House, the former residence of the commanding British military officer, is the oldest Western-style building in Hong Kong. Some (including the Court of Final Appeal Building and the Hong Kong Observatory) retain their original function, and others have been adapted and reused; the Former Marine Police Headquarters was redeveloped into a commercial and retail complex, and Béthanie (built in 1875 as a sanatorium) houses the Hong Kong Academy for Performing Arts. The Tin Hau Temple, dedicated to the sea goddess Mazu (originally built in 1012 and rebuilt in 1266), is the territory's oldest existing structure. The Ping Shan Heritage Trail has architectural examples of several imperial Chinese dynasties, including the Tsui Sing Lau Pagoda (Hong Kong's only remaining pagoda). Tong lau, mixed-use tenement buildings constructed during the colonial era, blended southern Chinese architectural styles with European influences. These were especially prolific during the immediate post-war period, when many were rapidly built to house large numbers of Chinese migrants. Examples include Lui Seng Chun, the Blue House in Wan Chai, and the Shanghai Street shophouses in Mong Kok. Mass-produced public-housing estates, built since the 1960s, are mainly constructed in modernist style. The Census and Statistics Department estimated Hong Kong's population at 7,413,070 in 2021. The overwhelming majority (91.6%) is Han Chinese, most of whom are Taishanese, Teochew, Hakka, and other Cantonese peoples. The remaining 8.4% are non-ethnic Chinese minorities, primarily Filipinos, Indonesians, and South Asians. However, most Filipinos and Indonesians in Hong Kong are short-term workers. According to a 2021 thematic report by the Hong Kong government, after excluding foreign domestic helpers, the real number of non-Chinese ethnic minorities in the city was 301,344, or 4% of Hong Kong's population. About half the population have some form of British nationality, a legacy of colonial rule; 3.4 million residents have British National (Overseas) status, and 260,000 British citizens live in the territory. The vast majority also hold Chinese nationality, automatically granted to all ethnic Chinese residents at the transfer of sovereignty. Headline population density exceeds 7,060 people/km, and is the fourth-highest in the world. The predominant language is Cantonese, a variety of Chinese originating in Guangdong. It is spoken by 93.7% of the population, 88.2% as a first language and 5.5% as a second language. Slightly over half the population (58.7%) speaks English, the other official language; 4.6% are native speakers, and 54.1% speak English as a second language. Code-switching, mixing English and Cantonese in informal conversation, is common among the bilingual population. Post-handover governments have promoted Mandarin, which is currently about as prevalent as English; 54.2% of the population speak Mandarin, with 2.3% native speakers and 51.9% as a second language. Traditional Chinese characters are used in writing, rather than the simplified characters used in the mainland. Among the religious population, the traditional "three teachings" of China, Buddhism, Confucianism, and Taoism, have the most adherents (20%), followed by Christianity (12%) and Islam (4%). Followers of other religions, including Sikhism, Hinduism, and Judaism, generally originate from regions where their religion predominates. Life expectancy in Hong Kong was 81.3 years for males and 87.2 years for females in 2022, one of the highest in the world. Cancer, pneumonia, heart disease, cerebrovascular disease, and accidents are the territory's five leading causes of death. The universal public healthcare system is funded by general-tax revenue, and treatment is highly subsidised; on average, 95% of healthcare costs are covered by the government. The city has a severe amount of income inequality, which has risen since the transfer of sovereignty, as the region's ageing population has gradually added to the number of nonworking people. Although median household income steadily increased during the decade to 2016, the wage gap remained high; the 90th percentile of earners receive 41% of all income. The city has the most billionaires per capita, with one billionaire per 109,657 people, as well as the second-highest number of billionaires of any city in the world, the highest number of billionaires of any city in Asia, and the largest concentration of ultra high-net-worth individuals of any city in the world. Despite government efforts to reduce the growing disparity, median income for the top 10% of earners is 44 times that of the bottom 10%. One of the world's most significant financial centres and commercial ports, Hong Kong has a market economy focused on services, characterised by low taxation, minimal government market intervention, and an established international financial market. It is the world's 35th-largest economy, with a nominal GDP of approximately US$373 billion. Hong Kong's economy ranked at the top of the Heritage Foundation's economic freedom index between 1995 and 2021. However, Hong Kong was removed from the index by the Heritage Foundation in 2021, with the Foundation citing a "loss of political freedom and autonomy ... [making Hong Kong] almost indistinguishable in many respects from other major Chinese commercial centers like Shanghai and Beijing". Hong Kong is highly developed, and ranks fourth on the UN Human Development Index. The Hong Kong Stock Exchange is the seventh-largest in the world, with a market capitalisation of HK$30.4 trillion (US$3.87 trillion) as of December 2018. Hong Kong is ranked as the 17th most innovative territory in the Global Innovation Index in 2023, and 3rd in the Global Financial Centres Index. The city is sometimes referred to as "Silicon Harbor", a nickname derived from Silicon Valley in California. Hong Kong hosts several high tech and innovation companies, including several multinational companies. Hong Kong is the ninth largest trading entity in exports and eighth largest in imports (2021), trading more goods in value than its gross domestic product. Over half of its cargo throughput consists of transshipments (goods travelling through Hong Kong). Products from mainland China account for about 40% of that traffic. The city's location allowed it to establish a transportation and logistics infrastructure which includes the world's seventh-busiest container port and the busiest airport for international cargo. The territory's largest export markets are mainland China and the United States. Hong Kong is a key part of the 21st Century Maritime Silk Road. It has little arable land and few natural resources, importing most of its food and raw materials. More than 90% of Hong Kong's food is imported, including nearly all of its meat and rice. Agricultural activity is 0.1% of GDP and consists of growing premium food and flower varieties. Although the territory had one of Asia's largest manufacturing economies during the latter half of the colonial era, Hong Kong's economy is now dominated by the service sector. The sector generates 92.7% of economic output, with the public sector accounting for about 10%. Between 1961 and 1997 Hong Kong's gross domestic product increased by a factor of 180, and per capita GDP increased by a factor of 87. The territory's GDP relative to mainland China's peaked at 27% in 1993; it fell to less than 3% in 2017, as the mainland developed and liberalised its economy. Economic and infrastructure integration with China has increased significantly since the 1978 start of market liberalisation on the mainland. Since resumption of cross-boundary train service in 1979, many rail and road links have been improved and constructed, facilitating trade between regions. The Closer Economic Partnership Arrangement formalised a policy of free trade between the two areas, with each jurisdiction pledging to remove remaining obstacles to trade and cross-boundary investment. A similar economic partnership with Macau details the liberalisation of trade between the special administrative regions. Chinese companies have expanded their economic presence in the territory since the transfer of sovereignty. Mainland firms represent over half of the Hang Seng Index value, up from 5% in 1997. As the mainland liberalised its economy, Hong Kong's shipping industry faced intense competition from other Chinese ports. Half of China's trade goods were routed through Hong Kong in 1997, dropping to about 13% by 2015. The territory's minimal taxation, common law system, and civil service attract overseas corporations wishing to establish a presence in Asia. The city has the second-highest number of corporate headquarters in the Asia-Pacific region. Hong Kong is a gateway for foreign direct investment in China, giving investors open access to mainland Chinese markets through direct links with the Shanghai and Shenzhen stock exchanges. The territory was the first market outside mainland China for renminbi-denominated bonds, and is one of the largest hubs for offshore renminbi trading. In November 2020, Hong Kong's Financial Services and the Treasury Bureau proposed a new law that will restrict cryptocurrency trading to professional investors only, leaving amateur traders (93% of Hong Kong's trading population) out of the market. The Hong Kong dollar, the local currency, is the eighth most traded currency in the world. Due to extremely compact house sizes and the extremely high housing density, the city has the most expensive housing market in the world. The government has had a passive role in the economy. Colonial governments had little industrial policy and implemented almost no trade controls. Under the doctrine of "positive non-interventionism", post-war administrations deliberately avoided the direct allocation of resources; active intervention was considered detrimental to economic growth. While the economy transitioned to a service basis during the 1980s, late colonial governments introduced interventionist policies. Post-handover administrations continued and expanded these programmes, including export-credit guarantees, a compulsory pension scheme, a minimum wage, anti-discrimination laws, and a state mortgage backer. Tourism is a major part of the economy, accounting for 5% of GDP. In 2016, 26.6 million visitors contributed HK$258 billion (US$32.9 billion) to the territory, making Hong Kong the 14th most popular destination for international tourists. It is the most popular Chinese city for tourists, receiving over 70% more visitors than its closest competitor (Macau). The city is ranked as one of the most expensive cities for expatriates. However, since 2020, there has been a sharp decline in incoming visitors due to tight COVID-19 travel restrictions. Additionally, due to the closure of Russian airspace in 2022, multiple airlines decided to cease their operations in Hong Kong. In an attempt to attract tourists back to Hong Kong, the Hong Kong government announced plans to give away 500,000 free airline tickets in 2023. Hong Kong has a highly developed, sophisticated transport network. Over 90% of daily trips are made on public transport, the highest percentage in the world. The Octopus card, a contactless smart payment card, is widely accepted on railways, trams, buses and ferries, and can be used for payment in most retail stores. The Peak Tram, Hong Kong's first public transport system, has provided funicular rail transport between Central and Victoria Peak since 1888. The Central and Western District has an extensive system of escalators and moving pavements, including the Mid-Levels escalator (the world's longest outdoor covered escalator system). Hong Kong Tramways covers a portion of Hong Kong Island. The Mass Transit Railway (MTR) is an extensive passenger rail network, connecting 93 metro stations throughout the territory. With a daily ridership of almost five million, the system serves 41% of all public transit passengers in the city and has an on-time rate of 99.9%. Cross-boundary train service to Shenzhen is offered by the East Rail line, and longer-distance inter-city trains to Guangzhou, Shanghai, and Beijing are operated from Hung Hom station. Connecting service to the national high-speed rail system is provided at West Kowloon railway station. Although public transport systems handle most passenger traffic, there are over 500,000 private vehicles registered in Hong Kong. Automobiles drive on the left (unlike in mainland China), because of historical influence of the British Empire. Vehicle traffic is extremely congested in urban areas, exacerbated by limited space to expand roads and an increasing number of vehicles. More than 18,000 taxicabs, easily identifiable by their bright colour, are licensed to carry riders in the territory. Bus services operate more than 700 routes across the territory, with smaller public light buses (also known as minibuses) serving areas standard buses do not reach as frequently or directly. Highways, organised with the Hong Kong Strategic Route and Exit Number System, connect all major areas of the territory. The Hong Kong–Zhuhai–Macau Bridge provides a direct route to the western side of the Pearl River estuary. Hong Kong International Airport is the territory's primary airport. Over 100 airlines operate flights from the airport, including locally based Cathay Pacific (flag carrier), Hong Kong Airlines, low-cost airline HK Express and cargo airline Air Hong Kong. It is the eighth-busiest airport by passenger traffic pre-COVID and handles the most air-cargo traffic in the world. Most private recreational aviation traffic flies through Shek Kong Airfield, under the supervision of the Hong Kong Aviation Club. The Star Ferry operates two lines across Victoria Harbour for its 53,000 daily passengers. Ferries also serve outlying islands inaccessible by other means. Smaller kai-to boats serve the most remote coastal settlements. Ferry travel to Macau and mainland China is also available. Junks, once common in Hong Kong waters, are no longer widely available and are used privately and for tourism. The large size of the port gives Hong Kong the classification of Large-Port Metropolis. Hong Kong generates most of its electricity locally. The vast majority of this energy comes from fossil fuels, with 46% from coal and 47% from petroleum. The rest is from other imports, including nuclear energy generated in mainland China. Renewable sources account for a negligible amount of energy generated for the territory. Small-scale wind-power sources have been developed, and a small number of private homes and public buildings have installed solar panels. With few natural lakes and rivers, high population density, inaccessible groundwater sources, and extremely seasonal rainfall, the territory does not have a reliable source of freshwater. The Dongjiang River in Guangdong supplies 70% of the city's water, and the remaining demand is filled by harvesting rainwater. Toilets in most built-up areas of the territory flush with seawater, greatly reducing freshwater use. Broadband Internet access is widely available, with 92.6% of households connected. Connections over fibre-optic infrastructure are increasingly prevalent, contributing to the high regional average connection speed of 21.9 Mbit/s (the world's fourth-fastest). Mobile-phone use is ubiquitous; there are more than 18 million mobile-phone accounts, more than double the territory's population. Hong Kong is characterised as a hybrid of East and West. Traditional Chinese values emphasising family and education blend with Western ideals, including economic liberty and the rule of law. Although the vast majority of the population is ethnically Chinese, Hong Kong has developed a distinct identity. The territory diverged from the mainland through its long period of colonial administration and a different pace of economic, social, and cultural development. Mainstream culture is derived from immigrants originating from various parts of China. This was influenced by British-style education, a separate political system, and the territory's rapid development during the late 20th century. Most migrants of that era fled poverty and war, reflected in the prevailing attitude toward wealth; Hongkongers tend to link self-image and decision-making to material benefits. Residents' sense of local identity has markedly increased post-handover: The majority of the population (52%) identifies as "Hongkongers", while 11% describe themselves as "Chinese". The remaining population purport mixed identities, 23% as "Hongkonger in China" and 12% as "Chinese in Hong Kong". Traditional Chinese family values, including family honour, filial piety, and a preference for sons, are prevalent. Nuclear families are the most common households, although multi-generational and extended families are not unusual. Spiritual concepts such as feng shui are observed; large-scale construction projects often hire consultants to ensure proper building positioning and layout. The degree of its adherence to feng shui is believed to determine the success of a business. Bagua mirrors are regularly used to deflect evil spirits, and buildings often lack floor numbers with a 4; the number has a similar sound to the word for "die" in Cantonese. Food in Hong Kong is primarily based on Cantonese cuisine, despite the territory's exposure to foreign influences and its residents' varied origins. Rice is the staple food, and is usually served plain with other dishes. Freshness of ingredients is emphasised. Poultry and seafood are commonly sold live at wet markets, and ingredients are used as quickly as possible. There are five daily meals: breakfast, lunch, afternoon tea, dinner, and siu yeh. Dim sum, as part of yum cha (brunch), is a dining-out tradition with family and friends. Dishes include congee, cha siu bao, siu yuk, egg tarts, and mango pudding. Local versions of Western food are served at cha chaan teng (Hong Kong-style cafes). Common cha chaan teng menu items include macaroni in soup, deep-fried French toast, and Hong Kong-style milk tea. Hong Kong developed into a filmmaking hub during the late 1940s as a wave of Shanghai filmmakers migrated to the territory, and these movie veterans helped build the colony's entertainment industry over the next decade. By the 1960s, the city was well known to overseas audiences through films such as The World of Suzie Wong. When Bruce Lee's The Way of the Dragon was released in 1972, local productions became popular outside Hong Kong. During the 1980s, films such as A Better Tomorrow, As Tears Go By, and Zu Warriors from the Magic Mountain expanded global interest beyond martial arts films; locally made gangster films, romantic dramas, and supernatural fantasies became popular. Hong Kong cinema continued to be internationally successful over the following decade with critically acclaimed dramas such as Farewell My Concubine, To Live, and Chungking Express. The city's martial arts film roots are evident in the roles of the most prolific Hong Kong actors. Jackie Chan, Donnie Yen, Jet Li, Chow Yun-fat, and Michelle Yeoh frequently play action-oriented roles in foreign films. Hong Kong films have also grown popular in oversea markets such as Japan, South Korea, and Southeast Asia, earning the city the moniker "Hollywood of the East". At the height of the local movie industry in the early 1990s, over 400 films were produced each year; since then, industry momentum shifted to mainland China. The number of films produced annually has declined to about 60 in 2017. Cantopop is a genre of Cantonese popular music which emerged in Hong Kong during the 1970s. Evolving from Shanghai-style shidaiqu, it is also influenced by Cantonese opera and Western pop. Local media featured songs by artists such as Sam Hui, Anita Mui, Leslie Cheung, and Alan Tam; during the 1980s, exported films and shows exposed Cantopop to a global audience. The genre's popularity peaked in the 1990s, when the Four Heavenly Kings dominated Asian record charts. Despite a general decline since late in the decade, Cantopop remains dominant in Hong Kong; contemporary artists such as Eason Chan, Joey Yung, and Twins are popular in and beyond the territory. Western classical music has historically had a strong presence in Hong Kong and remains a large part of local musical education. The publicly funded Hong Kong Philharmonic Orchestra, the territory's oldest professional symphony orchestra, frequently hosts musicians and conductors from overseas. The Hong Kong Chinese Orchestra, composed of classical Chinese instruments, is the leading Chinese ensemble and plays a significant role in promoting traditional music in the community. Hong Kong has never had a separate national anthem to the country that controlled it; its current official national anthem is therefore that of China, March of the Volunteers. The song Glory to Hong Kong has been used by protestors as an unofficial anthem of the city. Despite its small area, the territory is home to a variety of sports and recreational facilities. The city has hosted numerous major sporting events, including the 2009 East Asian Games, the 2008 Summer Olympics equestrian events, and the 2007 Premier League Asia Trophy. The territory regularly hosts the Hong Kong Sevens, Hong Kong Marathon, Hong Kong Tennis Classic and Lunar New Year Cup, and hosted the inaugural AFC Asian Cup and the 1995 Dynasty Cup. Hong Kong represents itself separately from mainland China, with its own sports teams in international competitions. The territory has participated in almost every Summer Olympics since 1952 and has earned nine medals. Lee Lai-shan won the territory's first Olympic gold medal at the 1996 Atlanta Olympics, and Cheung Ka Long won the second one in Tokyo 2020. Hong Kong athletes have won 126 medals at the Paralympic Games and 17 at the Commonwealth Games. No longer part of the Commonwealth of Nations, the city's last appearance in the latter was in 1994. Dragon boat races originated as a religious ceremony conducted during the annual Tuen Ng Festival. The race was revived as a modern sport as part of the Tourism Board's efforts to promote Hong Kong's image abroad. The first modern competition was organised in 1976, and overseas teams began competing in the first international race in 1993. The Hong Kong Jockey Club, the territory's largest taxpayer, has a monopoly on gambling and provides over 7% of government revenue. Three forms of gambling are legal in Hong Kong: lotteries, horse racing, and football. Education in Hong Kong is largely modelled on that of the United Kingdom, particularly the English system. Children are required to attend school from age 6 until completion of secondary education, generally at age 18. At the end of secondary schooling, all students take a public examination and are awarded the Hong Kong Diploma of Secondary Education upon successful completion. Of residents aged 15 and older, 81% completed lower-secondary education, 66% graduated from an upper secondary school, 32% attended a non-degree tertiary program, and 24% earned a bachelor's degree or higher. Mandatory education has contributed to an adult literacy rate of 95.7%. The literacy rate is lower than that of other developed economies because of the influx of refugees from mainland China during the post-war colonial era; much of the elderly population were not formally educated because of war and poverty. Comprehensive schools fall under three categories: public schools, which are government-run; subsidised schools, including government aid-and-grant schools; and private schools, often those run by religious organisations and that base admissions on academic merit. These schools are subject to the curriculum guidelines as provided by the Education Bureau. Private schools subsidised under the Direct Subsidy Scheme; international schools fall outside of this system and may elect to use differing curricula and teach using other languages. At primary and secondary school levels, the government maintains a policy of "mother tongue instruction"; most schools use Cantonese as the medium of instruction, with written education in both Chinese and English. Other languages being used as medium of instruction in non-international school education include English and Putonghua (Standard Mandarin Chinese). Secondary schools emphasise "bi-literacy and tri-lingualism", which has encouraged the proliferation of spoken Mandarin language education. English is the official medium of instruction and assessments for most university programmes in Hong Kong, although use of Cantonese is predominant in informal discussions among local students and professors. Hong Kong has eleven universities. The University of Hong Kong (HKU) was founded as the city's first institute of higher education during the early colonial period in 1911. The Chinese University of Hong Kong (CUHK) was established in 1963 to fill the need for a university that taught using Chinese as its primary language of instruction. Along with the Hong Kong University of Science and Technology (HKUST) established in 1991, these universities are consistently ranked among the top 50 or top 100 universities worldwide. The Hong Kong Polytechnic University (PolyU) and City University of Hong Kong (CityU), both granted university status in 1994, are consistently ranked among the top 100 or top 200 universities worldwide. The Hong Kong Baptist University (HKBU) was granted university status in 1994 and is a liberal arts institution. Lingnan University, Education University of Hong Kong, Hong Kong Metropolitan University (formerly Open University of Hong Kong), Hong Kong Shue Yan University and Hang Seng University of Hong Kong all attained full university status in subsequent years. Most of the newsapapers in Hong Kong are written in Chinese but there are also a few English-language newspapers. The major one is the South China Morning Post, with The Standard serving as a business-oriented alternative. A variety of Chinese-language newspapers are published daily; the most prominent are Ming Pao and Oriental Daily News. Local publications are often politically affiliated, with pro-Beijing or pro-democracy sympathies. The central government has a print-media presence in the territory through the state-owned Ta Kung Pao and Wen Wei Po. Several international publications have regional operations in Hong Kong, including The Wall Street Journal, Financial Times, The New York Times International Edition, USA Today, Yomiuri Shimbun, and The Nikkei. Three free-to-air television broadcasters operate in the territory; TVB, HKTVE, and Hong Kong Open TV air eight digital channels. TVB, Hong Kong's dominant television network, has an 80% viewer share. Pay TV services operated by Cable TV Hong Kong and PCCW offer hundreds of additional channels and cater to a variety of audiences. RTHK is the public broadcaster, providing seven radio channels and three television channels. Ten non-domestic broadcasters air programming for the territory's foreign population. Access to media and information over the Internet is not subject to mainland Chinese regulations, including the Great Firewall, yet local control applies. Government Trade Maps 22°18′N 114°12′E / 22.3°N 114.2°E / 22.3; 114.2
[[File:4th_Panzer_Division_logo_2.svg|thumb|upright|The 4th Panzer Division used ''blitzkrieg'' successfully.]] {{War}} '''Blitzkrieg''' is a [[German language|German]] word that means "lightning war" and refers to the high speed of a [[lightning bolt]]. In ''blitzkrieg'', the attacking motorized infantry armies move quickly and are helped by [[tank|tanks]] and [[aircraft]]. Slower-moving enemy units are overrun or surrounded and often captured with little fighting. The slower units often become disorganized and are not ready to fight when they are captured. The combined arms tactics of ''blitzkrieg'' were developed in the 1920s and 1930s, especially in the German [[Wehrmacht]]. They seldom used this word, however. This method worked well early in [[World War II]] during the invasions of [[Poland]] and [[France]] and was mostly successful during [[Operation Barbarossa]]. Later in the war, the Allies learned to defeat German ''blitzkrieg'' attacks by [[defence in depth]] and by attacking the flanks of the attackers with reserve forces. '''The Blitz''' refers to the German bombing of [[Britain]], particularly [[London]], during World War II. It destroyed over a million homes, killed over 40,000 people and was supposed to destroy industry and morale quickly to make the British pressure their government to end the war. The Blitz started in September 1940, in response to the British bombing of German cities by the [[Royal Air Force]] and continued until May 1941. [[Category:Military of Germany]] [[Category:World War II]] {{stub}}
Blitzkrieg (/ˈblɪtskriːɡ/ BLITS-kreeg, German: [ˈblɪtskʁiːk] ; from Blitz 'lightning' + Krieg 'war') or Bewegungskrieg is a word used to describe a combined arms surprise attack using a rapid, overwhelming force concentration that may consist of armored and motorized or mechanized infantry formations, together with artillery, air assault and close air support, with intent to break through the opponent's lines of defense, dislocate the defenders, unbalance the enemies by making it difficult to respond to the continuously changing front, and defeat them in a decisive Vernichtungsschlacht: a battle of annihilation. During the interwar period, aircraft and tank technologies matured and were combined with systematic application of the traditional German tactic of Bewegungskrieg (maneuver warfare), deep penetrations and the bypassing of enemy strong points to encircle and destroy enemy forces in a Kesselschlacht (cauldron battle/battle of encirclement). During the invasion of Poland, Western journalists adopted the term blitzkrieg to describe that form of armored warfare. The term had appeared in 1935, in a German military periodical Deutsche Wehr ("German Defence"), in connection to quick or lightning warfare. German maneuver operations were successful in the campaigns of 1939–1941, and by 1940, the term blitzkrieg was extensively used in Western media. Blitzkrieg operations capitalised on surprise penetrations such as that of the Ardennes forest region, the general Allies' unreadiness and their inability to match the pace of the German attack. During the Battle of France, the French made attempts to reform defensive lines along rivers but were frustrated when German forces arrived first and pressed on. Despite being common in German and English-language journalism during World War II, the word Blitzkrieg was never used by the Wehrmacht as an official military term except for propaganda. According to David Reynolds, "Hitler himself called the term Blitzkrieg 'A completely idiotic word' (ein ganz blödsinniges Wort)". Some senior officers, including Kurt Student, Franz Halder and Johann Adolf von Kielmansegg, even disputed the idea that it was a military concept. Kielmansegg asserted that what many regarded as blitzkrieg was nothing more than "ad hoc solutions that simply popped out of the prevailing situation". Student described it as ideas that "naturally emerged from the existing circumstances" as a response to operational challenges. The Wehrmacht never officially adopted it as a concept or doctrine. In 2005, the historian Karl-Heinz Frieser summarized blitzkrieg as the result of German commanders using the latest technology in the most advantageous way according to traditional military principles and employing "the right units in the right place at the right time". Modern historians now understand blitzkrieg as the combination of the traditional German military principles, methods and doctrines of the 19th century with the military technology of the interwar period. Modern historians use the term casually as a generic description for the style of maneuver warfare practiced by Germany during the early part of World War II, rather than as an explanation. According to Frieser, in the context of the thinking of Heinz Guderian on mobile combined arms formations, blitzkrieg can be used as a synonym for modern maneuver warfare on the operational level. The traditional meaning of "blitzkrieg" is that of German tactical and operational methodology during the first half of the Second World War that is often hailed as a new method of warfare. The word, meaning "lightning war" or "lightning attack" in its strategic sense describes a series of quick and decisive short battles to deliver a knockout blow to an enemy state before it can fully mobilize. Tactically, blitzkrieg is a coordinated military effort by tanks, motorized infantry, artillery and aircraft, to create an overwhelming local superiority in combat power, to defeat the opponent and break through its defences. Blitzkrieg as used by Germany had considerable psychological or "terror" elements, such as the Jericho Trompete, a noise-making siren on the Junkers Ju 87 dive bomber, to affect the morale of enemy forces. The devices were largely removed when the enemy became used to the noise after the Battle of France in 1940, and instead, bombs sometimes had whistles attached. It is also common for historians and writers to include psychological warfare by using fifth columnists to spread rumours and lies among the civilian population in the theatre of operations. The origin of the term blitzkrieg is obscure. It was never used in the title of a military doctrine or handbook of the German Army or Air Force, and no "coherent doctrine" or "unifying concept of blitzkrieg" existed, however German High Command mostly referred to the group of tactics as "Bewegungskrieg" (Maneuver Warfare). The term seems to have been rarely used in the German military press before 1939, and recent research at the German Militärgeschichtliches Forschungsamt, at Potsdam, found it in only two military articles from the 1930s. Both used the term to mean a swift strategic knockout, rather than a radically new military doctrine or approach to war. The first article (1935) dealt primarily with supplies of food and materiel in wartime. The term blitzkrieg was used in reference to German efforts to win a quick victory in the First World War but was not associated with the use of armored, mechanized or air forces. It argued that Germany must develop self-sufficiency in food because it might again prove impossible to deal a swift knockout to its enemies, which would lead to a long war. In the second article (1938), launching a swift strategic knockout was described as an attractive idea for Germany but difficult to achieve on land under modern conditions (especially against systems of fortification like the Maginot Line) unless an exceptionally high degree of surprise could be achieved. The author vaguely suggested that a massive strategic air attack might hold out better prospects, but the topic was not explored in detail. A third relatively early use of the term in German occurred in Die Deutsche Kriegsstärke (German War Strength) by Fritz Sternberg, a Jewish Marxist political economist and refugee from Nazi Germany, published in 1938 in Paris and in London as Germany and a Lightning War. Sternberg wrote that Germany was not prepared economically for a long war but might win a quick war ("Blitzkrieg"). He did not go into detail about tactics or suggest that the German armed forces had evolved a radically new operational method. His book offered scant clues as to how German lightning victories might be won. In English and other languages, the term had been used since the 1920s. The term was first used in the publications of Ferdinand Otto Miksche, first in the magazine "Army Quarterly", and in his 1941 book Blitzkrieg, in which he defined the concept. In September 1939, Time magazine termed the German military action as a "war of quick penetration and obliteration – Blitzkrieg, lightning war". After the invasion of Poland, the British press commonly used the term to describe German successes in that campaign. J. P. Harris called the term "a piece of journalistic sensationalism – a buzz-word with which to label the spectacular early successes of the Germans in the Second World War". The word was later applied to the bombing of Britain, particularly London, hence "The Blitz". The German popular press followed suit nine months later, after the Fall of France in 1940; thus, although the word had first been used in Germany, it was popularized by British journalism. Heinz Guderian referred to it as a word coined by the Allies: "as a result of the successes of our rapid campaigns our enemies ... coined the word Blitzkrieg". After the German failure in the Soviet Union in 1941, the use of the term began to be frowned upon in Nazi Germany, and Hitler then denied ever using the term and said in a speech in November 1941, "I have never used the word Blitzkrieg, because it is a very silly word". In early January 1942, Hitler dismissed it as "Italian phraseology". In 1914, German strategic thinking derived from the writings of Carl von Clausewitz (1 June 1780 – 16 November 1831), Helmuth von Moltke the Elder (26 October 1800 – 24 April 1891) and Alfred von Schlieffen (28 February 1833 – 4 January 1913), who advocated maneuver, mass and envelopment to create the conditions for a decisive battle (Vernichtungsschlacht). During the war, officers such as Willy Rohr developed tactics to restore maneuver on the battlefield. Specialist light infantry (Stosstruppen, "storm troops") were to exploit weak spots to make gaps for larger infantry units to advance with heavier weapons, exploit the success and leave isolated strong points to the troops that were following up. Infiltration tactics were combined with short hurricane artillery bombardments, which used massed artillery. Devised by Colonel Georg Bruchmüller, the attacks relied on speed and surprise, rather than on weight of numbers. The tactics met with great success in Operation Michael, the German spring offensive of 1918 and restored temporarily the war of movement once the Allied trench system had been overrun. The German armies pushed on towards Amiens and then Paris and came within 120 kilometres (75 mi) before supply deficiencies and Allied reinforcements halted the advance. The historian James Corum criticised the German leadership for failing to understand the technical advances of the First World War, conducting no studies of the machine gun prior to the war and giving tank production the lowest priority during the war. After Germany's defeat, the Treaty of Versailles limited the Reichswehr to a maximum of 100,000 men, which prevented the deployment of mass armies. The German General Staff was abolished by the treaty but continued covertly as the Truppenamt (Troop Office) and was disguised as an administrative body. Committees of veteran staff officers were formed within the Truppenamt to evaluate 57 issues of the war to revise German operational theories. By the time of the Second World War, their reports had led to doctrinal and training publications, including H. Dv. 487, Führung und Gefecht der verbundenen Waffen ("Command and Battle of the Combined Arms)", known as Das Fug (1921–1923) and Truppenführung (1933–1934), containing standard procedures for combined-arms warfare. The Reichswehr was influenced by its analysis of pre-war German military thought, particularly infiltration tactics since at the end of the war, they had seen some breakthroughs on the Western Front and the maneuver warfare which dominated the Eastern Front. On the Eastern Front, the war did not bog down into trench warfare since the German and the Russian Armies fought a war of maneuver over thousands of miles, which gave the German leadership unique experience that was unavailable to the trench-bound Western Allies. Studies of operations in the East led to the conclusion that small and coordinated forces possessed more combat power than large uncoordinated forces. After the war, the Reichswehr expanded and improved infiltration tactics. The commander in chief, Hans von Seeckt, argued that there had been an excessive focus on encirclement and emphasised speed instead. Seeckt inspired a revision of Bewegungskrieg (maneuver warfare) thinking and its associated Auftragstaktik in which the commander expressed his goals to subordinates and gave them discretion in how to achieve them. The governing principle was "the higher the authority, the more general the orders were"; it was the responsibility of the lower echelons to fill in the details. Implementation of higher orders remained within limits that were determined by the training doctrine of an elite officer corps. Delegation of authority to local commanders increased the tempo of operations, which had great influence on the success of German armies in the early war period. Seeckt, who believed in the Prussian tradition of mobility, developed the German army into a mobile force and advocated technical advances that would lead to a qualitative improvement of its forces and better coordination between motorized infantry, tanks, and planes. The British Army took lessons from the successful infantry and artillery offensives on the Western Front in late 1918. To obtain the best co-operation between all arms, emphasis was placed on detailed planning, rigid control and adherence to orders. Mechanization of the army, as part of a combined-arms theory of war, was considered a means to avoid mass casualties and the indecisive nature of offensives. The four editions of Field Service Regulations that were published after 1918 held that only combined-arms operations could create enough fire power to enable mobility on a battlefield. That theory of war also emphasised consolidation and recommended caution against overconfidence and ruthless exploitation. During the Sinai and Palestine campaign, operations involved some aspects of what would later be called blitzkrieg. The decisive Battle of Megiddo included concentration, surprise and speed. Success depended on attacking only in terrain favouring the movement of large formations around the battlefield and tactical improvements in the British artillery and infantry attack. General Edmund Allenby used infantry to attack the strong Ottoman front line in co-operation with supporting artillery, augmented by the guns of two destroyers. Through constant pressure by infantry and cavalry, two Ottoman armies in the Judean Hills were kept off-balance and virtually encircled during the Battles of Sharon and Nablus (Battle of Megiddo). The British methods induced "strategic paralysis" among the Ottomans and led to their rapid and complete collapse. In an advance of 65 miles (105 km), captures were estimated to be "at least 25,000 prisoners and 260 guns". Liddell Hart considered that important aspects of the operation had been the extent to which Ottoman commanders were denied intelligence on the British preparations for the attack through British air superiority and air attacks on their headquarters and telephone exchanges, which paralyzed attempts to react to the rapidly-deteriorating situation. Norman Stone detects early blitzkrieg operations in offensives by French Generals Charles Mangin and Marie-Eugène Debeney in 1918. However, French doctrine in the interwar years became defence-oriented. Colonel Charles de Gaulle advocated concentration of armor and airplanes. His opinions appeared in his 1934 book Vers l'Armée de métier ("Towards the Professional Army"). Like von Seeckt, de Gaulle concluded that France could no longer maintain the huge armies of conscripts and reservists that had fought the First World War, and he sought to use tanks, mechanized forces and aircraft to allow a smaller number of highly trained soldiers to have greater impact in battle. His views endeared him little to the French high command but are claimed by some to have influenced Heinz Guderian. In 1916, General Alexei Brusilov had used surprise and infiltration tactics during the Brusilov Offensive. Later, Marshal Mikhail Tukhachevsky (1893-1937), Georgii Isserson [ru] (1898-1976) and other members of the Red Army developed a concept of deep battle from the experience of the Polish–Soviet War of 1919–1920. Those concepts would guide the Red Army doctrine throughout the Second World War. Realising the limitations of infantry and cavalry, Tukhachevsky advocated mechanized formations and the large-scale industrialisation that they required. Robert Watt (2008) wrote that blitzkrieg has little in common with Soviet deep battle. In 2002, H. P. Willmott had noted that deep battle contained two important differences from blitzkrieg by being a doctrine of total war, not of limited operations, and rejecting decisive battle in favour of several large simultaneous offensives. The Reichswehr and the Red Army began a secret collaboration in the Soviet Union to evade the Treaty of Versailles occupational agent, the Inter-Allied Commission. In 1926 war games and tests began at Kazan and Lipetsk, in the Soviet Russia. The centers served to field-test aircraft and armored vehicles up to the battalion level and housed aerial- and armoured-warfare schools through which officers rotated. After becoming Chancellor of Germany in 1933, Adolf Hitler ignored the provisions of the Treaty of Versailles Treaty. Within the Wehrmacht, which was established in 1935, the command for motorized armored forces was named the Panzerwaffe in 1936. The Luftwaffe, the German air force, was officially established in February 1935, and development began on ground-attack aircraft and doctrines. Hitler strongly supported the new strategy. He read Guderian's 1937 book Achtung – Panzer! and upon observing armored field exercises at Kummersdorf, he remarked, "That is what I want – and that is what I will have". Guderian summarized combined-arms tactics as the way to get the mobile and motorized armored divisions to work together and support each other to achieve decisive success. In his 1950 book, Panzer Leader, he wrote: In this year, 1929, I became convinced that tanks working on their own or in conjunction with infantry could never achieve decisive importance. My historical studies, the exercises carried out in England and our own experience with mock-ups had persuaded me that the tanks would never be able to produce their full effect until the other weapons on whose support they must inevitably rely were brought up to their standard of speed and of cross-country performance. In such formation of all arms, the tanks must play primary role, the other weapons being subordinated to the requirements of the armor. It would be wrong to include tanks in infantry divisions; what was needed were armored divisions which would include all the supporting arms needed to allow the tanks to fight with full effect. Guderian believed that developments in technology were required to support the theory, especially by equipping armored divisions, tanks foremost, with wireless communications. Guderian insisted in 1933 to the high command that every tank in the German armored force must be equipped with a radio. At the start of World War II, only the German Army was thus prepared with all tanks being "radio-equipped". That proved critical in early tank battles in which German tank commanders exploited the organizational advantage over the Allies that radio communication gave them. All Allied armies would later copy that innovation. During the Polish campaign, the performance of armored troops, under the influence of Guderian's ideas, won over a number of skeptics who had initially expressed doubt about armored warfare, such as von Rundstedt and Rommel. According to David A. Grossman, by the Twelfth Battle of Isonzo (October–November 1917), while he was conducting a light-infantry operation, Rommel had perfected his maneuver-warfare principles, which were the very same ones that were applied during the blitzkrieg against France in 1940 and were repeated in the Coalition ground offensive against Iraq in the 1991 Gulf War. During the Battle of France and against his staff advisor's advice, Hitler ordered that everything should be completed in a few weeks. Fortunately for the Germans, Rommel and Guderian disobeyed the General Staff's orders (particularly those of General Paul Ludwig Ewald von Kleist) and forged ahead making quicker progress than anyone had expected, on the way "inventing the idea of Blitzkrieg". It was Rommel who created the new archetype of Blitzkrieg by leading his division far ahead of flanking divisions. MacGregor and Williamson remark that Rommel's version of blitzkrieg displayed a significantly better understanding of combined-arms warfare than that of Guderian. General Hermann Hoth submitted an official report in July 1940 which declared that Rommel had "explored new paths in the command of Panzer divisions". Schwerpunktprinzip was a heuristic device (conceptual tool or thinking formula) that was used in the German Army since the nineteenth century to make decisions from tactics to strategy about priority. Schwerpunkt has been translated as center of gravity, crucial, focal point and point of main effort. None of those forms is sufficient to describe the universal importance of the term and the concept of Schwerpunktprinzip. Every unit in the army, from the company to the supreme command, decided on a Schwerpunkt by schwerpunktbildung, as did the support services, which meant that commanders always knew what was the most important and why. The German army was trained to support the Schwerpunkt even when risks had to be taken elsewhere to support the point of main effort and to attack with overwhelming firepower. Schwerpunktbildung allowed the German Army to achieve superiority at the Schwerpunkt, whether attacking or defending, to turn local success at the Schwerpunkt into the progressive disorganisation of the opposing force and to create more opportunities to exploit that advantage even if the Germans were numerically and strategically inferior in general. In the 1930s, Guderian summarized that as Klotzen, nicht kleckern! ("Kick, don't spatter them!") Having achieved a breakthrough of the enemy's line, units comprising the Schwerpunkt were not supposed to become decisively engaged with enemy front line units to the right and the left of the breakthrough area. Units pouring through the hole were to drive upon set objectives behind the enemy front line. During the Second World War, German Panzer forces used their motorized mobility to paralyze the opponent's ability to react. Fast-moving mobile forces seized the initiative, exploited weaknesses and acted before the opposing forces could respond. Central to that was the decision cycle (tempo). Through superior mobility and faster decision-making cycles, mobile forces could act faster than the forces opposing them. Directive control was a fast and flexible method of command. Rather than receiving an explicit order, a commander would be told of his superior's intent and the role that his unit was to fill in that concept. The method of execution was then a matter for the discretion of the subordinate commander. The staff burden was reduced at the top and spread among tiers of command with knowledge about their situation. Delegation and the encouragement of initiative aided implementation, and important decisions could be taken quickly and communicated verbally or with only brief written orders. The last part of an offensive operation was the destruction of unsubdued pockets of resistance, which had been enveloped earlier and bypassed by the fast-moving armored and motorized spearheads. The Kesselschlacht ("cauldron battle") was a concentric attack on such pockets. It was there that most losses were inflicted upon the enemy, primarily through the mass capture of prisoners and weapons. During Operation Barbarossa, huge encirclements in 1941 produced nearly 3.5 million Soviet prisoners, along with masses of equipment. Close air support was provided in the form of the dive bomber and medium bomber, which would support the focal point of attack from the air. German successes are closely related to the extent to which the German Luftwaffe could control the air war in early campaigns in Western and Central Europe and in the Soviet Union. However, the Luftwaffe was a broadly based force with no constricting central doctrine other than its resources should be used generally to support national strategy. It was flexible and could carry out both operational-tactical, and strategic bombing. Flexibility was the strength of the Luftwaffe in 1939 to 1941. Paradoxically, that later became its weakness. While Allied Air Forces were tied to the support of the Army, the Luftwaffe deployed its resources in a more general operational way. It switched from air superiority missions to medium-range interdiction, to strategic strikes to close support duties, depending on the need of the ground forces. In fact, far from it being a specialist panzer spearhead arm, less than 15 percent of the Luftwaffe was intended for close support of the army in 1939. Methamphetamine, known as "pervitin," use is believed to have played a role in the speed of Germany's initial Blitzkrieg since military success with combined arms demanded long hours of continuous operations with minimal rest. The concepts associated with the term blitzkrieg (deep penetrations by armor, large encirclements, and combined arms attacks) were largely dependent upon terrain and weather conditions. Wherever the ability for rapid movement across "tank country" was not possible, armored penetrations often were avoided or resulted in failure. The terrain would ideally be flat, firm, unobstructed by natural barriers or fortifications, and interspersed with roads and railways. If it were instead hilly, wooded, marshy, or urban, armor would be vulnerable to infantry in close-quarters combat and unable to break out at full speed. Additionally, units could be halted by mud (thawing along the Eastern Front regularly slowed both sides) or extreme snow. Operation Barbarossa helped confirm that armor effectiveness and the requisite aerial support depended on weather and terrain. It should, however, be noted that the disadvantages of terrain could be nullified if surprise was achieved over the enemy by an attack in areas that had been considered natural obstacles, as occurred during the Battle of France in which the German blitzkrieg-style attack went through the Ardennes. Since the French thought that the Ardennes unsuitable for massive troop movement, particularly for tanks, the area was left with only light defences, which were quickly overrun by the Wehrmacht. The Germans quickly advanced through the forest and knocked down the trees that the French had thought would impede that tactic. The influence of air forces over forces on the ground changed significantly over the course of the Second World War. Early German successes were conducted when Allied aircraft could not make a significant impact on the battlefield. In May 1940, there was near parity in numbers of aircraft between the Luftwaffe and the Allies, but the Luftwaffe had been developed to support Germany's ground forces, had liaison officers with the mobile formations and operated a higher number of sorties per aircraft. In addition, the Germans' air parity or superiority allowed the unencumbered movement of ground forces, their unhindered assembly into concentrated attack formations, aerial reconnaissance, aerial resupply of fast moving formations and close air support at the point of attack. The Allied air forces had no close air support aircraft, training or doctrine. The Allies flew 434 French and 160 British sorties a day but methods of attacking ground targets had yet to be developed and so Allied aircraft caused negligible damage. Against the Allies' 600 sorties, the Luftwaffe on average flew 1,500 sorties a day. On 13 May, Fliegerkorps VIII flew 1,000 sorties in support of the crossing of the Meuse. The following day the Allies made repeated attempts to destroy the German pontoon bridges, but German fighter aircraft, ground fire and Luftwaffe flak batteries with the panzer forces destroyed 56 percent of the attacking Allied aircraft, and the bridges remained intact. Allied air superiority became a significant hindrance to German operations during the later years of the war. By June 1944, the Western Allies had the complete control of the air over the battlefield, and their fighter-bomber aircraft were very effective at attacking ground forces. On D-Day, the Allies flew 14,500 sorties over the battlefield area alone, not including sorties flown over Northwestern Europe. Against them the Luftwaffe flew some 300 sorties on 6 June. Though German fighter presence over Normandy increased over the next days and weeks, it never approached the numbers that the Allies commanded. Fighter-bomber attacks on German formations made movement during daylight almost impossible. Subsequently, shortages soon developed in food, fuel and ammunition and severely hampered the German defenders. German vehicle crews and even flak units experienced great difficulty moving during daylight. Indeed, the final German offensive operation in the west, Operation Wacht am Rhein, was planned to take place during poor weather to minimise interference by Allied aircraft. Under those conditions, it was difficult for German commanders to employ the "armored idea", if at all. Blitzkrieg is vulnerable to an enemy that is robust enough to weather the shock of the attack and does not panic at the idea of enemy formations in its rear area. That is especially true if the attacking formation lacks the reserve to keep funnelling forces into the spearhead or the mobility to provide infantry, artillery and supplies into the attack. If the defender can hold the shoulders of the breach, it has the opportunity to counter-attack into the flank of the attacker and potentially to cut it off the van, as what happened to Kampfgruppe Peiper in the Ardennes. During the Battle of France in 1940, the 4th Armoured Division (Major-General Charles de Gaulle) and elements of the 1st Army Tank Brigade (British Expeditionary Force) made probing attacks on the German flank and pushed into the rear of the advancing armored columns at times. That may have been a reason for Hitler to call a halt to the German advance. Those attacks combined with Maxime Weygand's hedgehog tactic would become the major basis for responding to blitzkrieg attacks in the future. Deployment in depth, or permitting enemy or "shoulders" of a penetration, was essential to channelling the enemy attack; artillery, properly employed at the shoulders, could take a heavy toll of attackers. Allied forces in 1940 lacked the experience to develop those strategies successfully, which resulted in the French armistice with heavy losses, those strategies characterized later Allied operations. At the Battle of Kursk, the Red Army used a combination of defence in great depth, extensive minefields and tenacious defense of breakthrough shoulders. In that way, they depleted German combat power even as German forces advanced. The reverse can be seen in the Russian summer offensive of 1944, Operation Bagration, which resulted in the destruction of Army Group Center. German attempts to weather the storm and fight out of encirclements failed because of the Soviets' ability to continue to feed armored units into the attack, maintain the mobility and strength of the offensive and arrive in force deep in the rear areas faster than the Germans could regroup. Although effective in quick campaigns against Poland and France, mobile operations could not be sustained by Germany in later years. Strategies based on maneuver have the inherent danger of the attacking force overextending its supply lines and can be defeated by a determined foe who is willing and able to sacrifice territory for time in which to regroup and rearm, as the Soviets did on the Eastern Front, as opposed to, for example, the Dutch, who had no territory to sacrifice. Tank and vehicle production was a constant problem for Germany. Indeed, late in the war, many panzer "divisions" had no more than a few dozen tanks. As the end of the war approached, Germany also experienced critical shortages in fuel and ammunition stocks as a result of Anglo-American strategic bombing and blockade. Although the production of Luftwaffe fighter aircraft continued, they could not fly because of lack of fuel. What fuel there was went to panzer divisions, and even then, they could not operate normally. Of the Tiger tanks lost against the US Army, nearly half of them were abandoned for lack of fuel. German volunteers first used armor in live field-conditions during the Spanish Civil War (1936-1939). Armor commitment consisted of Panzer Battalion 88, a force built around three companies of Panzer I tanks that functioned as a training cadre for Spain's Nationalists. The Luftwaffe deployed squadrons of fighters, dive-bombers and transport aircraft as the Condor Legion. Guderian said that the tank deployment was "on too small a scale to allow accurate assessments to be made". (The true test of his "armored idea" would have to wait for the Second World War.) However, the Luftwaffe also provided volunteers to Spain to test both tactics and aircraft in combat, including the first combat use of the Stuka. During the war, the Condor Legion undertook the 1937 bombing of Guernica, which had a tremendous psychological effect on the populations of Europe. The results were exaggerated, and the Western Allies concluded that the "city-busting" techniques were now part of the German way in war. The targets of the German aircraft were actually the rail lines and bridges, but lacking the ability to hit them with accuracy (only three or four Ju 87s saw action in Spain), the Luftwaffe chose a method of carpet bombing, resulting in heavy civilian casualties. Although journalists popularized the term Blitzkrieg during the September 1939 invasion of Poland, the historians Matthew Cooper and J. P. Harris have written that German operations during the campaign were consistent with traditional methods. The Wehrmacht strategy was more in line with Vernichtungsgedanke, a focus on envelopment to create pockets in broad-front annihilation. The German generals dispersed Panzer forces among the three German concentrations with little emphasis on independent use. They deployed tanks to create or destroy close pockets of Polish forces and to seize operational-depth terrain in support of the largely-unmotorized infantry, which followed. The Wehrmacht used available models of tanks, Stuka dive-bombers and concentrated forces in the Polish campaign, but the majority of the fighting involved conventional infantry and artillery warfare, and most Luftwaffe action was independent of the ground campaign. Matthew Cooper wrote: Throughout the Polish Campaign, the employment of the mechanised units revealed the idea that they were intended solely to ease the advance and to support the activities of the infantry.... Thus, any strategic exploitation of the armoured idea was still-born. The paralysis of command and the breakdown of morale were not made the ultimate aim of the... German ground and air forces, and were only incidental by-products of the traditional maneuvers of rapid encirclement and of the supporting activities of the flying artillery of the Luftwaffe, both of which had as their purpose the physical destruction of the enemy troops. Such was the Vernichtungsgedanke of the Polish campaign. John Ellis wrote that "there is considerable justice in Matthew Cooper's assertion that the panzer divisions were not given the kind of strategic mission that was to characterize authentic armored blitzkrieg, and were almost always closely subordinated to the various mass infantry armies". Steven Zaloga wrote, "Whilst Western accounts of the September campaign have stressed the shock value of the panzer and Stuka attacks, they have tended to underestimate the punishing effect of German artillery on Polish units. Mobile and available in significant quantity, artillery shattered as many units as any other branch of the Wehrmacht." The German invasion of France, with subsidiary attacks on Belgium and the Netherlands, consisted of two phases, Operation Yellow (Fall Gelb) and Operation Red (Fall Rot). Yellow opened with a feint conducted against the Netherlands and Belgium by two armored corps and paratroopers. Most of the German armored forces were placed in Panzer Group Kleist, which attacked through the Ardennes, a lightly defended sector that the French planned to reinforce if necessary before the Germans could bring up heavy and siege artillery. There was no time for the French to send such reinforcement, as the Germans did not wait for siege artillery but reached the Meuse and achieved a breakthrough at the Battle of Sedan in three days. Panzer Group Kleist raced to the English Channel, reached the coast at Abbeville and cut off the BEF, the Belgian Army and some of the best-equipped divisions of the French Army in northern France. Armored and motorized units under Guderian, Rommel and others advanced far beyond the marching and horse-drawn infantry divisions and far in excess of what Hitler and the German high command had expected or wished. When the Allies counter-attacked at Arras by using the heavily armored British Matilda I and Matilda II tanks, a brief panic ensued in the German High Command. Hitler halted his armored and motorized forces outside the port of Dunkirk, which the Royal Navy had started using to evacuate the Allied forces. Hermann Göring promised that the Luftwaffe would complete the destruction of the encircled armies, but aerial operations failed to prevent the evacuation of the majority of the Allied troops. In Operation Dynamo, some 330,000 French and British troops escaped. Case Yellow surprised everyone by overcoming the Allies' 4,000 armored vehicles, many of which were better than their German equivalents in armor and gunpower. The French and British frequently used their tanks in the dispersed role of infantry support, rather than by concentrating force at the point of attack, to create overwhelming firepower. The French armies were much reduced in strength and the confidence of their commanders shaken. With much of their own armor and heavy equipment lost in Northern France, they lacked the means to fight a mobile war. The Germans followed their initial success with Operation Red, a triple-pronged offensive. The XV Panzer Corps attacked towards Brest, XIV Panzer Corps attacked east of Paris, towards Lyon and the XIX Panzer Corps encircled the Maginot Line. The French, hard pressed to organise any sort of counter-attack, were continually ordered to form new defensive lines and found that German forces had already bypassed them and moved on. An armored counter-attack, organized by Colonel Charles de Gaulle, could not be sustained, and he had to retreat. Prior to the German offensive in May, Winston Churchill had said, "Thank God for the French Army". The same French Army collapsed after barely two months of fighting. That was in shocking contrast to the four years of trench warfare on which French forces had engaged during the First World War. French Prime Minister Paul Reynaud, analyzed the collapse in a speech on 21 May 1940: The truth is that our classic conception of the conduct of war has come up against a new conception. At the basis of this... there is not only the massive use of heavy armoured divisions or cooperation between them and airplanes, but the creation of disorder in the enemy's rear by means of parachute raids. The Germans had not used paratroopry attacks in France and made only one large drop in the Netherlands to capture three bridges; some small glider-landings were conducted in Belgium to take bottlenecks on routes of advance before the arrival of the main force (the most renowned being the landing on Fort Eben-Emael in Belgium). Use of armored forces was crucial for both sides on the Eastern Front. Operation Barbarossa, the German invasion of the Soviet Union in June 1941, involved a number of breakthroughs and encirclements by motorized forces. Its goal, according to Führer Directive 21 (18 December 1940), was "to destroy the Russian forces deployed in the West and to prevent their escape into the wide-open spaces of Russia". The Red Army was to be destroyed west of the Dvina and Dnieper rivers, which were about 500 kilometres (310 mi) east of the Soviet border, to be followed by a mopping-up operation. The surprise attack resulted in the near annihilation of the Voyenno-Vozdushnye Sily (VVS, Soviet Air Force) by simultaneous attacks on airfields, allowing the Luftwaffe to achieve total air supremacy over all the battlefields within the first week. On the ground, four German panzer groups outflanked and encircled disorganized Red Army units, and the marching infantry completed the encirclements and defeated the trapped forces. In late July, after 2nd Panzer Group (commanded by Guderian) captured the watersheds of the Dvina and Dnieper rivers near Smolensk, the panzers had to defend the encirclement, because the marching infantry divisions remained hundreds of kilometers to the west. The Germans conquered large areas of the Soviet Union, but their failure to destroy the Red Army before the winter of 1941-1942 was a strategic failure and made German tactical superiority and territorial gains irrelevant. The Red Army had survived enormous losses and regrouped with new formations far to the rear of the front line. During the Battle of Moscow (October 1941 to January 1942), the Red Army defeated the German Army Group Center and for the first time in the war seized the strategic initiative. In the summer of 1942, Germany launched another offensive and this time focusing on Stalingrad and the Caucasus in the southern Soviet Union. The Soviets again lost tremendous amounts of territory, only to counter-attack once more during winter. The German gains were ultimately limited because Hitler diverted forces from the attack on Stalingrad and drove towards the Caucasus oilfields simultaneously. The Wehrmacht became overstretched. Although it won operationally, it could not inflict a decisive defeat as the durability of the Soviet Union's manpower, resources, industrial base and aid from the Western Allies began to take effect. In July 1943, the Wehrmacht conducted Operation Zitadelle (Citadel) against a salient at Kursk, which Soviet troop heavily defended. Soviet defensive tactics had by now hugely improved, particularly in the use of artillery and air support. By April 1943, the Stavka had learned of German intentions through intelligence supplied by front-line reconnaissance and Ultra intercepts. In the following months, the Red Army constructed deep defensive belts along the paths of the planned German attack. The Soviets made a concerted effort to disguise their knowledge of German plans and the extent of their own defensive preparations, and the German commanders still hoped to achieve operational surprise when the attack commenced. The Germans did not achieve surprise and could not outflank or break through into enemy rear areas during the operation. Several historians assert that Operation Citadel was planned and intended to be a blitzkrieg operation. Many of the German participants who wrote about the operation after the war, including Erich von Manstein, make no mention of blitzkrieg in their accounts. In 2000, Niklas Zetterling and Anders Frankson characterised only the southern pincer of the German offensive as a "classical blitzkrieg attack". Pier Battistelli wrote that the operational planning marked a change in German offensive thinking away from blitzkrieg and that more priority was given to brute force and fire power than to speed and maneuver. In 1995, David Glantz stated that blitzkrieg was at Kursk for the first time defeated in summer, and the opposing Soviet forces mounted a successful counter-offensive. The Battle of Kursk ended with two Soviet counter-offensives and the revival of deep operations. In the summer of 1944, the Red Army destroyed Army Group Centre in Operation Bagration by using combined-arms tactics for armor, infantry and air power in a coordinated strategic assault, known as deep operations, which led to an advance of 600 kilometres (370 mi) in six weeks. Allied armies began using combined-arms formations and deep-penetration strategies that Germany had used in the opening years of the war. Many Allied operations in the Western Desert and on the Eastern Front, relied on firepower to establish breakthroughs by fast-moving armored units. The artillery-based tactics were also decisive in Western Front operations after 1944's Operation Overlord, and the British Commonwealth and American armies developed flexible and powerful systems for using artillery support. What the Soviets lacked in flexibility, they made up for in number of rocket launchers, guns and mortars. The Germans never achieved the kind of fire concentrations that their enemies were achieving 1944. After the Allied landings in Normandy (June 1944), the Germans began a counter-offensive to overwhelm the landing force with armored attacks, but they failed because of a lack of co-ordination and to Allied superiority in anti-tank defense and in the air. The most notable attempt to use deep-penetration operations in Normandy was Operation Luttich at Mortain, which only hastened the Falaise Pocket and the destruction of German forces in Normandy. The Mortain counter-attack was defeated by the American 12th Army Group with little effect on its own offensive operations. The last German offensive on the Western front, the Battle of the Bulge (Operation Wacht am Rhein), was an offensive launched towards the port of Antwerp in December 1944. Launched in poor weather against a thinly-held Allied sector, it achieved surprise and initial success as Allied air-power was grounded due to cloud cover. Determined defense by American troops in places throughout the Ardennes, the lack of good roads and German supply shortages caused delays. Allied forces deployed to the flanks of the German penetration, and as soon as the skies cleared, Allied aircraft returned to the battlefield. Allied counter-attacks soon forced back the Germans, who abandoned much equipment for lack of fuel. Blitzkrieg had been called a Revolution in Military Affairs (RMA), but many writers and historians have concluded that the Germans did not invent a new form of warfare but applied new technologies to traditional ideas of Bewegungskrieg (maneuver warfare) to achieve decisive victory. In 1965, Captain Robert O'Neill, Professor of the History of War at the University of Oxford produced an example of the popular view. In Doctrine and Training in the German Army 1919–1939, O'Neill wrote: What makes this story worth telling is the development of one idea: the blitzkrieg. The German Army had a greater grasp of the effects of technology on the battlefield, and went on to develop a new form of warfare by which its rivals when it came to the test were hopelessly outclassed. Other historians wrote that blitzkrieg was an operational doctrine of the German armed forces and a strategic concept on which the leadership of Nazi Germany based its strategic and economic planning. Military planners and bureaucrats in the war economy appear rarely, if ever, to have employed the term blitzkrieg in official documents. That the German army had a "blitzkrieg doctrine" was rejected in the late 1970s by Matthew Cooper. The concept of a blitzkrieg Luftwaffe was challenged by Richard Overy in the late 1970s and by Williamson Murray in the mid-1980s. That Nazi Germany went to war on the basis of "blitzkrieg economics" was criticized by Richard Overy in the 1980s, and George Raudzens described the contradictory senses in which historians have used the word. The notion of a German blitzkrieg concept or doctrine survives in popular history and many historians still support the thesis. Frieser wrote that after the failure of the Schlieffen Plan in 1914, the German army concluded that decisive battles were no longer possible in the changed conditions of the twentieth century. Frieser wrote that the Oberkommando der Wehrmacht (OKW), which was created in 1938 had intended to avoid the decisive battle concepts of its predecessors and planned for a long war of exhaustion (Ermattungskrieg). It was only after the improvised plan for the Battle of France in 1940 was unexpectedly successful that the German General Staff came to believe that Vernichtungskrieg was still feasible. German thinking reverted to the possibility of a quick and decisive war for the Balkan campaign and Operation Barbarossa. Most academic historians regard the notion of blitzkrieg as military doctrine to be a myth. Shimon Naveh wrote, "The striking feature of the blitzkrieg concept is the complete absence of a coherent theory which should have served as the general cognitive basis for the actual conduct of operations". Naveh described it as an "ad hoc solution" to operational dangers, thrown together at the last moment. Overy disagreed with the idea that Hitler and the Nazi regime ever intended a blitzkrieg war because the once-popular belief that the Nazi state organized its economy to carry out its grand strategy in short campaigns was false. Hitler had intended for a rapid unlimited war to occur much later than 1939, but Germany's aggressive foreign policy forced the state into war before it was ready. The planning of Hitler and the Wehrmacht in the 1930s did not reflect a blitzkrieg method but the opposite. J. P. Harris wrote that the Wehrmacht never used the word, and it did not appear in German army or air force field manuals. The word was coined in September 1939 by a Times newspaper reporter. Harris also found no evidence that German military thinking developed a blitzkrieg mentality. Karl-Heinz Frieser and Adam Tooze reached similar conclusions to Overy and Naveh that the notions of blitzkrieg economy and strategy are myths. Frieser wrote that surviving German economists and General Staff officers denied that Germany went to war with a blitzkrieg strategy. Robert M. Citino argues: Blitzkrieg was not a doctrine, or an operational scheme, or even a tactical system. In fact, it simply doesn't exist, at least not in the way we usually think it does. The Germans never used the term Blitzkrieg in any precise sense, and almost never used it outside of quotations. It simply meant a rapid and decisive victory (lightning war)... The Germans didn't invent anything new in the interwar period, but rather used new technologies like tanks and air and radio-controlled command to restore an old way of war that they still found to be valid, Bewegungskrieg. The historian Victor Davis Hanson states that Blitzkrieg "played on the myth of German technological superiority and industrial dominance" and adds that German successes, particularly that of its Panzer divisions were "instead predicated on the poor preparation and morale of Germany's enemies". Hanson also reports that at a Munich public address in November 1941, Hitler had "disowned" the concept of Blitzkrieg by calling it an "idiotic word". Further, successful Blitzkrieg operations were predicated on superior numbers, air support and were possible for only short periods of time without sufficient supply lines. For all intents and purposes, Blitzkrieg ended at the Eastern Front once the German forces had given up Stalingrad, after they faced hundreds of new T-34 tanks, when the Luftwaffe became unable to assure air dominance, and after the stalemate at Kursk. To that end, Hanson concludes that German military success was not accompanied by the adequate provisioning of its troops with food and materiel far from the source of supply, which contributed to its ultimate failures. Despite its later disappointments as German troops extended their lines at too great a distance, the very specter of armored Blitzkrieg forces initially proved victorious against the Polish, Dutch, Belgian and French Armies early in the war. In the 1960s, Alan Milward developed a theory of blitzkrieg economics: Germany could not fight a long war and chose to avoid comprehensive rearmament and armed in breadth to win quick victories. Milward described an economy positioned between a full war economy and a peacetime economy. The purpose of the blitzkrieg economy was to allow the German people to enjoy high living standards in the event of hostilities and avoid the economic hardships of the First World War. Overy wrote that blitzkrieg as a "coherent military and economic concept has proven a difficult strategy to defend in light of the evidence". Milward's theory was contrary to Hitler's and German planners' intentions. The Germans, aware of the errors of the First World War, rejected the concept of organizing its economy to fight only a short war. Therefore, focus was given to the development of armament in depth for a long war, instead of armament in breadth for a short war. Hitler claimed that relying on surprise alone was "criminal" and that "we have to prepare for a long war along with surprise attack". During the winter of 1939–1940, Hitler demobilized many troops from the army to return as skilled workers to factories because the war would be decided by production, not a quick "Panzer operation". In the 1930s, Hitler had ordered rearmament programs that cannot be considered limited. In November 1937, he had indicated that most of the armament projects would be completed by 1943–1945. The rearmament of the Kriegsmarine was to have been completed in 1949 and the Luftwaffe rearmament program was to have matured in 1942, with a force capable of strategic bombing with heavy bombers. The construction and the training of motorized forces and a full mobilization of the rail networks would not begin until 1943 and 1944, respectively. Hitler needed to avoid war until these projects were complete but his misjudgements in 1939 forced Germany into war before rearmament was complete. After the war, Albert Speer claimed that the German economy achieved greater armaments output not because of diversions of capacity from civilian to military industry but by streamlining of the economy. Overy pointed out some 23 percent of German output was military by 1939. Between 1937 and 1939, 70 percent of investment capital went into the rubber, synthetic fuel, aircraft and shipbuilding industries. Hermann Göring had consistently stated that the task of the Four Year Plan was to rearm Germany for total war. Hitler's correspondence with his economists also reveals that his intent was to wage war in 1943–1945, when the resources of central Europe had been absorbed into Nazi Germany. Living standards were not high in the late 1930s. Consumption of consumer goods had fallen from 71 percent in 1928 to 59 percent in 1938. The demands of the war economy reduced the amount of spending in non-military sectors to satisfy the demand for the armed forces. On 9 September, Göring, as Head of the Reich Defense Council, called for complete "employment" of living and fighting power of the national economy for the duration of the war. Overy presents that as evidence that a "blitzkrieg economy" did not exist. Adam Tooze wrote that the German economy was being prepared for a long war. The expenditure for the war was extensive and put the economy under severe strain. The German leadership were concerned less with how to balance the civilian economy and the needs of civilian consumption but to figure out how to best prepare the economy for total war. Once war had begun, Hitler urged his economic experts to abandon caution and expend all available resources on the war effort, but the expansion plans only gradually gained momentum in 1941. Tooze wrote that the huge armament plans in the pre-war period did not indicate any clear-sighted blitzkrieg economy or strategy. Frieser wrote that the Heer (German pronunciation: [ˈheːɐ̯]) was not ready for blitzkrieg at the start of the war. A blitzkrieg method called for a young, highly skilled mechanized army. In 1939–1940, 45 percent of the army was 40 years old and 50 percent of the soldiers had only a few weeks' training. The German Army, contrary to the blitzkrieg legend, was not fully motorized and had only 120,000 vehicles, compared to the 300,000 of the French Army. The British also had an "enviable" contingent of motorized forces. Thus, "the image of the German 'Blitzkrieg' army is a figment of propaganda imagination". During the First World War, the German army used 1.4 million horses for transport and in the Second World War 2.7 million horses. Only ten percent of the army was motorized in 1940. Half of the German divisions available in 1940 were combat ready, but they were less well-equipped than the British and French or the Imperial German Army of 1914. In the spring of 1940, the German army was semi-modern in which a small number of well-equipped and "elite" divisions were offset by many second and third rate divisions". In 2003, John Mosier wrote that while the French soldiers in 1940 were better trained than German soldiers, as were the Americans later and that the German Army was the least mechanized of the major armies, its leadership cadres were larger and better and that the high standard of leadership was the main reason for the successes of the German army in World War II, as it had been in World War I. James Corum wrote that it was a myth that the Luftwaffe had a doctrine of terror bombing in which civilians were attacked to break the will or aid the collapse of an enemy by the Luftwaffe in blitzkrieg operations. After the bombing of Guernica in 1937 and the Rotterdam Blitz in 1940, it was commonly assumed that terror bombing was a part of Luftwaffe doctrine. During the interwar period, the Luftwaffe leadership rejected the concept of terror bombing in favour of battlefield support and interdiction operations: The vital industries and transportation centers that would be targeted for shutdown were valid military targets. Civilians were not to be targeted directly, but the breakdown of production would affect their morale and will to fight. German legal scholars of the 1930s carefully worked out guidelines for what type of bombing was permissible under international law. While direct attacks against civilians were ruled out as "terror bombing", the concept of the attacking the vital war industries – and probable heavy civilian casualties and breakdown of civilian morale – was ruled as acceptable. Corum continued: General Walther Wever compiled a doctrine known as The Conduct of the Aerial War. This document, which the Luftwaffe adopted, rejected Giulio Douhet's theory of terror bombing. Terror bombing was deemed to be "counter-productive", increasing rather than destroying the enemy's will to resist. Such bombing campaigns were regarded as diversion from the Luftwaffe's main operations; destruction of the enemy armed forces. The bombings of Guernica, Rotterdam and Warsaw were tactical missions in support of military operations and were not intended as strategic terror attacks. J. P. Harris wrote that most Luftwaffe leaders from Goering through the general staff believed, as did their counterparts in Britain and the United States, that strategic bombing was the chief mission of the air force and that given such a role, the Luftwaffe would win the next war and that Nearly all lectures concerned the strategic uses of airpower; virtually none discussed tactical co-operation with the Army. Similarly in the military journals, emphasis centred on 'strategic' bombing. The prestigious Militärwissenschaftliche Rundschau, the War Ministry's journal, which was founded in 1936, published a number of theoretical pieces on future developments in air warfare. Nearly all discussed the use of strategic airpower, some emphasising that aspect of air warfare to the exclusion of others. One author commented that European military powers were increasingly making the bomber force the heart of their airpower. The manoeuvrability and technical capability of the next generation of bombers would be 'as unstoppable as the flight of a shell. The Luftwaffe ended up with an air force consisting mainly of relatively short-range aircraft, but that does not prove that the German air force was solely interested in "tactical" bombing. It happened because the German aircraft industry lacked the experience to build a long-range bomber fleet quickly and because Hitler was insistent on the very rapid creation of a numerically large force. It is also significant that Germany's position in the centre of Europe to a large extent obviated the need to make a clear distinction between bombers suitable only for "tactical" purposes and those necessary for strategic purposes in the early stages of a likely future war. The British theorists John Frederick Charles Fuller and Captain Basil Henry Liddell Hart have often been associated with the development of blitzkrieg, but that is a matter of controversy. In recent years historians have uncovered that Liddell Hart distorted and falsified facts to make it appear as if his ideas has been adopted. After the war Liddell Hart imposed his own perceptions after the event by claiming that the mobile tank warfare has been practiced by the Wehrmacht was a result of his influence. By manipulation and contrivance, Liddell Hart distorted the actual circumstances of the blitzkrieg formation, and he obscured its origins. By his indoctrinated idealization of an ostentatious concept, he reinforced the myth of blitzkrieg. Imposing retrospectively his own perceptions of mobile warfare upon the shallow concept of blitzkrieg, he "created a theoretical imbroglio that has taken 40 years to unravel". Blitzkrieg was not an official doctrine, and historians in recent times have come to the conclusion that it did not exist as such: It was the opposite of a doctrine. Blitzkrieg consisted of an avalanche of actions that were sorted out less by design and more by success. In hindsight—and with some help from Liddell Hart—this torrent of action was squeezed into something it never was: an operational design. The early 1950s literature transformed blitzkrieg into a historical military doctrine, which carried the signature of Liddell Hart and Guderian. The main evidence of Liddell Hart's deceit and "tendentious" report of history can be found in his letters to Erich von Manstein, Heinz Guderian, and the relatives and associates of Erwin Rommel. Liddell Hart, in letters to Guderian, "imposed his own fabricated version of blitzkrieg on the latter and compelled him to proclaim it as original formula". Kenneth Macksey found Liddell Hart's original letters to Guderian in the latter's papers. Liddell Hart requested Guderian to give him credit for "impressing him" with his ideas of armored warfare. When Liddell Hart was questioned about this in 1968 and the discrepancy between the English and German editions of Guderian's memoirs, "he gave a conveniently unhelpful though strictly truthful reply. ('There is nothing about the matter in my file of correspondence with Guderian himself except... that I thanked him... for what he said in that additional paragraph'.)". During the First World War, Fuller had been a staff officer attached to the new tank corps. He developed Plan 1919 for massive independent tank operations, which he claimed were subsequently studied by the German military. It is variously argued that Fuller's wartime plans and post-war writings were inspirations or that his readership was low and German experiences during the war received more attention. The German view of themselves as the losers of the war may be linked to the senior and experienced officers' undertaking a thorough review in studying and rewriting of all of their Army doctrine and training manuals. Fuller and Liddell Hart were "outsiders". Liddell Hart was unable to serve as a soldier after 1916 after being gassed on the Somme, and Fuller's abrasive personality resulted in his premature retirement in 1933. Their views had limited impact in the British army; the War Office permitted the formation of an Experimental Mechanized Force on 1 May 1927, composed of tanks, motorized infantry, self-propelled artillery and motorized engineers but the force was disbanded in 1928 on the grounds that it had served its purpose. A new experimental brigade was intended for the next year and became a permanent formation in 1933, during the cuts of the 1932/33–1934/35 financial years. It has been argued that blitzkrieg was not and thae that the Germans did not invent something called blitzkrieg in the 1920s and 1930s. Rather, the German concept of wars of movement and concentrated force were seen in wars of Prussia and the German Wars of Unification. The first European general to introduce rapid movement, concentrated power and integrated military effort was Swedish King Gustavus Adolphus during the Thirty Years' War. The appearance of the aircraft and tank in the First World War, called an RMA, offered the German military a chance to get back to the traditional war of movement as practiced by Moltke the Elder. The so-called "blitzkrieg campaigns" of 1939 to around 1942 were well within that operational context. At the outbreak of war, the German army had no radically new theory of war. The operational thinking of the German army had not changed significantly since the First World War or since the late 19th century. J. P. Harris and Robert M. Citino point out that the Germans had always had a marked preference for short decisive campaigns but were unable to achieve short-order victories in First World War conditions. The transformation from the stalemate of the First World War into tremendous initial operational and strategic success in the Second World War was partly the employment of a relatively-small number of mechanized divisions, most importantly the Panzer divisions, and the support of an exceptionally powerful air force. Heinz Guderian is widely regarded as being highly influential in developing the military methods of warfare used by Germany's tank men at the start of the Second World War. That style of warfare brought the maneuver back to the fore and placed an emphasis on the offensive. Along with the shockingly-rapid collapse in the armies that opposed it, that came to be branded as blitzkrieg warfare. Aftee Germany's military reforms of the Guderian emerged as a strong proponent of mechanized forces. Within the Inspectorate of Transport Troops, Guderian and colleagues performed theoretical and field exercise work. Guderian met with opposition from some in the General Staff, who were distrustful of the new weapons and who continued to view the infantry as the primary weapon of the army. Among them, Guderian claimed, was Chief of the General Staff Ludwig Beck (1935–1938), who he alleged was skeptical that armored forces could be decisive. That claim has been disputed by later historians. James Corum wrote: Guderian expressed a hearty contempt for General Ludwig Beck, chief of the General Staff from 1935 to 1938, whom he characterized as hostile to ideas of modern mechanised warfare: [Corum quoting Guderian] "He [Beck] was a paralysing element wherever he appeared.... [S]ignificantly of his way of thought was his much-boosted method of fighting which he called delaying defence". This is a crude caricature of a highly competent general who authored Army Regulation 300 (Troop Leadership) in 1933, the primary tactical manual of the German Army in World War II, and under whose direction the first three panzer divisions were created in 1935, the largest such force in the world of the time. By Guderian's account, he single-handedly created the German tactical and operational methodology. Between 1922 and 1928 Guderian wrote a number of articles concerning military movement. As the ideas of making use of the combustible engine in a protected encasement to bring mobility back to warfare developed in the German army, Guderian was a leading proponent of the formations that would be used for this purpose. He was later asked to write an explanatory book, which was titled Achtung Panzer! (1937) in which he explained the theories of the tank men and defended them. Guderian argued that the tank would be the decisive weapon of the next war. "If the tanks succeed, then victory follows", he wrote. In an article addressed to critics of tank warfare, he wrote that "until our critics can produce some new and better method of making a successful land attack other than self-massacre, we shall continue to maintain our beliefs that tanks—properly employed, needless to say—are today the best means available for land attack". Addressing the faster rate at which defenders could reinforce an area than attackers could penetrate it during the First World War, Guderian wrote that "since reserve forces will now be motorized, the building up of new defensive fronts is easier than it used to be; the chances of an offensive based on the timetable of artillery and infantry co-operation are, as a result, even slighter today than they were in the last war." He continued, "We believe that by attacking with tanks we can achieve a higher rate of movement than has been hitherto obtainable, and—what is perhaps even more important—that we can keep moving once a breakthrough has been made". Guderian additionally required for tactical radios to be widely used to facilitate coordination and command by having one installed in all tanks. Guderian's leadership was supported, fostered and institutionalized by his supporters in the Reichswehr General Staff system, which worked the Army to greater and greater levels of capability through massive and systematic Movement Warfare war games in the 1930s. Guderian's book incorporated the work of theorists such as Ludwig Ritter von Eimannsberger, whose book, The Tank War (Der Kampfwagenkrieg) (1934) gained a wide audience in the German Army. Another German theorist, Ernst Volckheim, wrote a huge amount on tank and combined arms tactics and was influential to German thinking on the use of armored formations, but his work was not acknowledged in Guderian's writings.
[[File:BALTICO.jpg|thumb|T-34 used in [[World War II]]]] A '''tank''' is an armored fighting vehicle, typically armed with a lethal [[cannon|gun]] mounted on a turret and a few machine guns. A tank is covered in thick [[armour]] to protect it from enemy weapons. Tanks have tracks that wrap around its wheels to spread out its weight and let it cross rough terrain. Most tanks have a powerful gun and one or more [[machine gun]]s. A tank normally has 3-5 [[crew]] members with at least a driver, commander and gunner. There may also be a loader, who handles the [[ammunition]] for the main gun (so the gunner does not have to take eyes off the target). Some [[World War II|WWII]] tanks also had a separate soldier responsible for the [[radio]]. Since then, tank crews have evolved to have each person do one of these things, but in their early stages, the commander was often the loader and radio operator, and in a few tanks was also the driver. == Using == [[File:Frot Laffly 28 March 1915.jpg|thumb|Early French tank, 1915]] [[File:Leopard 2 A7.JPG|alt=|thumb|A [[Leopard 2]]A7 tank in Germany.]] [[File:Imperial German Heavy Tank A7V Adalbert.jpg|left|thumb|The [[A7V]] tank was the German response to the British and French tanks, and was used in WW1]] The first tanks were made by the British [[Royal Navy]] and French car manufacturers during [[World War I]] as a way of attacking enemy [[trench]]es. They were called '''tanks''' to trick the Germans into thinking they were water carriers for the [[Middle Eastern theatre of World War I]]. Their use in a surprise attack in the [[Battle of the Somme]] caused fear among the German soldiers but their small numbers and poor reliability prevented them from making much difference. Tanks became a main weapon during [[World War II]], when the German Army introduced a way of using them called [[Blitzkrieg]]. Battles between great numbers of tanks were fought, especially between the Germans and the Soviet Union. The Battle of [[Kursk]] was the biggest. Well-known World War II tanks were the German [[Panzer]] IV, Panther and [[Tiger I]], Soviet T-34 (made in the largest numbers of any tank of the war and second most ever), the British Matilda, Churchill and Cromwell, and the American [[M4 sherman]] (second most produced tank of the war) and Stuart tanks. There were plans for much larger and more heavily armoured tanks such as the [[Maus]] but they were determined to be of little use. Their huge weight would have made moving them very difficult, and their huge size would have made them easy targets for heavy [[artillery]]. Old tanks are often modified for other uses like carrying soldiers or equipment. [[Military engineering|Combat engineers]] use special tank-based vehicles, for example minesweeper tanks or bridge-layer tanks. The tank is a major part of all large armies today. Tanks have replaced the [[cavalry]] everywhere and do the things that soldiers on horseback did in the past. Most modern tanks are of the heavy or [[Main Battle Tank]] type, able to fight other tanks. Compared to older types MBTs are very heavy (Russian MBTS being about 40 tons and Western around 60 tons) with a 120&nbsp;mm (Western) and 125&nbsp;mm (Russian) calibre cannon and 2-3 machine guns. [[Marine (military)|Marines]] and other specialized forces also use some light tanks. == Anti-tank warfare == Armies have invented many ways to fight against enemy tanks and to defend their own tanks. Artillery was found to be effective when tanks were first introduced, and special [[anti-tank gun]]<nowiki/>s were made. Elephant guns, large versions of ordinary rifles, were tried. They worked against the earliest tanks but later tanks have tougher armor. Obstacles such as tank traps and anti-tank trenches can catch tanks so artillery can hit them. Anti-tank [[land mine]]<nowiki/>s can also catch or even destroy them. The main weapons infantry uses for fighting against tanks are [[Anti-tank guided missile|anti tank missiles]] and [[Rocket-propelled grenade|rocket propelled grenades]]. Often a tank force has its own infantry troops for protection against those weapons. Tanks are much used to fight tanks, and some tanks are specially made to be good at destroying other tanks. Heavy artillery and [[Bomb|bombs]] can destroy tanks with a blast wave. Starting in the early 21st century, combat drones have sometimes destroyed tanks. Ground attack aircraft can destroy them with guns, bombs, and guided missiles. == Related pages == * [[Anti-tank gun]] * [[Military vehicle#Combat vehicles]] == Other websites == {{commons-inline}} [[Category:Tanks| ]] [[Category:Army]]
A tank is an armoured fighting vehicle intended as a primary offensive weapon in front-line ground combat. Tank designs are a balance of heavy firepower, strong armour, and good battlefield mobility provided by tracks and a powerful engine; usually their main armament is mounted in a turret. They are a mainstay of modern 20th and 21st century ground forces and a key part of combined arms combat. Modern tanks are versatile mobile land weapons platforms whose main armament is a large-caliber tank gun mounted in a rotating gun turret, supplemented by machine guns or other ranged weapons such as anti-tank guided missiles or rocket launchers. They have heavy vehicle armour which provides protection for the crew, the vehicle's munition storage, fuel tank and propulsion systems. The use of tracks rather than wheels provides improved operational mobility which allows the tank to overcome rugged terrain and adverse conditions such as mud and ice/snow better than wheeled vehicles, and thus be more flexibly positioned at advantageous locations on the battlefield. These features enable the tank to perform well in a variety of intense combat situations, simultaneously both offensively (with direct fire from their powerful main gun) and defensively (as fire support and defilade for friendly troops due to the near invulnerability to common infantry small arms and good resistance against heavier weapons, although anti-tank weapons used in 2022, some of them man-portable, have demonstrated the ability to destroy older generations of tanks with single shots), all while maintaining the mobility needed to exploit changing tactical situations. Fully integrating tanks into modern military forces spawned a new era of combat: armoured warfare. Until the arrival of the main battle tank, tanks were typically categorized either by weight class (light, medium, heavy or superheavy tanks) or doctrinal purpose (breakthrough-, cavalry-, infantry-, cruiser-, or reconnaissance tanks). Some are larger and very heavily armoured and with large guns, while others are smaller, lightly armoured, and equipped with a smaller caliber and lighter gun. These smaller tanks move over terrain with speed and agility and can perform a reconnaissance role in addition to engaging enemy targets. The smaller, faster tank would not normally engage in battle with a larger, heavily armoured tank, except during a surprise flanking manoeuvre. The word tank was first applied to British "landships" in 1915 to keep their nature secret before they entered service. On 24 December 1915, a meeting took place of the Inter-Departmental Conference (including representatives of the Director of Naval Construction's Committee, the Admiralty, the Ministry of Munitions, and the War Office). Its purpose was to discuss the progress of the plans for what were described as "Caterpillar Machine Gun Destroyers or Land Cruisers." In his autobiography, Albert Gerald Stern (Secretary to the Landship Committee, later head of the Mechanical Warfare Supply Department) says that at that meeting: Mr. (Thomas J.) Macnamara (M.P., and Parliamentary and Financial Secretary to the Admiralty) then suggested, for secrecy's sake, to change the title of the Landship Committee. Mr. d'Eyncourt agreed that it was very desirable to retain secrecy by all means, and proposed to refer to the vessel as a "Water Carrier". In Government offices, committees and departments are always known by their initials. For this reason I, as Secretary, considered the proposed title totally unsuitable. In our search for a synonymous term, we changed the word "Water Carrier" to "Tank," and became the "Tank Supply" or "T.S." Committee. That is how these weapons came to be called Tanks. He incorrectly added, "and the name has now been adopted by all countries in the world." Lieutenant-Colonel Ernest Swinton, who was secretary to the meeting, says that he was instructed to find a non-committal word when writing his report of the proceedings. In the evening he discussed it with a fellow officer, Lt-Col Walter Dally Jones, and they chose the word "tank". "That night, in the draft report of the conference, the word 'tank' was employed in its new sense for the first time." Swinton's Notes on the Employment of Tanks, in which he uses the word throughout, was published in January 1916. In July 1918, Popular Science Monthly reported: Because a fellow of the Royal Historical Society* has unintentionally misled the British public as to the origin of the famous "tanks", Sir William Tritton, who designed and built them, has published the real story of their name ... Since it was obviously inadvisable to herald "Little Willie's" reason for existence to the world he was known as the "Instructional Demonstration Unit." "Little Willie's" hull was called in the shop orders a "water carrier for Mesopotamia"; no one knew that the hull was intended to be mounted on a truck. Naturally, the water carrier began to be called a "tank". So the name came to be used by managers and foremen of the shop, until now it has a place in the army vocabulary and will probably be so known in history for all time. (*F.J. Gardiner, F.R.Hist.S.) D'Eyncourt's account differs from Swinton's and Tritton's: ... when the future arrangements were under discussion for transporting the first landships to France a question arose as to how, from a security point of view, the consignment should be labelled. To justify their size we decided to call them 'water-carriers for Russia' —the idea being that they should be taken for some new method of taking water to forward troops in the battle areas. Lt.-Col. Swinton ... raised a humorous objection to this, remarking that the War Office pundits would probably contract the description to 'W.C.'s for Russia', and that we had better forestall this by merely labelling the packages 'Tanks'. So tanks they became, and tanks they have remained." This appears to be an imperfect recollection. He says that the name problem arose "when we shipped the first two vehicles to France the following year" (August 1916), but by that time the name "tank" had been in use for eight months. The tanks were labelled "With Care to Petrograd," but the belief was encouraged that they were a type of snowplough. The term "tank" is used throughout the English speaking world, but other countries use different terminology. In France, the second country to use tanks in battle, the word tank or tanque was adopted initially, but was then, largely at the insistence of Colonel J.B.E. Estienne, rejected in favour of char d'assaut ("assault vehicle") or simply char ("vehicle"). During World War I, German sources tended to refer to British tanks as Tanks and to their own as Kampfwagen. Later, tanks became referred to as "Panzer" (lit. "armour"), a shortened form of the full term "Panzerkampfwagen", literally "armoured fighting vehicle". In the Arab world, tanks are called Dabbāba (after a type of siege engine). In Italian, a tank is a "carro armato" (lit. "armed wagon"), without reference to its armour. Norway uses the term stridsvogn and Sweden the similar stridsvagn (lit. "battle wagon", also used for "chariots"), whereas Denmark uses kampvogn (lit. fight wagon). Finland uses panssarivaunu (armoured wagon), although tankki is also used colloquially. The Polish name czołg, derived from verb czołgać się ("to crawl"), is used, depicting the way of machine's movement and its speed. In Hungarian the tank is called harckocsi (combat wagon), albeit tank is also common. In Japanese, the term sensha (戦車, lit. "battle vehicle") is taken from Chinese and used, and this term is likewise borrowed into Korean as jeoncha (전차/戰車); more recent Chinese literature uses the English-derived 坦克 tǎnkè (tank) as opposed to 戰車 zhànchē (battle vehicle) used in earlier days. The modern tank is the result of a century of development from the first primitive armoured vehicles, due to improvements in technology such as the internal combustion engine, which allowed the rapid movement of heavy armoured vehicles. As a result of these advances, tanks underwent tremendous shifts in capability in the years since their first appearance. Tanks in World War I were developed separately and simultaneously by Great Britain and France as a means to break the deadlock of trench warfare on the Western Front. The first British prototype, nicknamed Little Willie, was constructed at William Foster & Co. in Lincoln, England in 1915, with leading roles played by Major Walter Gordon Wilson who designed the gearbox and hull, and by William Tritton of William Foster and Co., who designed the track plates. This was a prototype of a new design that would become the British Army's Mark I tank, the first tank used in combat in September 1916 during the Battle of the Somme. The name "tank" was adopted by the British during the early stages of their development, as a security measure to conceal their purpose (see etymology). While the British and French built thousands of tanks in World War I, Germany was unconvinced of the tank's potential, and did not have enough resources, thus it built only twenty. Tanks of the interwar period evolved into the much larger and more powerful designs of World War II. Important new concepts of armoured warfare were developed; the Soviet Union launched the first mass tank/air attack at Khalkhin Gol (Nomonhan) in August 1939, and later developed the T-34, one of the predecessors of the main battle tank. Less than two weeks later, Germany began their large-scale armoured campaigns that would become known as blitzkrieg ("lightning war") – massed concentrations of tanks combined with motorized and mechanized infantry, artillery and air power designed to break through the enemy front and collapse enemy resistance. The widespread introduction of high-explosive anti-tank warheads during the second half of World War II led to lightweight infantry-carried anti-tank weapons such as the Panzerfaust, which could destroy some types of tanks. Tanks in the Cold War were designed with these weapons in mind, and led to greatly improved armour types during the 1960s, especially composite armour. Improved engines, transmissions and suspensions allowed tanks of this period to grow larger. Aspects of gun technology changed significantly as well, with advances in shell design and aiming technology. During the Cold War, the main battle tank concept arose and became a key component of modern armies. In the 21st century, with the increasing role of asymmetrical warfare and the end of the Cold War, that also contributed to the increase of cost-effective anti-tank rocket propelled grenades (RPGs) worldwide and its successors, the ability of tanks to operate independently has declined. Modern tanks are more frequently organized into combined arms units which involve the support of infantry, who may accompany the tanks in infantry fighting vehicles, and supported by reconnaissance or ground-attack aircraft. The tank is the 20th-century realization of an ancient concept: that of providing troops with mobile protection and firepower. The internal combustion engine, armour plate, and continuous track were key innovations leading to the invention of the modern tank. Many sources imply that Leonardo da Vinci and H. G. Wells in some way foresaw or "invented" the tank. Leonardo's late-15th-century drawings of what some describe as a "tank" show a man-powered, wheeled vehicle with cannons all around it. However the human crew would not have enough power to move it over larger distance, and usage of animals was problematic in a space so confined. In the 15th century, Jan Žižka built armoured wagons containing cannons and used them effectively in several battles. The continuous "caterpillar track" arose from attempts to improve the mobility of wheeled vehicles by spreading their weight, reducing ground pressure, and increasing their traction. Experiments can be traced back as far as the 17th century, and by the late nineteenth they existed in various recognizable and practical forms in several countries. It is frequently claimed that Richard Lovell Edgeworth created a caterpillar track. It is true that in 1770 he patented a "machine, that should carry and lay down its own road", but this was Edgeworth's choice of words. His own account in his autobiography is of a horse-drawn wooden carriage on eight retractable legs, capable of lifting itself over high walls. The description bears no similarity to a caterpillar track. Armoured trains appeared in the mid-19th century, and various armoured steam and petrol-engined vehicles were also proposed. The machines described in Wells's 1903 short story The Land Ironclads are a step closer, insofar as they are armour-plated, have an internal power plant, and are able to cross trenches. Some aspects of the story foresee the tactical use and impact of the tanks that later came into being. However, Wells's vehicles were driven by steam and moved on pedrail wheels, technologies that were already outdated at the time of writing. After seeing British tanks in 1916, Wells denied having "invented" them, writing, "Yet let me state at once that I was not their prime originator. I took up an idea, manipulated it slightly, and handed it on." It is, though, possible that one of the British tank pioneers, Ernest Swinton, was subconsciously or otherwise influenced by Wells's tale. The first combinations of the three principal components of the tank appeared in the decade before World War One. In 1903, Captain Léon René Levavasseur of the French artillery proposed mounting a field gun in an armoured box on tracks. Major William E. Donohue, of the British Army's Mechanical Transport Committee, suggested fixing a gun and armoured shield on a British type of track-driven vehicle. The first armoured car was produced in Austria in 1904. However, all were restricted to rails or reasonably passable terrain. It was the development of a practical caterpillar track that provided the necessary independent, all-terrain mobility. In a memorandum of 1908, Antarctic explorer Robert Falcon Scott presented his view that man-hauling to the South Pole was impossible and that motor traction was needed. Snow vehicles did not yet exist, however, and so his engineer Reginald Skelton developed the idea of a caterpillar track for snow surfaces. These tracked motors were built by the Wolseley Tool and Motor Car Company in Birmingham and tested in Switzerland and Norway, and can be seen in action in Herbert Ponting's 1911 documentary film of Scott's Antarctic Terra Nova Expedition. Scott died during the expedition in 1912, but expedition member and biographer Apsley Cherry-Garrard credited Scott's "motors" with the inspiration for the British World War I tanks, writing: "Scott never knew their true possibilities; for they were the direct ancestors of the 'tanks' in France". In 1911, a Lieutenant Engineer in the Austrian Army, Günther Burstyn, presented to the Austrian and Prussian War Ministries plans for a light, three-man tank with a gun in a revolving turret, the so-called Burstyn-Motorgeschütz. In the same year an Australian civil engineer named Lancelot de Mole submitted a basic design for a tracked, armoured vehicle to the British War Office. In Russia, Vasiliy Mendeleev designed a tracked vehicle containing a large naval gun. All of these ideas were rejected and, by 1914, forgotten (although it was officially acknowledged after the war that de Mole's design was at least the equal to the initial British tanks). Various individuals continued to contemplate the use of tracked vehicles for military applications, but by the outbreak of the War no one in a position of responsibility in any army seems to have given much thought to tanks. The direct military impact of the tank can be debated but its effect on the Germans was immense, it caused bewilderment, terror and concern in equal measure. It was also a huge boost to the civilians at home. After facing the Zeppelins, at last Britain had a wonder weapon. Tanks were taken on tours and treated almost like film stars. From late 1914 a small number of middle-ranking British Army officers tried to persuade the War Office and the Government to consider the creation of armoured vehicles. Amongst their suggestions was the use of caterpillar tractors, but although the Army used many such vehicles for towing heavy guns, it could not be persuaded that they could be adapted as armoured vehicles. The consequence was that early tank development in the United Kingdom was carried out by the Royal Navy. As the result of an approach by Royal Naval Air Service officers who had been operating armoured cars on the Western Front, the First Lord of the Admiralty, Winston Churchill, formed the Landship Committee, on 20 February 1915. The Director of Naval Construction for the Royal Navy, Eustace Tennyson d'Eyncourt, was appointed to head the Committee in view of his experience with the engineering methods it was felt might be required; the two other members were naval officers, and a number of industrialists were engaged as consultants. So many played a part in its long and complicated development that it is not possible to name any individual as the sole inventor of the tank. However leading roles were played by Lt Walter Gordon Wilson R.N. who designed the gearbox and developed practical tracks and by William Tritton whose agricultural machinery company, William Foster & Co. in Lincoln, Lincolnshire, England built the prototypes. On 22 July 1915, a commission was placed to design a machine that could cross a trench 4 ft wide. Secrecy surrounded the project with the designers locking themselves in a room at the White Hart Hotel in Lincoln. The committee's first design, Little Willie, ran for the first time in September 1915 and served to develop the form of the track but an improved design, better able to cross trenches, swiftly followed and in January 1916 the prototype, nicknamed "Mother", was adopted as the design for future tanks. The first order for tanks was placed on 12 February 1916, and a second on 21 April. Fosters built 37 (all "male"), and Metropolitan Railway Carriage and Wagon Company, of Birmingham, 113 (38 "male" and 75 "female"), a total of 150. Production models of "Male" tanks (armed with naval cannon and machine guns) and "Females" (carrying only machine-guns) would go on to fight in history's first tank action at the Somme in September 1916. Great Britain produced about 2,600 tanks of various types during the war. The first tank to engage in battle was designated D1, a British Mark I Male, during the Battle of Flers-Courcelette (part of the wider Somme offensive) on 15 September 1916. Bert Chaney, a nineteen-year-old signaller with the 7th London Territorial Battalion, reported that "three huge mechanical monsters such as [he] had never seen before" rumbled their way onto the battlefield, "frightening the Jerries out of their wits and making them scuttle like frightened rabbits." When the news of the first use of the tanks emerged, Prime Minister David Lloyd George commented, It is really to Mr Winston Churchill that the credit is due more than to anyone else. He took up with enthusiasm the idea of making them a long time ago, and he met with many difficulties. He converted me, and at the Ministry of Munitions he went ahead and made them. The admiralty experts were invaluable, and gave the greatest possible assistance. They are, of course, experts in the matter of armour plating. Major Stern, (formerly an officer in the Royal Naval Air Service) a business man at the Ministry of Munitions had charge of the work of getting them built, and he did the task very well. Col Swinton and others also did valuable work. Whilst several experimental machines were investigated in France, it was a colonel of artillery, J.B.E. Estienne, who directly approached the Commander-in-Chief with detailed plans for a tank on caterpillar tracks, in late 1915. The result was two largely unsatisfactory types of tank, 400 each of the Schneider and Saint-Chamond, both based on the Holt Tractor. The following year, the French pioneered the use of a full 360° rotation turret in a tank for the first time, with the creation of the Renault FT light tank, with the turret containing the tank's main armament. In addition to the traversable turret, another innovative feature of the FT was its engine located at the rear. This pattern, with the gun located in a mounted turret and the engine at the back, has become the standard for most succeeding tanks across the world even to this day. The FT was the most numerous tank of the war; over 3,000 were made by late 1918. Germany fielded very few tanks during World War I, and started development only after encountering British tanks on the Somme. The A7V, the only type made, was introduced in March 1918 with just 20 being produced during the war. The first tank versus tank action took place on 24 April 1918 at the Second Battle of Villers-Bretonneux, France, when three British Mark IVs met three German A7Vs. Captured British Mk IVs formed the bulk of Germany's tank forces during World War I; about 35 were in service at any one time. Plans to expand the tank programme were under way when the War ended. The United States Tank Corps used tanks supplied by France and Great Britain during World War I. Production of American-built tanks had just begun when the War came to an end. Italy also manufactured two Fiat 2000s towards the end of the war, too late to see service. Russia independently built and trialed two prototypes early in the War; the tracked, two-man Vezdekhod and the huge Lebedenko, but neither went into production. A tracked self-propelled gun was also designed but not produced. Although tank tactics developed rapidly during the war, piecemeal deployments, mechanical problems, and poor mobility limited the military significance of the tank in World War I, and the tank did not fulfil its promise of rendering trench warfare obsolete. Nonetheless, it was clear to military thinkers on both sides that tanks in some way could have a significant role in future conflicts. In the interwar period tanks underwent further mechanical development. In terms of tactics, J.F.C. Fuller's doctrine of spearhead attacks with massed tank formations was the basis for work by Heinz Guderian in Germany, Percy Hobart in Britain, Adna R. Chaffee, Jr., in the US, Charles de Gaulle in France, and Mikhail Tukhachevsky in the USSR. Liddell Hart held a more moderate view that all arms – cavalry, infantry and artillery – should be mechanized and work together. The British formed the all-arms Experimental Mechanized Force to test the use of tanks with supporting forces. In the Second World War only Germany would initially put the theory into practice on a large scale, and it was their superior tactics and French blunders, not superior weapons, that made the "blitzkrieg" so successful in May 1940. For information regarding tank development in this period, see tank development between the wars. Germany, Italy and the Soviet Union all experimented heavily with tank warfare during their clandestine and "volunteer" involvement in the Spanish Civil War, which saw some of the earliest examples of successful mechanized combined arms —such as when Republican troops, equipped with Soviet-supplied tanks and supported by aircraft, eventually routed Italian troops fighting for the Nationalists in the seven-day Battle of Guadalajara in 1937. However, of the nearly 700 tanks deployed during this conflict, only about 64 tanks representing the Franco faction and 331 from the Republican side were equipped with cannon, and of those 64 nearly all were World War I vintage Renault FT tanks, while the 331 Soviet supplied machines had 45mm main guns and were of 1930s manufacture. The balance of Nationalist tanks were machine gun armed. The primary lesson learned from this war was that machine gun armed tanks had to be equipped with cannon, with the associated armour inherent to modern tanks. The five-month-long war between the Soviet Union and the Japanese 6th Army at Khalkhin Gol (Nomonhan) in 1939 brought home some lessons. In this conflict, the Soviets fielded over two thousand tanks, to the around 73 cannon armed tanks deployed by the Japanese, the major difference being that Japanese armour were equipped with diesel engines as opposed to the Russian tanks equipped with petrol engines. After General Georgy Zhukov inflicted a defeat on the Japanese 6th Army with his massed combined tank and air attack, the Soviets learned a lesson on the use of gasoline engines, and quickly incorporated those newly found experiences into their new T-34 medium tank during World War II. Prior to World War II, the tactics and strategy of deploying tank forces underwent a revolution. In August 1939, Soviet General Georgy Zhukov used the combined force of tanks and airpower at Nomonhan against the Japanese 6th Army; Heinz Guderian, a tactical theoretician who was heavily involved in the formation of the first independent German tank force, said "Where tanks are, the front is", and this concept became a reality in World War II. Guderian's armoured warfare ideas, combined with Germany's existing doctrines of Bewegungskrieg ("maneuver warfare") and infiltration tactics from World War I, became the basis of blitzkrieg in the opening stages of World War II. During World War II, the first conflict in which armoured vehicles were critical to battlefield success, the tank and related tactics developed rapidly. Armoured forces proved capable of tactical victory in an unprecedentedly short amount of time, yet new anti-tank weaponry showed that the tank was not invulnerable. During the Invasion of Poland, tanks performed in a more traditional role in close cooperation with infantry units, but in the Battle of France deep independent armoured penetrations were executed by the Germans, a technique later called blitzkrieg. Blitzkrieg used innovative combined arms tactics and radios in all of the tanks to provide a level of tactical flexibility and power that surpassed that of the Allied armour. The French Army, with tanks equal or superior to the German tanks in both quality and quantity, employed a linear defensive strategy in which the armoured cavalry units were made subservient to the needs of the infantry armies to cover their entrenchment in Belgium. In addition, they lacked radios in many of their tanks and headquarters, which limited their ability to respond to German attacks. In accordance with blitzkrieg methods, German tanks bypassed enemy strongpoints and could radio for close air support to destroy them, or leave them to the infantry. A related development, motorized infantry, allowed some of the troops to keep up with the tanks and create highly mobile combined arms forces. The defeat of a major military power within weeks shocked the rest of the world, spurring tank and anti-tank weapon development. The North African Campaign also provided an important battleground for tanks, as the flat, desolate terrain with relatively few obstacles or urban environments was ideal for conducting mobile armoured warfare. However, this battlefield also showed the importance of logistics, especially in an armoured force, as the principal warring armies, the German Afrika Korps and the British Eighth Army, often outpaced their supply trains in repeated attacks and counter-attacks on each other, resulting in complete stalemate. This situation would not be resolved until 1942, when during the Second Battle of El Alamein, the Afrika Korps, crippled by disruptions in their supply lines, had 95% of its tanks destroyed and was forced to retreat by a massively reinforced Eighth Army, the first in a series of defeats that would eventually lead to the surrender of the remaining Axis forces in Tunisia. When Germany launched its invasion of the Soviet Union, Operation Barbarossa, the Soviets had a superior tank design, the T-34. A lack of preparations for the Axis surprise attack, mechanical problems, poor training of the crews and incompetent leadership caused the Soviet machines to be surrounded and destroyed in large numbers. However, interference from Adolf Hitler, the geographic scale of the conflict, the dogged resistance of the Soviet combat troops, and the Soviets' massive advantages in manpower and production capability prevented a repeat of the German successes of 1940. Despite early successes against the Soviets, the Germans were forced to up-gun their Panzer IVs, and to design and build both the larger and more expensive Tiger heavy tank in 1942, and the Panther medium tank the following year. In doing so, the Wehrmacht denied the infantry and other support arms the production priorities that they needed to remain equal partners with the increasingly sophisticated tanks, in turn violating the principle of combined arms they had pioneered. Soviet developments following the invasion included upgunning the T-34, development of self-propelled anti-tank guns such as the SU-152, and deployment of the IS-2 in the closing stages of the war, with the T-34 being the most produced tank of World War II, totalling up to some 65,000 examples by May 1945. Much like the Soviets, when entering World War II six months later (December 1941), the United States' mass production capacity enabled it to rapidly construct thousands of relatively cheap M4 Sherman medium tanks. A compromise all round, the Sherman was reliable and formed a large part of the Anglo-American ground forces, but in a tank-versus-tank battle was no match for the Panther or Tiger. Numerical and logistical superiority and the successful use of combined arms allowed the Allies to overrun the German forces during the Battle of Normandy. Upgunned versions with the 76 mm gun M1 and the 17-pounder were introduced to improve the M4's firepower, but concerns about protection remained—despite the apparent armour deficiencies, a total of some 42,000 Shermans were built and delivered to the Allied nations using it during the war years, a total second only to the T-34. Tank hulls were modified to produce flame tanks, mobile rocket artillery, and combat engineering vehicles for tasks including mine-clearing and bridging. Specialized self-propelled guns, most of which could double as tank destroyers, were also both developed by the Germans—with their Sturmgeschütz, Panzerjäger and Jagdpanzer vehicles—and the Samokhodnaya ustanovka families of AFV's for the Soviets: such turretless, casemate-style tank destroyers and assault guns were less complex, stripped down tanks carrying heavy guns, solely firing forward. The firepower and low cost of these vehicles made them attractive but as manufacturing techniques improved and larger turret rings made larger tank guns feasible, the gun turret was recognized as the most effective mounting for the main gun to allow movement in a different direction from firing, enhancing tactical flexibility. During the Cold War, tension between the Warsaw Pact countries and North Atlantic Treaty Organization (NATO) countries created an arms race that ensured that tank development proceeded largely as it had during World War II. The essence of tank designs during the Cold War had been hammered out in the closing stages of World War II. Large turrets, capable suspension systems, greatly improved engines, sloped armour and large-caliber (90 mm and larger) guns were standard. Tank design during the Cold War built on this foundation and included improvements to fire control, gyroscopic gun stabilization, communications (primarily radio) and crew comfort and saw the introduction of laser rangefinders and infrared night vision equipment. Armour technology progressed in an ongoing race against improvements in anti-tank weapons, especially antitank guided missiles like the TOW. Medium tanks of World War II evolved into the main battle tank (MBT) of the Cold War and took over the majority of tank roles on the battlefield. This gradual transition occurred in the 1950s and 1960s due to anti-tank guided missiles, sabot ammunition and high-explosive anti-tank warheads. World War II had shown that the speed of a light tank was no substitute for armour & firepower and medium tanks were vulnerable to newer weapon technology, rendering them obsolete. In a trend started in World War II, economies of scale led to serial production of progressively upgraded models of all major tanks during the Cold War. For the same reason many upgraded post-World War II tanks and their derivatives (for example, the T-55 and T-72) remain in active service around the world, and even an obsolete tank may be the most formidable weapon on battlefields in many parts of the world. Among the tanks of the 1950s were the British Centurion and Soviet T-54/55 in service from 1946, and the US M48 from 1951. These three vehicles formed the bulk of the armoured forces of NATO and the Warsaw Pact throughout much of the Cold War. Lessons learned from tanks such as the Leopard 1, M48 Patton series, Chieftain, and T-72 led to the contemporary Leopard 2, M1 Abrams, Challenger 2, C1 Ariete, T-90 and Merkava IV. Tanks and anti-tank weapons of the Cold War era saw action in a number of proxy wars like the Korean War, Vietnam War, Indo-Pakistani War of 1971, Soviet–Afghan War and Arab-Israeli conflicts, culminating with the Yom Kippur War. The T-55, for example, has seen action in no fewer than 32 conflicts. In these wars the U.S. or NATO countries and the Soviet Union or China consistently backed opposing forces. Proxy wars were studied by Western and Soviet military analysts and provided a contribution to the Cold War tank development process. The role of tank vs. tank combat is becoming diminished. Tanks work in concert with infantry in urban warfare by deploying them ahead of the platoon. When engaging enemy infantry, tanks can provide covering fire on the battlefield. Conversely, tanks can spearhead attacks when infantry are deployed in personnel carriers. Tanks were used to spearhead the initial US invasion of Iraq in 2003. As of 2005, there were 1,100 M1 Abrams used by the United States Army in the course of the Iraq War, and they have proven to have an unexpectedly high level of vulnerability to roadside bombs. A relatively new type of remotely detonated mine, the explosively formed penetrator has been used with some success against American armoured vehicles (particularly the Bradley fighting vehicle). However, with upgrades to their armour in the rear, M1s have proven invaluable in fighting insurgents in urban combat, particularly at the Battle of Fallujah, where the US Marines brought in two extra brigades. Israeli Merkava tanks contain features that enable them to support infantry in low intensity conflicts (LIC) and counter-terrorism operations. Such features are the rear door and rear corridor, enabling the tank to carry infantry and embark safely; the IMI APAM-MP-T multi-purpose ammunition round, advanced C4IS systems and recently: TROPHY active protection system which protects the tank from shoulder-launched anti-tank weapons. During the Second Intifada further modifications were made, designated as "Merkava Mk. 3d Baz LIC". In terms of firepower, the focus of 2010s-era R&D was increased detection capability such as thermal imagers, automated fire control systems for the guns and increased muzzle energy from the gun to improve range, accuracy and armour penetration. The most mature future gun technology is the electrothermal-chemical gun. The XM291 electrothermal-chemical tank gun has gone through successful multiple firing sequences on a modified American M8 Armored Gun System chassis. To improve tank protection, one field of research involves making the tank invisible to radar by adapting stealth technologies originally designed for aircraft. Improvements to camouflage or and attempts to render it invisible through active camouflage, which changes according to where the tank is located, are being pursued. Research is also ongoing in electromagnetic armour systems to disperse or deflect incoming shaped charges, as well as various forms of active protection systems to prevent incoming projectiles (RPGs, missiles, etc.) from striking the tank. Mobility may be enhanced in future tanks by the use of diesel-electric or turbine-electric series hybrid drives—first used in a primitive, gasoline-engined form with Porsche's Elefant German tank destroyer of 1943—improving fuel efficiency while reducing the size and weight of the power plant. Furthermore, advances in gas turbine technology, including the use of advanced recuperators, have allowed for reduction in engine volume and mass to less than 1 m and 1 metric ton, respectively, while maintaining fuel efficiency similar to that of a diesel engine. In line with the new doctrine of network-centric warfare, the 2010s-era modern battle tank shows increasing sophistication in its electronics and communication systems. The future of tanks has been challenged by the proliferation of relatively inexpensive anti tank guided missiles and rockets during the Russo-Ukrainian War. The three traditional factors determining a tank's capability effectiveness are its firepower, protection, and mobility. Firepower is the ability of a tank's crew to identify, engage, and destroy enemy tanks and other targets using its large-caliber cannon. Protection is the degree to which the tank's armour, profile and camouflage enables the tank crew to evade detection, protect themselves from enemy fire, and retain vehicle functionality during and after combat. Mobility includes how well the tank can be transported by rail, sea, or air to the operational staging area; from the staging area by road or over terrain towards the enemy; and tactical movement by the tank over the battlefield during combat, including traversing of obstacles and rough terrain. The variations of tank designs have been determined by the way these three fundamental features are blended. For instance, in 1937, the French doctrine focused on firepower and protection more than mobility because tanks worked in intimate liaison with the infantry. There was also the case of the development of a heavy cruiser tank, which focused on armour and firepower to challenge Germany's Tiger and Panther tanks. Tanks have been classified by weight, role, or other criteria, that has changed over time and place. Classification is determined by the prevailing theories of armoured warfare, which have been altered in turn by rapid advances in technology. No one classification system works across all periods or all nations; in particular, weight-based classification is inconsistent between countries and eras. In World War I, the first tank designs focused on crossing wide trenches, requiring very long and large vehicles, such as the British Mark I; these became classified as heavy tanks. Tanks that fulfilled other combat roles were smaller, like the French Renault FT; these were classified as light tanks or tankettes. Many late-war and inter-war tank designs diverged from these according to new, though mostly untried, concepts for future tank roles and tactics. Tank classifications varied considerably according to each nation's own tank development, such as "cavalry tanks", "fast tanks", and "breakthrough tanks". During World War II, many tank concepts were found unsatisfactory and discarded, mostly leaving the more multi-role tanks; these became easier to classify. Tank classes based on weight (and the corresponding transport and logistical needs) led to new definitions of heavy and light tank classes, with medium tanks covering the balance of those between. The British maintained cruiser tanks, focused on speed, and infantry tanks that traded speed for more armour. Tank destroyers are tanks or other armoured fighting vehicles specifically designed to defeat enemy tanks. Assault guns are armoured fighting vehicles that could combine the roles of infantry tanks and tank destroyers. Some tanks were converted to flame tanks, specializing on close-in attacks on enemy strongholds with flamethrowers. As the war went on, tanks tended to become larger and more powerful, shifting some tank classifications and leading to super-heavy tanks. Experience and technology advances during the Cold War continued to consolidate tank roles. With the worldwide adoption of the modern main battle tank designs, which favour a modular universal design, most other classifications are dropped from modern terminology. All main battle tanks tend to have a good balance of speed, armour, and firepower, even while technology continues to improve all three. Being fairly large, main battle tanks can be complemented with light tanks, armoured personnel carriers, infantry fighting vehicles or similar relatively lighter armoured fighting vehicles, typically in the roles of armoured reconnaissance, amphibious or air assault operations, or against enemies lacking main battle tanks. The main weapon of modern tanks is typically a single, large-caliber cannon mounted in a fully traversing (rotating) gun turret. The typical modern tank gun is a smoothbore weapon capable of firing a variety of ammunition, including armour-piercing kinetic energy penetrators (KEP), also known as armour-piercing discarding sabot (APDS), and/or armour-piercing fin-stabilized discarding sabot (APFSDS) and high-explosive anti-tank (HEAT) shells, and/or high-explosive squash head (HESH) and/or anti-tank guided missiles (ATGM) to destroy armoured targets, as well as high-explosive (HE) shells for shooting at "soft" targets (unarmoured vehicles or troops) or fortifications. Canister shot may be used in close or urban combat situations where the risk of hitting friendly forces with shrapnel from HE rounds is unacceptably high. A gyroscope is used to stabilise the main gun, allowing it to be effectively aimed and fired at the "short halt" or on the move. Modern tank guns are also commonly fitted with insulating thermal sleeves to reduce gun-barrel warping caused by uneven thermal expansion, bore evacuators to minimise gun firing fumes entering the crew compartment and sometimes muzzle brakes to minimise the effect of recoil on accuracy and rate of fire. Traditionally, target detection relied on visual identification. This was accomplished from within the tank through telescopic periscopes; often, however, tank commanders would open up the hatch to view the outside surroundings, which improved situational awareness but incurred the penalty of vulnerability to sniper fire. Though several developments in target detection have taken place, these methods are still common practice. In the 2010s, more electronic target detection methods are available. In some cases spotting rifles were used to confirm proper trajectory and range to a target. These spotting rifles were mounted co-axially to the main gun, and fired tracer ammunition ballistically matched to the gun itself. The gunner would track the movement of the tracer round in flight, and upon impact with a hard surface, it would give off a flash and a puff of smoke, after which the main gun was immediately fired. However this slow method has been mostly superseded by laser rangefinding equipment. Modern tanks also use sophisticated light intensification and thermal imaging equipment to improve fighting capability at night, in poor weather and in smoke. The accuracy of modern tank guns is pushed to the mechanical limit by computerized fire-control systems. A fire-control system uses a laser rangefinder to determine the range to the target, a thermocouple, anemometer and wind vane to correct for weather effects and a muzzle referencing system to correct for gun-barrel temperature, warping and wear. Two sightings of a target with the range-finder enable calculation of the target movement vector. This information is combined with the known movement of the tank and the principles of ballistics to calculate the elevation and aim point that maximises the probability of hitting the target. Usually, tanks carry smaller caliber armament for short-range defense where fire from the main weapon would be ineffective or wasteful, for example when engaging infantry, light vehicles or close air support aircraft. A typical complement of secondary weapons is a general-purpose machine gun mounted coaxially with the main gun, and a heavier anti-aircraft-capable machine gun on the turret roof. Some tanks also have a hull-mounted machine gun. These weapons are often modified variants of those used by infantry, and so use the same kinds of ammunition. The measure of a tank's protection is the combination of its ability to avoid detection (due to having a low profile and through the use of camouflage), to avoid being hit by enemy fire, its resistance to the effects of enemy fire, and its capacity to sustain damage whilst still completing its objective, or at least protecting its crew. This is done by a variety of countermeasures, such as armour plating and reactive defenses, as well as more complex ones such as heat-emissions reduction. In common with most unit types, tanks are subject to additional hazards in dense wooded and urban combat environments which largely negate the advantages of the tank's long-range firepower and mobility, limit the crew's detection capabilities and can restrict turret traverse. Despite these disadvantages, tanks retain high survivability against previous-generation rocket-propelled grenades aimed at the most-armoured sections. However, as effective and advanced as armour plating has become, tank survivability against newer-generation tandem-warhead anti-tank missiles is a concern for military planners. Tandem-warhead RPGs use two warheads to fool active protection systems; a first dummy warhead is fired first, to trigger the active defenses, with the real warhead following it. For example, the RPG-29 from the 1980s is able to penetrate the frontal hull armour of the Challenger II and also managed to damage a M1 Abrams. As well, even tanks with advanced armour plating can have their tracks or gear cogs damaged by RPGs, which may render them immobile or hinder their mobility. Despite all of the advances in armour plating, a tank with its hatches open remains vulnerable to Molotov cocktail (gasoline bombs) and grenades. Even a "buttoned up" tank may have components which are vulnerable to Molotov cocktails, such as optics, extra gas cans and extra ammunition stored on the outside of the tank. A tank avoids detection using the doctrine of countermeasures known as CCD: camouflage (looks the same as the surroundings), concealment (cannot be seen) and deception (looks like something else). Camouflage can include disruptive painted shapes on the tank to break up the distinctive appearance and silhouette of a tank. Netting or actual branches from the surrounding landscape are also used. Prior to development of infrared technology, tanks were often given a coating of camouflage paint that, depending on environmental region or season, would allow it to blend in with the rest of its environment. A tank operating in wooded areas would typically get a green and brown paintjob; a tank in a winter environment would get white paint (often mixed with some darker colors); tanks in the desert often get khaki paintjobs. The Russian Nakidka camouflage kit was designed to reduce the optical, thermal, infrared, and radar signatures of a tank, so that acquisition of the tank would be difficult. According to Nii Stali, the designers of Nakidka, Nakidka would reduce the probabilities of detection via "visual and near-IR bands by 30%, the thermal band by 2–3-fold, radar band by 6-fold, and radar-thermal band to near-background levels. Concealment can include hiding the tank among trees or digging in the tank by having a combat bulldozer dig out part of a hill, so that much of the tank will be hidden. A tank commander can conceal the tank by using "hull down" approaches to going over upward-sloping hills, so that she or he can look out the commander's cupola without the distinctive-looking main cannon cresting over the hill. Adopting a turret-down or hull-down position reduces the visible silhouette of a tank as well as providing the added protection of a position in defilade. Working against efforts to avoid detection is the fact that a tank is a large metallic object with a distinctive, angular silhouette that emits copious heat and engine noise. A tank that is operating in cold weather or which needs to use its radio or other communications or target-detecting electronics will need to start its engine regularly to maintain its battery power, which will create engine noise. Consequently, it is difficult to effectively camouflage a tank in the absence of some form of cover or concealment (e.g., woods) it can hide its hull behind. The tank becomes easier to detect when moving (typically, whenever it is in use) due to the large, distinctive auditory, vibration and thermal signature of its engine and power plant. Tank tracks and dust clouds also betray past or present tank movement. Switched-off tanks are vulnerable to infra-red detection due to differences between the thermal conductivity and therefore heat dissipation of the metallic tank and its surroundings. At close range the tank can be detected even when powered down and fully concealed due to the column of warmer air above the tank and the smell of diesel or gasoline. Thermal blankets slow the rate of heat emission and some thermal camouflage nets use a mix of materials with differing thermal properties to operate in the infra-red as well as the visible spectrum. Grenade launchers can rapidly deploy a smoke screen that is opaque to infrared light, to hide it from the thermal viewer of another tank. In addition to using its own grenade launchers, a tank commander could call in an artillery unit to provide smoke cover. Some tanks can produce a smoke screen. Sometimes camouflage and concealment are used at the same time. For example, a camouflage-painted and branch-covered tank (camouflage) may be hidden in a behind a hill or in a dug-in-emplacement (concealment). Some armoured recovery vehicles (often tracked, tank chassis-based "tow trucks" for tanks) have dummy turrets and cannons. This makes it less likely that enemy tanks will fire on these vehicles. Some armies have fake "dummy" tanks made of wood which troops can carry into position and hide behind obstacles. These "dummy" tanks may cause the enemy to think that there are more tanks than are actually possessed. To effectively protect the tank and its crew, tank armour must counter a wide variety of antitank threats. Protection against kinetic energy penetrators and high-explosive anti-tank (HEAT) shells fired by other tanks is of primary importance, but tank armour also aims to protect against infantry mortars, grenades, rocket-propelled grenades, anti-tank guided missiles, anti-tank mines, anti-tank rifles, bombs, direct artillery hits, and (less often) nuclear, biological and chemical threats, any of which could disable or destroy a tank or its crew. Steel armour plate was the earliest type of armour. The Germans pioneered the use of face hardened steel during World War II and the Soviets also achieved improved protection with sloped armour technology. World War II developments led to the obsolescence of homogeneous steel armour with the development of shaped-charge warheads, exemplified by the Panzerfaust and bazooka infantry-carried weapons which were effective, despite some early success with spaced armour. Magnetic mines led to the development of anti-magnetic paste and paint. From WWII to the modern era, troops have added improvised armour to tanks while in combat settings, such as sandbags or pieces of old armour plating. British tank researchers took the next step with the development of Chobham armour, or more generally composite armour, incorporating ceramics and plastics in a resin matrix between steel plates, which provided good protection against HEAT weapons. High-explosive squash head warheads led to anti-spall armour linings, and kinetic energy penetrators led to the inclusion of exotic materials like a matrix of depleted uranium into a composite armour configuration. Reactive armour consists of small explosive-filled metal boxes that detonate when hit by the metallic jet projected by an exploding HEAT warhead, causing their metal plates to disrupt it. Tandem warheads defeat reactive armour by causing the armour to detonate prematurely. Modern reactive armour protects itself from Tandem warheads by having a thicker front metal plate to prevent the precursor charge from detonating the explosive in the reactive armour. Reactive armours can also reduce the penetrative abilities of kinetic energy penetrators by deforming the penetrator with the metal plates on the Reactive armour, thereby reducing its effectiveness against the main armour of the tank. The latest generation of protective measures for tanks are active protection systems. The term "active" is used to contrast these approaches with the armour used as the primary protective approach in earlier tanks. The mobility of a tank is described by its battlefield or tactical mobility, its operational mobility, and its strategic mobility. Tank agility is a function of the weight of the tank due to its inertia while manoeuvring and its ground pressure, the power output of the installed power plant and the tank transmission and track design. In addition, rough terrain effectively limits the tank's speed through the stress it puts on the suspension and the crew. A breakthrough in this area was achieved during World War II when improved suspension systems were developed that allowed better cross-country performance and limited firing on the move. Systems like the earlier Christie or later torsion-bar suspension developed by Ferdinand Porsche dramatically improved the tank's cross-country performance and overall mobility. Tanks are highly mobile and able to travel over most types of terrain due to their continuous tracks and advanced suspension. The tracks disperse the weight of the vehicle over a large area, resulting in less ground pressure. A tank can travel at approximately 40 kilometres per hour (25 mph) across flat terrain and up to 70 kilometres per hour (43 mph) on roads, but due to the mechanical strain this places on the vehicle and the logistical strain on fuel delivery and tank maintenance, these must be considered "burst" speeds that invite mechanical failure of engine and transmission systems. Consequently, wheeled tank transporters and rail infrastructure is used wherever possible for long-distance tank transport. The limitations of long-range tank mobility can be viewed in sharp contrast to that of wheeled armoured fighting vehicles. The majority of blitzkrieg operations were conducted at the pedestrian pace of 5 kilometres per hour (3.1 mph), and that was only achieved on the roads of France. The tank's power plant supplies kinetic energy to move the tank, and electric power via a generator to components such as the turret rotation motors and the tank's electronic systems. The tank power plant evolved from predominantly petrol and adapted large-displacement aeronautical or automotive engines to diesel engines. Japan was the first to begin transitioning to this engine type beginning with the Type 89B in 1934. The main advantage of diesel is their higher fuel economy, which allows for greater operating ranges. Diesel engines can also run on a variety of fuels, such as aviation kerosene and even gasoline. Advanced multi-fuel diesel engines have been adopted. Powerful (per unit weight) but fuel-hungry gas turbines have seen some limited use, namely by the Soviet T-80 and American M1 Abrams. The transition from petrol to diesel In the absence of combat engineers, most tanks are limited to fording small rivers. The typical fording depth for MBTs is approximately 1 m (3.3 ft), being limited by the height of the engine air intake and driver's position. Modern tanks such as the Russian T-90 and the German Leopard 1 and Leopard 2 tanks can ford to a depth of 3 to 4 m (9.8 to 13.1 ft) when properly prepared and equipped with a snorkel to supply air for the crew and engine. Tank crews usually have a negative reaction towards deep fording but it adds considerable scope for surprise and tactical flexibility in water crossing operations by opening new and unexpected avenues of attack. Amphibious tanks are specially designed or adapted for water operations, such as by including snorkels and skirts, but they are rare in modern armies, being replaced by purpose-built amphibious assault vehicles or armoured personnel carriers in amphibious assaults. Advances such as the EFA mobile bridge and armoured vehicle-launched scissors bridges have also reduced the impediment to tank advance that rivers posed in World War II. Most modern tanks most often have four crew members, or three if an auto-loader is installed. These are the: Operating a tank is a team effort. For example, the loader is assisted by the rest of the crew in stowing ammunition. The driver is assisted in maintaining the automotive features. Historically, crews have varied from just two members to a dozen. First World War tanks were developed with immature technologies; in addition to the crew needed to man the multiple guns and machine guns, up to four crewmen were needed to drive the tank: the driver, acting as the vehicle commander and manning the brakes, drove via orders to his gears-men; a co-driver operated the gearbox and throttle; and two gears-men, one on each track, steered by setting one side or the other to idle, allowing the track on the other side to slew the tank to one side. Pre-World War II French tanks were noted for having a two-man crew, in which the overworked commander had to load and fire the gun in addition to commanding the tank. With World War II the multi-turreted tanks proved impracticable, and as the single turret on a low hull design became standard, crews became standardized around a crew of four or five. In those tanks with a fifth crew member, usually three were located in the turret (as described above) while the fifth was most often seated in the hull next to the driver, and operated the hull machine gun in addition to acting as a co-driver or radio operator. Well-designed crew stations, giving proper considerations to comfort and ergonomics, are an important factor in the combat effectiveness of a tank, as it limits fatigue and speeds up individual actions. A noted author on the subject of tank design engineering, Richard Ogorkiewicz, outlined the following basic engineering sub-systems that are commonly incorporated into tank's technological development: To the above can be added unit communication systems and electronic anti-tank countermeasures, crew ergonomic and survival systems (including flame suppression), and provision for technological upgrading. Few tank designs have survived their entire service lives without some upgrading or modernization, particularly during wartime, including some that have changed almost beyond recognition, such as the latest Israeli Magach versions. The characteristics of a tank are determined by the performance criteria required for the tank. The obstacles that must be traversed affect the vehicles front and rear profiles. The terrain that is expected to be traversed determines the track ground pressure that may be allowed to be exerted for that particular terrain. Tank design is a compromise between its technological and budgetary constraints and its tactical capability requirements. It is not possible to maximise firepower, protection and mobility simultaneously while incorporating the latest technology and retain affordability for sufficient procurement quantity to enter production. For example, in the case of tactical capability requirements, increasing protection by adding armour will result in an increase in weight and therefore decrease in mobility; increasing firepower by installing a larger gun will force the designer team to increase armour, the therefore weight of the tank by retaining same internal volume to ensure crew efficiency during combat. In the case of the Abrams MBT which has good firepower, speed and armour, these advantages are counterbalanced by its engine's notably high fuel consumption, which ultimately reduces its range, and in a larger sense its mobility. Since the Second World War, the economics of tank production governed by the complexity of manufacture and cost, and the impact of a given tank design on logistics and field maintenance capabilities, have also been accepted as important in determining how many tanks a nation can afford to field in its force structure. Some tank designs that were fielded in significant numbers, such as Tiger I and M60A2 proved to be too complex or expensive to manufacture, and made unsustainable demands on the logistics services support of the armed forces. The affordability of the design therefore takes precedence over the combat capability requirements. Nowhere was this principle illustrated better than during the Second World War when two Allied designs, the T-34 and the M4 Sherman, although both simple designs which accepted engineering compromises, were used successfully against more sophisticated designs by Germany that were more complex and expensive to produce, and more demanding on overstretched logistics of the Wehrmacht. Given that a tank crew will spend most of its time occupied with maintenance of the vehicle, engineering simplicity has become the primary constraint on tank design since the Second World War despite advances in mechanical, electrical and electronics technologies. Since the Second World War, tank development has incorporated experimenting with significant mechanical changes to the tank design while focusing on technological advances in the tank's many subsystems to improve its performance. However, a number of novel designs have appeared throughout this period with mixed success, including the Soviet IT-1 and T-64 in firepower, and the Israeli Merkava and Swedish S-tank in protection, while for decades the US's M551 remained the only light tank deployable by parachute. Commanding and coordinating tanks in the field has always been subject to particular problems, particularly in the area of communications, but in modern armies these problems have been partially alleviated by networked, integrated systems that enable communications and contribute to enhanced situational awareness. armoured bulkheads, engine noise, intervening terrain, dust and smoke, and the need to operate "buttoned up" (with hatches closed) are severe detriments to communication and lead to a sense of isolation for small tank units, individual vehicles, and tank crew. Radios were not then portable or robust enough to be mounted in a tank, although Morse code transmitters were installed in some Mark IVs at Cambrai as messaging vehicles. Attaching a field telephone to the rear would become a practice only during the next war. During World War I when these failed or were unavailable, situation reports were sent back to headquarters by some crews releasing carrier pigeons through loopholes or hatches and communications between vehicles was accomplished using hand signals, handheld semaphore flags which continued in use in the Red Army/Soviet Army through the Second and Cold wars, or by foot or horse-mounted messengers. From the beginning, the German military stressed wireless communications, equipping their combat vehicles with radios, and drilled all units to rely on disciplined radio use as a basic element of tactics. This allowed them to respond to developing threats and opportunities during battles, giving the Germans a notable tactical advantage early in the war; even where Allied tanks initially had better firepower and armour, they generally lacked individual radios. By mid-war, Western Allied tanks adopted full use of radios, although Russian use of radios remained relatively limited. On the modern battlefield an intercom mounted in the crew helmet provides internal communications and a link to the radio network, and on some tanks an external intercom on the rear of the tank provides communication with co-operating infantry. Radio networks employ radio voice procedure to minimize confusion and "chatter". A recent development in AFV equipment and doctrine is integration of information from the fire control system, laser rangefinder, Global Positioning System and terrain information via hardened military specification electronics and a battlefield network to display information on enemy targets and friendly units on a monitor in the tank. The sensor data can be sourced from nearby tanks, planes, UAVs or, in the future infantry (such as the US Future Force Warrior project). This improves the tank commander's situational awareness and ability to navigate the battlefield and select and engage targets. In addition to easing the reporting burden by automatically logging all orders and actions, orders are sent via the network with text and graphical overlays. This is known as Network-centric warfare by the US, Network Enabled Capability (UK) or Digital Army Battle Management System צי"ד (Israel). Advanced battle tanks, including the K-2 Black Panther, have taken up the first major step forward in adopting a fully radar integrated Fire Control System which allows it to detect tanks from a further distance and identify it as a friend-or-foe as well as increasing the tank's accuracy as well as its capability to lock onto tanks. Performing situational awareness and communication is the one of four primary MBT functions in the 21st century. To improve the crew's situational awareness MBTs use circular review system with a combination of Augmented reality and Artificial Intelligence technologies. Further advancements in tank defense systems have led to the development of active protection systems. These involve either one of two options:
[[File:Evening view across Carlton Hill towards the Castle (13164472813).jpg|thumb|Edinburgh at night]] [[File:Edinburgh from Calton Hill.jpg|thumb|right|230px|A view of the city from [[Calton Hill]]]] [[File:Edinburgh Castle princes.jpg|thumb|right|230px|[[Edinburgh Castle]] viewed from [[Princes Street]]]] [[File:Edinburgh map.png|thumb|right|230px|Map showing the Old and New towns]] [[Image:Flag of Edinburgh.svg|thumb|150px|Flag]] [[Image:Arms of Edinburgh.png|thumb|90px|coat of arms]] '''Edinburgh''' ([[Scottish Gaelic]]: ''Dùn Èideann'') is the [[capital city]] and second largest city in [[Scotland]]. Edinburgh lies on the east coast, where the [[River Forth]] flows into the sea. The central part is [[Edinburgh Castle]], at the top of a steep hill. The castle has a military display every year, called a ''tattoo'', where soldiers show their skills at marching and competitions, and there are brass bands and bands of [[bagpipes]]. Edinburgh has a very large [[festival]] every year, where thousands of performers come to put on shows. The [[Edinburgh International Festival]] takes place in [[August]] and [[September]]. At the same time there is the [[Edinburgh Fringe]]. The shows are of all kinds, and range from large ones with famous people, to very small ones by new or unknown actors. The city is served by [[Edinburgh Airport]], and Haymarket and Waverley railway stations. == Old and New towns == The Old Town of Edinburgh is the oldest part of the city, and with the 18th-century New Town, it is a [[World Heritage Site|UNESCO World Heritage Site]]. It has preserved its [[mediaeval]] plan and many [[Reformation]]-era buildings. The Old and New Towns are known locally simply as "Town". It is the city centre. The main part of the Old Town is a complex of streets called the [[Royal Mile]] while [[Princes Street]] is the main business area of the New Town. == Sport == The city's main [[football]] teams are [[Heart of Midlothian F.C.]] and [[Hibernian F.C.]] [[Murrayfield stadium]] in the city is the home of the Scottish national [[Rugby football|rugby]] team and Edinburgh Rugby. The nearby Murrayfield Ice Rink is home to the city's [[ice hockey]] team, the Edinburgh Capitals. See also [[Leith Athletic F.C.]] == Buildings == Edinburgh has lots of important buildings. The [[Balmoral Hotel]] was opened in 1902 and designed by [[W. Hamilton Beattie]]. [[Saint Giles' Cathedral]] was built in [[Mediaeval]] times. There was a big fire there in 1385. The [[Royal Museum]] was designed by [[architect]] [[Francis Fowke]] and built between 1861 and 1888. Its sister museum, the [[National Museum of Scotland|Museum of Scotland]], was designed by [[Benson and Forsyth]] in 1998. [[Holyrood Palace]] was once the home of the Scottish kings, and is open to the public. It now serves as the [[Monarchy of the United Kingdom|King of Britain]]'s Scottish residence. The [[Scottish Parliament]] is located next to Holyrood Palace. == Governing == Edinburgh is one of the 32 [[Council areas of Scotland|local government areas of Scotland]]. It has its own council which is led by the [[Provost (civil)|Lord Provost]] (similar to a [[mayor]]). The city is divided into 17 "wards" which elect their councillors. ==Famous people from Edinburgh== * [[Walter Scott|Sir Walter Scott]], who wrote many historical stories. * [[Alexander Graham Bell]], who invented the [[telephone]], was born in Edinburgh.<ref name="WDL">{{cite web |url = http://www.wdl.org/en/item/11375/ |title = Alexander Graham Bell Laboratory Notebook, 1875-1876 |website = [[World Digital Library]] |date = 1875–1876 |accessdate = 2013-07-24 }}</ref> * [[Arthur Conan Doyle|Sir Arthur Conan Doyle]] who wrote the [[Sherlock Holmes]] stories. * [[Robert Louis Stevenson]] who wrote ''Kidnapped'', ''Treasure Island'', ''Doctor Jekyll'' ''and Mr Hyde'' and many other stories. * [[MurielSpark|Muriel Spark]], who wrote ''The Prime of Miss Jean Brodie'' and many other stories. * [[Sean Connery]], who acted as ''[[James Bond]]'' in films. == Twin Towns == Edinburgh is [[Twin town|twinned]] with: * {{flagicon|Italy}} [[Florence]] * {{flagicon|New Zealand}} [[Dunedin]] * {{flagicon|France}} [[Nice]] * {{flagicon|Poland}} [[Kraków]] * {{flagicon|USA}} [[San Diego]] * {{flagicon|Denmark}} [[Aalborg]] * {{flagicon|Ukraine}} [[Kyiv]] * {{flagicon|Japan}} [[Kyoto]] * {{flagicon|China}} [[Xi'an]] * {{flagicon|Canada}} [[Vancouver]] * {{flagicon|Russia}} [[Saint Petersburg|St. Petersburg]] * {{flagicon|Germany}} [[Munich]] ==References== {{reflist}}2. [https://www.bestroute.uk/maps/scotland/edinburgh/ Map of Edinburgh]{{Hanseatic League}} {{British Cities|nocat=yes}} [[Category:Edinburgh| ]] [[Category:Port cities and towns of the North Sea]] [[Category:Port cities and towns of Scotland]]
Edinburgh (/ˈɛdɪnbərə/ Scots: [ˈɛdɪnbʌrə]; Scottish Gaelic: Dùn Èideann [ˌt̪un ˈeːtʲən̪ˠ]) is the capital city of Scotland and one of its 32 council areas. The city is located in south-east Scotland, and is bounded to the north by the Firth of Forth estuary and to the south by the Pentland Hills. Edinburgh had a population of 506,520 in mid-2020, making it the second-most populous city in Scotland and the seventh-most populous in the United Kingdom. Recognised as the capital of Scotland since at least the 15th century, Edinburgh is the seat of the Scottish Government, the Scottish Parliament, the highest courts in Scotland, and the Palace of Holyroodhouse, the official residence of the British monarch in Scotland. It is also the annual venue of the General Assembly of the Church of Scotland. The city has long been a centre of education, particularly in the fields of medicine, Scottish law, literature, philosophy, the sciences and engineering. The University of Edinburgh, founded in 1582 and now one of three in the city, is considered one of the best research institutions in the world. It is the second-largest financial centre in the United Kingdom, the fourth largest in Europe, and the thirteenth largest internationally. The city is a cultural centre, and is the home of institutions including the National Museum of Scotland, the National Library of Scotland and the Scottish National Gallery. The city is also known for the Edinburgh International Festival and the Fringe, the latter being the world's largest annual international arts festival. Historic sites in Edinburgh include Edinburgh Castle, the Palace of Holyroodhouse, the churches of St. Giles, Greyfriars and the Canongate, and the extensive Georgian New Town built in the 18th/19th centuries. Edinburgh's Old Town and New Town together are listed as a UNESCO World Heritage Site, which has been managed by Edinburgh World Heritage since 1999. The city's historical and cultural attractions have made it the UK's second-most visited tourist destination, attracting 4.9 million visits, including 2.4 million from overseas in 2018. Edinburgh is governed by the City of Edinburgh Council, a unitary authority. The City of Edinburgh council area had an estimated population of 526,470 in mid-2021, and includes outlying towns and villages which are not part of Edinburgh proper. The city is in the Lothian region and was historically part of the shire of Midlothian (also called Edinburghshire). "Edin", the root of the city's name, derives from Eidyn, the name for the region in Cumbric, the Brittonic Celtic language formerly spoken there. The name's meaning is unknown. The district of Eidyn was centred on the stronghold of Din Eidyn, the dun or hillfort of Eidyn. This stronghold is believed to have been located at Castle Rock, now the site of Edinburgh Castle. A siege of Din Eidyn by Oswald, king of the Angles of Northumbria in 638 marked the beginning of three centuries of Germanic influence in south east Scotland that laid the foundations for the development of Scots, before the town was ultimately subsumed in 954 by the kingdom known to the English as Scotland. As the language shifted from Cumbric to Northumbrian Old English and then Scots, the Brittonic din in Din Eidyn was replaced by burh, producing Edinburgh. In Scottish Gaelic din becomes dùn, producing modern Dùn Èideann. The city is affectionately nicknamed Auld Reekie, Scots for Old Smoky, for the views from the country of the smoke-covered Old Town. In Walter Scott's 1820 novel The Abbot, a character observes that "yonder stands Auld Reekie—you may see the smoke hover over her at twenty miles' distance". In 1898, Thomas Carlyle comments on the phenomenon: "Smoke cloud hangs over old Edinburgh, for, ever since Aeneas Silvius's time and earlier, the people have the art, very strange to Aeneas, of burning a certain sort of black stones, and Edinburgh with its chimneys is called 'Auld Reekie' by the country people". 19th-century historian Robert Chambers argued that the sobriquet could not be traced before the reign of Charles II in the late 17th century. Instead, he attributed the name to a Fife laird, Durham of Largo, who regulated the bedtime of his children by the smoke rising above Edinburgh from the fires of the tenements. "It's time now bairns, to tak' the beuks, and gang to our beds, for yonder's Auld Reekie, I see, putting on her nicht -cap!". Edinburgh has been popularly called the Athens of the North since the early 19th century. References to Athens, such as Athens of Britain and Modern Athens, had been made as early as the 1760s. The similarities were seen to be topographical but also intellectual. Edinburgh's Castle Rock reminded returning grand tourists of the Athenian Acropolis, as did aspects of the neoclassical architecture and layout of New Town. Both cities had flatter, fertile agricultural land sloping down to a port several miles away (respectively, Leith and Piraeus). Intellectually, the Scottish Enlightenment, with its humanist and rationalist outlook, was influenced by Ancient Greek philosophy. In 1822, artist Hugh William Williams organized an exhibition that showed his paintings of Athens alongside views of Edinburgh, and the idea of a direct parallel between both cities quickly caught the popular imagination. When plans were drawn up in the early 19th century to architecturally develop Calton Hill, the design of the National Monument directly copied Athens' Parthenon. Tom Stoppard's character Archie of Jumpers said, perhaps playing on Reykjavík meaning "smoky bay", that the "Reykjavík of the South" would be more appropriate. The city has also been known by several Latin names, such as Edinburgum, while the adjectival forms Edinburgensis and Edinensis are used in educational and scientific contexts. Edina is a late 18th-century poetical form used by the Scots poets Robert Fergusson and Robert Burns. "Embra" or "Embro" are colloquialisms from the same time, as in Robert Garioch's Embro to the Ploy. Ben Jonson described it as "Britaine's other eye", and Sir Walter Scott referred to it as "yon Empress of the North". Robert Louis Stevenson, also a son of the city, wrote that Edinburgh "is what Paris ought to be". The earliest known human habitation in the Edinburgh area was at Cramond, where evidence was found of a Mesolithic camp site dated to c. 8500 BC. Traces of later Bronze Age and Iron Age settlements have been found on Castle Rock, Arthur's Seat, Craiglockhart Hill and the Pentland Hills. When the Romans arrived in Lothian at the end of the 1st century AD, they found a Brittonic Celtic tribe whose name they recorded as the Votadini. The Votadini transitioned into the Gododdin kingdom in the Early Middle Ages, with Eidyn serving as one of the kingdom's districts. During this period, the Castle Rock site, thought to have been the stronghold of Din Eidyn, emerged as the kingdom's major centre. The medieval poem Y Gododdin describes a war band from across the Brittonic world who gathered in Eidyn before a fateful raid; this may describe a historical event around AD 600. In 638, the Gododdin stronghold was besieged by forces loyal to King Oswald of Northumbria, and around this time control of Lothian passed to the Angles. Their influence continued for the next three centuries until around 950, when, during the reign of Indulf, son of Constantine II, the "burh" (fortress), named in the 10th-century Pictish Chronicle as oppidum Eden, was abandoned to the Scots. It thenceforth remained, for the most part, under their jurisdiction. The royal burgh was founded by King David I in the early 12th century on land belonging to the Crown, though the date of its charter is unknown. The first documentary evidence of the medieval burgh is a royal charter, c. 1124–1127, by King David I granting a toft in burgo meo de Edenesburg to the Priory of Dunfermline. The shire of Edinburgh seems to have also been created in the reign of David I, possibly covering all of Lothian at first, but by 1305 the eastern and western parts of Lothian had become Haddingtonshire and Linlithgowshire, leaving Edinburgh as the county town of a shire covering the central part of Lothian, which was called Edinburghshire or Midlothian (the latter name being an informal, but commonly used, alternative until the county's name was legally changed in 1947). Edinburgh was largely under English control from 1291 to 1314 and from 1333 to 1341, during the Wars of Scottish Independence. When the English invaded Scotland in 1298, Edward I of England chose not to enter Edinburgh but passed by it with his army. In the middle of the 14th century, the French chronicler Jean Froissart described it as the capital of Scotland (c. 1365), and James III (1451–88) referred to it in the 15th century as "the principal burgh of our kingdom". In 1482 James III "granted and perpetually confirmed to the said Provost, Bailies, Clerk, Council, and Community, and their successors, the office of Sheriff within the Burgh for ever, to be exercised by the Provost for the time as Sheriff, and by the Bailies for the time as Sheriffsdepute conjunctly and severally; with full power to hold Courts, to punish transgressors not only by banishment but by death, to appoint officers of Court, and to do everything else appertaining to the office of Sheriff; as also to apply to their own proper use the fines and escheats arising out of the exercise of the said office." Despite being burnt by the English in 1544, Edinburgh continued to develop and grow, and was at the centre of events in the 16th-century Scottish Reformation and 17th-century Wars of the Covenant. In 1582, Edinburgh's town council was given a royal charter by King James VI permitting the establishment of a university; founded as Tounis College (Town's College), the institution developed into the University of Edinburgh, which contributed to Edinburgh's central intellectual role in subsequent centuries. In 1603, King James VI of Scotland succeeded to the English throne, uniting the crowns of Scotland and England in a personal union known as the Union of the Crowns, though Scotland remained, in all other respects, a separate kingdom. In 1638, King Charles I's attempt to introduce Anglican church forms in Scotland encountered stiff Presbyterian opposition culminating in the conflicts of the Wars of the Three Kingdoms. Subsequent Scottish support for Charles Stuart's restoration to the throne of England resulted in Edinburgh's occupation by Oliver Cromwell's Commonwealth of England forces – the New Model Army – in 1650. In the 17th century, Edinburgh's boundaries were still defined by the city's defensive town walls. As a result, the city's growing population was accommodated by increasing the height of the houses. Buildings of 11 storeys or more were common, and have been described as forerunners of the modern-day skyscraper. Most of these old structures were replaced by the predominantly Victorian buildings seen in today's Old Town. In 1611 an act of parliament created the High Constables of Edinburgh to keep order in the city, thought to be the oldest statutory police force in the world. Following the Treaty of Union in 1706, the Parliaments of England and Scotland passed Acts of Union in 1706 and 1707 respectively, uniting the two kingdoms in the Kingdom of Great Britain effective from 1 May 1707. As a consequence, the Parliament of Scotland merged with the Parliament of England to form the Parliament of Great Britain, which sat at Westminster in London. The Union was opposed by many Scots, resulting in riots in the city. By the first half of the 18th century, Edinburgh was described as one of Europe's most densely populated, overcrowded and unsanitary towns. Visitors were struck by the fact that the social classes shared the same urban space, even inhabiting the same tenement buildings; although here a form of social segregation did prevail, whereby shopkeepers and tradesmen tended to occupy the cheaper-to-rent cellars and garrets, while the more well-to-do professional classes occupied the more expensive middle storeys. During the Jacobite rising of 1745, Edinburgh was briefly occupied by the Jacobite "Highland Army" before its march into England. After its eventual defeat at Culloden, there followed a period of reprisals and pacification, largely directed at the rebellious clans. In Edinburgh, the Town Council, keen to emulate London by initiating city improvements and expansion to the north of the castle, reaffirmed its belief in the Union and loyalty to the Hanoverian monarch George III by its choice of names for the streets of the New Town: for example, Rose Street and Thistle Street; and for the royal family, George Street, Queen Street, Hanover Street, Frederick Street and Princes Street (in honour of George's two sons). The consistently geometric layout of the plan for the extension of Edinburgh was the result of a major competition in urban planning staged by the Town Council in 1766. In the second half of the century, the city was at the heart of the Scottish Enlightenment, when thinkers like David Hume, Adam Smith, James Hutton and Joseph Black were familiar figures in its streets. Edinburgh became a major intellectual centre, earning it the nickname "Athens of the North" because of its many neo-classical buildings and reputation for learning, recalling ancient Athens. In the 18th-century novel The Expedition of Humphry Clinker by Tobias Smollett one character describes Edinburgh as a "hotbed of genius". Edinburgh was also a major centre for the Scottish book trade. The highly successful London bookseller Andrew Millar was apprenticed there to James McEuen. From the 1770s onwards, the professional and business classes gradually deserted the Old Town in favour of the more elegant "one-family" residences of the New Town, a migration that changed the city's social character. According to the foremost historian of this development, "Unity of social feeling was one of the most valuable heritages of old Edinburgh, and its disappearance was widely and properly lamented." Despite an enduring myth to the contrary, Edinburgh became an industrial centre with its traditional industries of printing, brewing and distilling continuing to grow in the 19th century and joined by new industries such as rubber works, engineering works and others. By 1821, Edinburgh had been overtaken by Glasgow as Scotland's largest city. The city centre between Princes Street and George Street became a major commercial and shopping district, a development partly stimulated by the arrival of railways in the 1840s. The Old Town became an increasingly dilapidated, overcrowded slum with high mortality rates. Improvements carried out under Lord Provost William Chambers in the 1860s began the transformation of the area into the predominantly Victorian Old Town seen today. More improvements followed in the early 20th century as a result of the work of Patrick Geddes, but relative economic stagnation during the two world wars and beyond saw the Old Town deteriorate further before major slum clearance in the 1960s and 1970s began to reverse the process. University building developments which transformed the George Square and Potterrow areas proved highly controversial. Since the 1990s a new "financial district", including the Edinburgh International Conference Centre, has grown mainly on demolished railway property to the west of the castle, stretching into Fountainbridge, a run-down 19th-century industrial suburb which has undergone radical change since the 1980s with the demise of industrial and brewery premises. This ongoing development has enabled Edinburgh to maintain its place as the United Kingdom's second largest financial and administrative centre after London. Financial services now account for a third of all commercial office space in the city. The development of Edinburgh Park, a new business and technology park covering 38 acres (15 ha), 4 mi (6 km) west of the city centre, has also contributed to the District Council's strategy for the city's major economic regeneration. In 1998, the Scotland Act, which came into force the following year, established a devolved Scottish Parliament and Scottish Executive (renamed the Scottish Government since September 2007). Both based in Edinburgh, they are responsible for governing Scotland while reserved matters such as defence, foreign affairs and some elements of income tax remain the responsibility of the Parliament of the United Kingdom in London. In 2022, Edinburgh was affected by the 2022 Scotland bin strikes. In 2023, Edinburgh became the first capital city in Europe to sign the global Plant Based Treaty, which was introduced at COP26 in 2021 in Glasgow. Green Party councillor Steve Burgess introduced the treaty. The Scottish Countryside Alliance and other farming groups called the treaty "anti-farming." Situated in Scotland's Central Belt, Edinburgh lies on the southern shore of the Firth of Forth. The city centre is 2+1⁄2 mi (4.0 km) southwest of the shoreline of Leith and 26 mi (42 km) inland, as the crow flies, from the east coast of Scotland and the North Sea at Dunbar. While the early burgh grew up near the prominent Castle Rock, the modern city is often said to be built on seven hills, namely Calton Hill, Corstorphine Hill, Craiglockhart Hill, Braid Hill, Blackford Hill, Arthur's Seat and the Castle Rock, giving rise to allusions to the seven hills of Rome. Occupying a narrow gap between the Firth of Forth to the north and the Pentland Hills and their outrunners to the south, the city sprawls over a landscape which is the product of early volcanic activity and later periods of intensive glaciation. Igneous activity between 350 and 400 million years ago, coupled with faulting, led to the creation of tough basalt volcanic plugs, which predominate over much of the area. One such example is the Castle Rock which forced the advancing ice sheet to divide, sheltering the softer rock and forming a 1 mi-long (1.6 km) tail of material to the east, thus creating a distinctive crag and tail formation. Glacial erosion on the north side of the crag gouged a deep valley later filled by the now drained Nor Loch. These features, along with another hollow on the rock's south side, formed an ideal natural strongpoint upon which Edinburgh Castle was built. Similarly, Arthur's Seat is the remains of a volcano dating from the Carboniferous period, which was eroded by a glacier moving west to east during the ice age. Erosive action such as plucking and abrasion exposed the rocky crags to the west before leaving a tail of deposited glacial material swept to the east. This process formed the distinctive Salisbury Crags, a series of teschenite cliffs between Arthur's Seat and the location of the early burgh. The residential areas of Marchmont and Bruntsfield are built along a series of drumlin ridges south of the city centre, which were deposited as the glacier receded. Other prominent landforms such as Calton Hill and Corstorphine Hill are also products of glacial erosion. The Braid Hills and Blackford Hill are a series of small summits to the south of the city centre that command expansive views looking northwards over the urban area to the Firth of Forth. Edinburgh is drained by the river named the Water of Leith, which rises at the Colzium Springs in the Pentland Hills and runs for 18 miles (29 km) through the south and west of the city, emptying into the Firth of Forth at Leith. The nearest the river gets to the city centre is at Dean Village on the north-western edge of the New Town, where a deep gorge is spanned by Thomas Telford's Dean Bridge, built in 1832 for the road to Queensferry. The Water of Leith Walkway is a mixed-use trail that follows the course of the river for 19.6 km (12.2 mi) from Balerno to Leith. Excepting the shoreline of the Firth of Forth, Edinburgh is encircled by a green belt, designated in 1957, which stretches from Dalmeny in the west to Prestongrange in the east. With an average width of 3.2 km (2 mi) the principal objectives of the green belt were to contain the outward expansion of the city and to prevent the agglomeration of urban areas. Expansion affecting the green belt is strictly controlled but developments such as Edinburgh Airport and the Royal Highland Showground at Ingliston lie within the zone. Similarly, suburbs such as Juniper Green and Balerno are situated on green belt land. One feature of the Edinburgh green belt is the inclusion of parcels of land within the city which are designated green belt, even though they do not connect with the peripheral ring. Examples of these independent wedges of green belt include Holyrood Park and Corstorphine Hill. Edinburgh includes former towns and villages that retain much of their original character as settlements in existence before they were absorbed into the expanding city of the nineteenth and twentieth centuries. Many areas, such as Dalry, contain residences that are multi-occupancy buildings known as tenements, although the more southern and western parts of the city have traditionally been less built-up with a greater number of detached and semi-detached villas. The historic centre of Edinburgh is divided in two by the broad green swathe of Princes Street Gardens. To the south, the view is dominated by Edinburgh Castle, built high on Castle Rock, and the long sweep of the Old Town descending towards Holyrood Palace. To the north lie Princes Street and the New Town. The West End includes the financial district, with insurance and banking offices as well as the Edinburgh International Conference Centre. Edinburgh's Old and New Towns were listed as a UNESCO World Heritage Site in 1995 in recognition of the unique character of the Old Town with its medieval street layout and the planned Georgian New Town, including the adjoining Dean Village and Calton Hill areas. There are over 4,500 listed buildings within the city, a higher proportion relative to area than any other city in the United Kingdom. The castle is perched on top of a rocky crag (the remnant of an extinct volcano) and the Royal Mile runs down the crest of a ridge from it terminating at Holyrood Palace. Minor streets (called closes or wynds) lie on either side of the main spine forming a herringbone pattern. Due to space restrictions imposed by the narrowness of this landform, the Old Town became home to some of the earliest "high rise" residential buildings. Multi-storey dwellings known as lands were the norm from the 16th century onwards with ten and eleven storeys being typical and one even reaching fourteen or fifteen storeys. Numerous vaults below street level were inhabited to accommodate the influx of incomers, particularly Irish immigrants, during the Industrial Revolution. The street has several fine public buildings such as St Giles' Cathedral, the City Chambers and the Law Courts. Other places of historical interest nearby are Greyfriars Kirkyard and Mary King's Close. The Grassmarket, running deep below the castle is connected by the steep double terraced Victoria Street. The street layout is typical of the old quarters of many Northern European cities. The New Town was an 18th-century solution to the problem of an increasingly crowded city which had been confined to the ridge sloping down from the castle. In 1766 a competition to design a "New Town" was won by James Craig, a 27-year-old architect. The plan was a rigid, ordered grid, which fitted in well with Enlightenment ideas of rationality. The principal street was to be George Street, running along the natural ridge to the north of what became known as the "Old Town". To either side of it are two other main streets: Princes Street and Queen Street. Princes Street has become Edinburgh's main shopping street and now has few of its Georgian buildings in their original state. The three main streets are connected by a series of streets running perpendicular to them. The east and west ends of George Street are terminated by St Andrew Square and Charlotte Square respectively. The latter, designed by Robert Adam, influenced the architectural style of the New Town into the early 19th century. Bute House, the official residence of the First Minister of Scotland, is on the north side of Charlotte Square. The hollow between the Old and New Towns was formerly the Nor Loch, which was created for the town's defence but came to be used by the inhabitants for dumping their sewage. It was drained by the 1820s as part of the city's northward expansion. Craig's original plan included an ornamental canal on the site of the loch, but this idea was abandoned. Soil excavated while laying the foundations of buildings in the New Town was dumped on the site of the loch to create the slope connecting the Old and New Towns known as The Mound. In the middle of the 19th century the National Gallery of Scotland and Royal Scottish Academy Building were built on The Mound, and tunnels for the railway line between Haymarket and Waverley stations were driven through it. The Southside is a residential part of the city, which includes the districts of St Leonards, Marchmont, Morningside, Newington, Sciennes, the Grange and Blackford. The Southside is broadly analogous to the area covered formerly by the Burgh Muir, and was developed as a residential area after the opening of the South Bridge in the 1780s. The Southside is particularly popular with families (many state and private schools are here), young professionals and students (the central University of Edinburgh campus is based around George Square just north of Marchmont and the Meadows), and Napier University (with major campuses around Merchiston and Morningside). The area is also well provided with hotel and "bed and breakfast" accommodation for visiting festival-goers. These districts often feature in works of fiction. For example, Church Hill in Morningside, was the home of Muriel Spark's Miss Jean Brodie, and Ian Rankin's Inspector Rebus lives in Marchmont and works in St Leonards. Leith was historically the port of Edinburgh, an arrangement of unknown date that was confirmed by the royal charter Robert the Bruce granted to the city in 1329. The port developed a separate identity from Edinburgh, which to some extent it still retains, and it was a matter of great resentment when the two burghs merged in 1920 into the City of Edinburgh. Even today the parliamentary seat is known as "Edinburgh North and Leith". The loss of traditional industries and commerce (the last shipyard closed in 1983) resulted in economic decline. The Edinburgh Waterfront development has transformed old dockland areas from Leith to Granton into residential areas with shopping and leisure facilities and helped rejuvenate the area. With the redevelopment, Edinburgh has gained the business of cruise liner companies which now provide cruises to Norway, Sweden, Denmark, Germany, and the Netherlands. The coastal suburb of Portobello is characterised by Georgian villas, Victorian tenements, a beach and promenade and cafés, bars, restaurants and independent shops. There are rowing and sailing clubs and a restored Victorian swimming pool, including Turkish baths. The urban area of Edinburgh is almost entirely within the City of Edinburgh Council boundary, merging with Musselburgh in East Lothian. Towns within easy reach of the city boundary include Inverkeithing, Haddington, Tranent, Prestonpans, Dalkeith, Bonnyrigg, Loanhead, Penicuik, Broxburn, Livingston and Dunfermline. Edinburgh lies at the heart of the Edinburgh & South East Scotland City region with a population in 2014 of 1,339,380. Like most of Scotland, Edinburgh has a cool, temperate, maritime climate which, despite its northerly latitude, is milder than places which lie at similar latitudes such as Moscow and Labrador. The city's proximity to the sea mitigates any large variations in temperature or extremes of climate. Winter daytime temperatures rarely fall below freezing while summer temperatures are moderate, rarely exceeding 22 °C (72 °F). The highest temperature recorded in the city was 31.6 °C (88.9 °F) on 25 July 2019 at Gogarbank, beating the previous record of 31 °C (88 °F) on 4 August 1975 at Edinburgh Airport. The lowest temperature recorded in recent years was −14.6 °C (5.7 °F) during December 2010 at Gogarbank. Given Edinburgh's position between the coast and hills, it is renowned as "the windy city", with the prevailing wind direction coming from the south-west, which is often associated with warm, unstable air from the North Atlantic Current that can give rise to rainfall – although considerably less than cities to the west, such as Glasgow. Rainfall is distributed fairly evenly throughout the year. Winds from an easterly direction are usually drier but considerably colder, and may be accompanied by haar, a persistent coastal fog. Vigorous Atlantic depressions, known as European windstorms, can affect the city between October and May. Located slightly north of the city centre, the weather station at the Royal Botanic Garden Edinburgh (RBGE) has been an official weather station for the Met Office since 1956. The Met Office operates its own weather station at Gogarbank on the city's western outskirts, near Edinburgh Airport. This slightly inland station has a slightly wider temperature span between seasons, is cloudier and somewhat wetter, but differences are minor. Temperature and rainfall records have been kept at the Royal Observatory since 1764. The most recent official population estimates (2020) are 506,520 for the locality (includes Currie), 530,990 for the Edinburgh settlement (includes Musselburgh). Edinburgh has a high proportion of young adults, with 19.5% of the population in their 20s (exceeded only by Aberdeen) and 15.2% in their 30s which is the highest in Scotland. The proportion of Edinburgh's population born in the UK fell from 92% to 84% between 2001 and 2011, while the proportion of White Scottish-born fell from 78% to 70%. Of those Edinburgh residents born in the UK, 335,000 or 83% were born in Scotland, with 58,000 or 14% being born in England. Some 13,000 people or 2.7% of the city's population are of Polish descent. 39,500 people or 8.2% of Edinburgh's population class themselves as Non-White which is an increase from 4% in 2001. Of the Non-White population, the largest group by far are Asian, totalling 26,264 people. Within the Asian population, people of Chinese descent are now the largest sub-group, with 8,076 people, amounting to about 1.7% of the city's total population. The city's population of Indian descent amounts to 6,470 (1.4% of the total population), while there are some 5,858 of Pakistani descent (1.2% of the total population). Although they account for only 1,277 people or 0.3% of the city's population, Edinburgh has the highest number and proportion of people of Bangladeshi descent in Scotland. Over 7,000 people were born in African countries (1.6% of the total population) and nearly 7,000 in the Americas. With the notable exception of Inner London, Edinburgh has a higher number of people born in the United States (over 3,700) than any other city in the UK. The proportion of people born outside the UK was 15.9% compared with 8% in 2001. A census by the Edinburgh presbytery in 1592 recorded a population of 8,003 adults spread equally north and south of the High Street which runs along the spine of the ridge sloping down from the Castle. In the 18th and 19th centuries, the population expanded rapidly, rising from 49,000 in 1751 to 136,000 in 1831, primarily due to migration from rural areas. As the population grew, problems of overcrowding in the Old Town, particularly in the cramped tenements that lined the present day Royal Mile and the Cowgate, were exacerbated. Poor sanitary arrangements resulted in a high incidence of disease, with outbreaks of cholera occurring in 1832, 1848 and 1866. The construction of the New Town from 1767 onwards witnessed the migration of the professional and business classes from the difficult living conditions in the Old Town to the lower density, higher quality surroundings taking shape on land to the north. Expansion southwards from the Old Town saw more tenements being built in the 19th century, giving rise to Victorian suburbs such as Dalry, Newington, Marchmont and Bruntsfield. Early 20th-century population growth coincided with lower-density suburban development. As the city expanded to the south and west, detached and semi-detached villas with large gardens replaced tenements as the predominant building style. Nonetheless, the 2001 census revealed that over 55% of Edinburgh's population were still living in tenements or blocks of flats, a figure in line with other Scottish cities, but much higher than other British cities, and even central London. From the early to mid 20th century, the growth in population, together with slum clearance in the Old Town and other areas, such as Dumbiedykes, Leith, and Fountainbridge, led to the creation of new estates such as Stenhouse and Saughton, Craigmillar and Niddrie, Pilton and Muirhouse, Piershill, and Sighthill. In 2018, the Church of Scotland had 20,956 members in 71 congregations in the Presbytery of Edinburgh. Its most prominent church is St Giles' on the Royal Mile, first dedicated in 1243 but believed to date from before the 12th century. Saint Giles is historically the patron saint of Edinburgh. St Cuthbert's, situated at the west end of Princes Street Gardens in the shadow of Edinburgh Castle and St Giles' can lay claim to being the oldest Christian sites in the city, though the present St Cuthbert's, designed by Hippolyte Blanc, was dedicated in 1894. Other Church of Scotland churches include Greyfriars Kirk, the Canongate Kirk, St Andrew's and St George's West Church and the Barclay Church. The Church of Scotland Offices are in Edinburgh, as is the Assembly Hall where the annual General Assembly is held. The Roman Catholic Archdiocese of St Andrews and Edinburgh has 27 parishes across the city. The Archbishop of St Andrews and Edinburgh has his official residence in Greenhill, the diocesan offices are in nearby Marchmont, and its cathedral is St Mary's Cathedral, Edinburgh. The Diocese of Edinburgh of the Scottish Episcopal Church has over 50 churches, half of them in the city. Its centre is the late 19th-century Gothic style St Mary's Cathedral in the West End's Palmerston Place. Orthodox Christianity is represented by Pan, Romanian and Russian Orthodox churches. There are several independent churches in the city, both Catholic and Protestant, including Charlotte Chapel, Carrubbers Christian Centre, Bellevue Chapel and Sacred Heart. There are also churches belonging to Quakers, Christadelphians, Seventh-day Adventists, Church of Christ, Scientist, The Church of Jesus Christ of Latter-day Saints (LDS Church) and Elim Pentecostal Church. Muslims have several places of worship across the city. Edinburgh Central Mosque, the largest Islamic place of worship, is located in Potterrow on the city's Southside, near Bristo Square. Construction was largely financed by a gift from King Fahd of Saudi Arabia and was completed in 1998. There is also an Ahmadiyya Muslim community. The first recorded presence of a Jewish community in Edinburgh dates back to the late 18th century. Edinburgh's Orthodox synagogue, opened in 1932, is in Salisbury Road and can accommodate a congregation of 2000. A Liberal Jewish congregation also meets in the city. A Sikh gurdwara and a Hindu mandir are located in Leith. The city also has a Brahma Kumaris centre in the Polwarth area. The Edinburgh Buddhist Centre, run by the Triratna Buddhist Community, formerly situated in Melville Terrace, now runs sessions at the Healthy Life Centre, Bread Street. Other Buddhist traditions are represented by groups which meet in the capital: the Community of Interbeing (followers of Thich Nhat Hanh), Rigpa, Samye Dzong, Theravadin, Pure Land and Shambala. There is a Sōtō Zen Priory in Portobello and a Theravadin Thai Buddhist Monastery in Slateford Road. Edinburgh is home to a Baháʼí community, and a Theosophical Society meets in Great King Street. Edinburgh has an Inter-Faith Association. Edinburgh has over 39 graveyards and cemeteries, many of which are listed and of historical character, including several former church burial grounds. Examples include Old Calton Burial Ground, Greyfriars Kirkyard and Dean Cemetery. Edinburgh has the strongest economy of any city in the United Kingdom outside London and the highest percentage of professionals in the UK with 43% of the population holding a degree-level or professional qualification. According to the Centre for International Competitiveness, it is the most competitive large city in the United Kingdom. It also has the highest gross value added per employee of any city in the UK outside London, measuring £57,594 in 2010. It was named European Best Large City of the Future for Foreign Direct Investment and Best Large City for Foreign Direct Investment Strategy in the Financial Times fDi magazine awards 2012/13. In the 19th century, Edinburgh's economy was known for banking and insurance, publishing and printing, and brewing and distilling. Today, its economy is based mainly on financial services, scientific research, higher education, and tourism. In March 2010, unemployment in Edinburgh was comparatively low at 3.6%, and it remains consistently below the Scottish average of 4.5%. Edinburgh is the second most visited city by foreign visitors in the UK after London. Banking has been a mainstay of the Edinburgh economy for over 300 years, since the Bank of Scotland was established by an act of the Scottish Parliament in 1695. Today, the financial services industry, with its particularly strong insurance and investment sectors, and underpinned by Edinburgh-based firms such as Scottish Widows and Standard Life Aberdeen, accounts for the city being the UK's second financial centre after London and Europe's fourth in terms of equity assets. The NatWest Group (formerly Royal Bank of Scotland Group) opened new global headquarters at Gogarburn in the west of the city in October 2005. The city is home to the headquarters of Bank of Scotland, Sainsbury's Bank, Tesco Bank, and TSB Bank. Tourism is also an important element in the city's economy. As a World Heritage Site, tourists visit historical sites such as Edinburgh Castle, the Palace of Holyroodhouse and the Old and New Towns. Their numbers are augmented in August each year during the Edinburgh Festivals, which attracts 4.4 million visitors, and generates over £100M for the local economy. As the centre of Scotland's government and legal system, the public sector plays a central role in Edinburgh's economy. Many departments of the Scottish Government are in the city. Other major employers include NHS Scotland and local government administration. When the £1.3bn Edinburgh & South East Scotland City Region Deal was signed in 2018, the region's Gross Value Added (GVA) contribution to the Scottish economy was cited as £33bn, or 33% of the country's output. The City Region Deal funds a range of "Data Driven Innovation" hubs which are using data to innovate in the region, recognising the region's strengths in technology and data science, the growing importance of the data economy, and the need to tackle the digital skills gap, as a route to social and economic prosperity. The city hosts a series of festivals that run between the end of July and early September each year. The best known of these events are the Edinburgh Festival Fringe, the Edinburgh International Festival, the Edinburgh Military Tattoo, the Edinburgh Art Festival and the Edinburgh International Book Festival. The longest established of these festivals is the Edinburgh International Festival, which was first held in 1947 and consists mainly of a programme of high-profile theatre productions and classical music performances, featuring international directors, conductors, theatre companies and orchestras. This has since been overtaken in size by the Edinburgh Fringe which began as a programme of marginal acts alongside the "official" Festival and has become the world's largest performing arts festival. In 2017, nearly 3400 different shows were staged in 300 venues across the city. Comedy has become one of the mainstays of the Fringe, with numerous well-known comedians getting their first 'break' there, often by being chosen to receive the Edinburgh Comedy Award. The Edinburgh Military Tattoo, occupies the Castle Esplanade every night for three weeks each August, with massed pipe bands and military bands drawn from around the world. Performances end with a short fireworks display. As well as the summer festivals, many other festivals are held during the rest of the year, including the Edinburgh International Film Festival and Edinburgh International Science Festival. The summer of 2020 was the first time in its 70-year history that the Edinburgh festival was not run, being cancelled due to the COVID-19 pandemic. This affected many of the tourist-focused businesses in Edinburgh which depend on the various festivals over summer to return an annual profit. The annual Edinburgh Hogmanay celebration was originally an informal street party focused on the Tron Kirk in the Old Town's High Street. Since 1993, it has been officially organised with the focus moved to Princes Street. In 1996, over 300,000 people attended, leading to ticketing of the main street party in later years up to a limit of 100,000 tickets. Hogmanay now covers four days of processions, concerts and fireworks, with the street party beginning on Hogmanay. Alternative tickets are available for entrance into the Princes Street Gardens concert and Cèilidh, where well-known artists perform and ticket holders can participate in traditional Scottish cèilidh dancing. The event attracts thousands of people from all over the world. On the night of 30 April the Beltane Fire Festival takes place on Calton Hill, involving a procession followed by scenes inspired by pagan old spring fertility celebrations. At the beginning of October each year the Dussehra Hindu Festival is also held on Calton Hill. Outside the Festival season, Edinburgh supports several theatres and production companies. The Royal Lyceum Theatre has its own company, while the King's Theatre, Edinburgh Festival Theatre and Edinburgh Playhouse stage large touring shows. The Traverse Theatre presents a more contemporary repertoire. Amateur theatre companies productions are staged at the Bedlam Theatre, Church Hill Theatre and King's Theatre among others. The Usher Hall is Edinburgh's premier venue for classical music, as well as occasional popular music concerts. It was the venue for the Eurovision Song Contest 1972. Other halls staging music and theatre include The Hub, the Assembly Rooms and the Queen's Hall. The Scottish Chamber Orchestra is based in Edinburgh. Edinburgh has one repertory cinema, The Cameo, and formerly, the Edinburgh Filmhouse as well as the independent Dominion Cinema and a range of multiplexes. Edinburgh has a healthy popular music scene. Occasionally large concerts are staged at Murrayfield and Meadowbank, while mid-sized events take place at smaller venues such as 'The Corn Exchange', 'The Liquid Rooms' and 'The Bongo Club'. In 2010, PRS for Music listed Edinburgh among the UK's top ten 'most musical' cities. Several city pubs are well known for their live performances of folk music. They include 'Sandy Bell's' in Forrest Road, 'Captain's Bar' in South College Street and 'Whistlebinkies' in South Bridge. Like many other cities in the UK, numerous nightclub venues host Electronic dance music events. Edinburgh is home to a flourishing group of contemporary composers such as Nigel Osborne, Peter Nelson, Lyell Cresswell, Hafliði Hallgrímsson, Edward Harper, Robert Crawford, Robert Dow and John McLeod. McLeod's music is heard regularly on BBC Radio 3 and throughout the UK. The main local newspaper is the Edinburgh Evening News. It is owned and published alongside its sister titles The Scotsman and Scotland on Sunday by JPIMedia. The city has two commercial radio stations: Forth 1, a station which broadcasts mainstream chart music, and Forth 2 on medium wave which plays classic hits. Capital Scotland, Heart Scotland and Eklipse Sports Radio also have transmitters covering Edinburgh. Along with the UK national radio stations, BBC Radio Scotland and the Gaelic language service BBC Radio nan Gàidheal are also broadcast. DAB digital radio is broadcast over two local multiplexes. BFBS Radio broadcasts from studios on the base at Dreghorn Barracks across the city on 98.5FM as part of its UK Bases network. Small scale DAB started October 2022 with numerous community stations onboard Television, along with most radio services, is broadcast to the city from the Craigkelly transmitting station situated in Fife on the opposite side of the Firth of Forth and the Black Hill transmitting station in North Lanarkshire to the west. There are no television stations based in the city. Edinburgh Television existed in the late 1990s to early 2003 and STV Edinburgh existed from 2015 to 2018. Edinburgh has many museums and libraries. These include the National Museum of Scotland, the National Library of Scotland, National War Museum, the Museum of Edinburgh, Surgeons' Hall Museum, the Writers' Museum, the Museum of Childhood and Dynamic Earth. The Museum on The Mound has exhibits on money and banking. Edinburgh Zoo, covering 82 acres (33 ha) on Corstorphine Hill, is the second most visited paid tourist attraction in Scotland, and home to two giant pandas, Tian Tian and Yang Guang, on loan from the People's Republic of China. Edinburgh is also home to The Royal Yacht Britannia, decommissioned in 1997 and now a five-star visitor attraction and evening events venue permanently berthed at Ocean Terminal. Edinburgh contains Scotland's three National Galleries of Art as well as numerous smaller art galleries. The national collection is housed in the Scottish National Gallery, located on The Mound, comprising the linked National Gallery of Scotland building and the Royal Scottish Academy building. Contemporary collections are shown in the Scottish National Gallery of Modern Art which occupies a split site at Belford. The Scottish National Portrait Gallery on Queen Street focuses on portraits and photography. The council-owned City Art Centre in Market Street mounts regular art exhibitions. Across the road, The Fruitmarket Gallery offers world-class exhibitions of contemporary art, featuring work by British and international artists with both emerging and established international reputations. The city hosts several of Scotland's galleries and organisations dedicated to contemporary visual art. Significant strands of this infrastructure include Creative Scotland, Edinburgh College of Art, Talbot Rice Gallery (University of Edinburgh), Collective Gallery (based at the City Observatory) and the Edinburgh Annuale. There are also many small private shops/galleries that provide space to showcase works from local artists. The locale around Princes Street is the main shopping area in the city centre, with souvenir shops, chain stores such as Boots the Chemist, Edinburgh Woollen Mill, and H&M. George Street, north of Princes Street, has several upmarket shops and independent stores. At the east end of Princes Street, the redeveloped St James Quarter opened its doors in June 2021, while next to the Balmoral Hotel and Waverley Station is Waverley Market. Multrees Walk is a pedestrian shopping district, dominated by the presence of Harvey Nichols, and other names including Louis Vuitton, Mulberry and Michael Kors. Edinburgh also has substantial retail parks outside the city centre. These include The Gyle Shopping Centre and Hermiston Gait in the west of the city, Cameron Toll Shopping Centre, Straiton Retail Park (actually just outside the city, in Midlothian) and Fort Kinnaird in the south and east, and Ocean Terminal in the north on the Leith waterfront. Following local government reorganisation in 1996, the City of Edinburgh Council constitutes one of the 32 council areas of Scotland. Like all other local authorities of Scotland, the council has powers over most matters of local administration such as housing, planning, local transport, parks, economic development and regeneration. The council comprises 63 elected councillors, returned from 17 multi-member electoral wards in the city. Following the 2007 City of Edinburgh Council election the incumbent Labour Party lost majority control of the council after 23 years to a Liberal Democrat/SNP coalition. After the 2017 election, the SNP and Labour formed a coalition administration, which lasted until the next election in 2022. The 2022 City of Edinburgh Council election resulted in the most politically balanced council in the UK, with 19 SNP, 13 Labour, 12 Liberal Democrat, 10 Green and 9 Conservative councillors. A minority Labour administration was formed, being voted in by Scottish Conservative and Scottish Liberal Democrat councillors. The SNP and Greens presented a coalition agreement, but could not command majority support in the Council. This caused controversy amongst the Scottish Labour Party group for forming an administration supported by Conservatives and led to the suspension of two Labour councillors on the Council for abstaining on the vote to approve the new administration. The city's coat of arms was registered by the Lord Lyon King of Arms in 1732. Edinburgh, like all of Scotland, is represented in the Scottish Parliament, situated in the Holyrood area of the city. For electoral purposes, the city is divided into six constituencies which, along with 3 seats outside of the city, form part of the Lothian region. Each constituency elects one Member of the Scottish Parliament (MSP) by the first past the post system of election, and the region elects seven additional MSPs to produce a result based on a form of proportional representation. As of the 2021 election, the Scottish National Party have four MSPs: Ash Denham for Edinburgh Eastern, Ben Macpherson for Edinburgh Northern and Leith and Gordon MacDonald for Edinburgh Pentlands and Angus Robertson for Edinburgh Central constituencies. Alex Cole-Hamilton, the Leader of the Scottish Liberal Democrats represents Edinburgh Western and Daniel Johnson of the Scottish Labour Party represents Edinburgh Southern constituency. In addition, the city is also represented by seven regional MSPs representing the Lothian electoral region: The Conservatives have three regional MSPs: Jeremy Balfour, Miles Briggs and Sue Webber, Labour have two regional MSPs: Sarah Boyack and Foysol Choudhury; two Scottish Green regional MSPs were elected: Green's Co-Leader Lorna Slater and Alison Johnstone. However, following her election as the Presiding Officer of the 6th Session of the Scottish Parliament on 13 May 2021, Alison Johnstone has abided by the established parliamentary convention for speakers and renounced all affiliation with her former political party for the duration of her term as Presiding Officer. So she presently sits as an independent MSP for the Lothians Region. Edinburgh is also represented in the House of Commons of the United Kingdom by five Members of Parliament. The city is divided into Edinburgh North and Leith, Edinburgh East, Edinburgh South, Edinburgh South West, and Edinburgh West, each constituency electing one member by the first past the post system. Since the 2019 UK General election, Edinburgh is represented by three Scottish National Party MPs (Deirdre Brock, Edinburgh North and Leith/Tommy Sheppard, Edinburgh East/Joanna Cherry, Edinburgh South West), one Liberal Democrat MP in Edinburgh West (Christine Jardine) and one Labour MP in Edinburgh South (Ian Murray). Edinburgh Airport is Scotland's busiest airport and the principal international gateway to the capital, handling over 14.7 million passengers; it was also the sixth-busiest airport in the United Kingdom by total passengers in 2019. In anticipation of rising passenger numbers, the former operator of the airport BAA outlined a draft masterplan in 2011 to provide for the expansion of the airfield and the terminal building. In June 2012, Global Infrastructure Partners purchased the airport for £807 million. The possibility of building a second runway to cope with an increased number of aircraft movements has also been mooted. Travel in Edinburgh is undertaken predominantly by bus. Lothian Buses, the successor company to Edinburgh Corporation Transport Department, operate the majority of city bus services within the city and to surrounding suburbs, with the most routes running via Princes Street. Services further afield operate from the Edinburgh Bus Station off St Andrew Square and Waterloo Place and are operated mainly by Stagecoach East Scotland, Scottish Citylink, National Express Coaches and Borders Buses. Lothian Buses and McGill's Scotland East operate the city's branded public tour buses. The night bus service and airport buses are mainly operated by Lothian Buses link. In 2019, Lothian Buses recorded 124.2 million passenger journeys. To tackle traffic congestion, Edinburgh is now served by six park & ride sites on the periphery of the city at Sheriffhall (in Midlothian), Ingliston, Riccarton, Inverkeithing (in Fife), Newcraighall and Straiton (in Midlothian). A referendum of Edinburgh residents in February 2005 rejected a proposal to introduce congestion charging in the city. Edinburgh Waverley is the second-busiest railway station in Scotland, with only Glasgow Central handling more passengers. On the evidence of passenger entries and exits between April 2015 and March 2016, Edinburgh Waverley is the fifth-busiest station outside London; it is also the UK's second biggest station in terms of the number of platforms and area size. Waverley is the terminus for most trains arriving from London King's Cross and the departure point for many rail services within Scotland operated by ScotRail. To the west of the city centre lies Haymarket station, which is an important commuter stop. Opened in 2003, Edinburgh Park station serves the Gyle business park in the west of the city and the nearby Gogarburn headquarters of the Royal Bank of Scotland. The Edinburgh Crossrail route connects Edinburgh Park with Haymarket, Edinburgh Waverley and the suburban stations of Brunstane and Newcraighall in the east of the city. There are also commuter lines to Edinburgh Gateway, South Gyle and Dalmeny, the latter serving South Queensferry by the Forth Bridges, and to Wester Hailes and Curriehill in the south-west of the city. Edinburgh Trams became operational on 31 May 2014. The city had been without a tram system since Edinburgh Corporation Tramways ceased on 16 November 1956. Following parliamentary approval in 2007, construction began in early 2008. The first stage of the project was expected to be completed by July 2011 but, following delays caused by extra utility work and a long-running contractual dispute between the council and the main contractor, Bilfinger SE, the project was rescheduled. The line opened in 2014 but had been cut short to 8.7 mi (14.0 km) in length, running from Edinburgh Airport To York Place in the east end of the city. The line was later extended north onto Leith and Newhaven opening a further eight stops to passengers in June 2023. The York Place stop was replaced by a new island stop at Picardy Place. The original plan would have seen a second line run from Haymarket through Ravelston and Craigleith to Granton Square on the Waterfront Edinburgh.This was shelved in 2011 but is now once again under consideration, as is another line potentially linking the south of the city and the Bioquarter. There were also long-term plans for lines running west from the airport to Ratho and Newbridge and another connecting Granton to Newhaven via Lower Granton Road Lothian Buses and Edinburgh Trams are both owned and operated by Transport for Edinburgh. Despite its modern transport links, in January of 2021 Edinburgh was named the most congested city in the UK for the fourth year running, though has since fallen to 7th place in 2022 There are three universities in Edinburgh: the University of Edinburgh, Heriot-Watt University and Edinburgh Napier University. Established by royal charter in 1583, the University of Edinburgh is one of Scotland's ancient universities and is the fourth oldest in the country after St Andrews, Glasgow and Aberdeen. Originally centred on Old College the university expanded to premises on The Mound, the Royal Mile and George Square. Today, the King's Buildings in the south of the city contain most of the schools within the College of Science and Engineering. In 2002, the medical school moved to purpose built accommodation adjacent to the new Royal Infirmary of Edinburgh at Little France. The university is placed 16th in the QS World University Rankings for 2022. Heriot-Watt University is based at the Riccarton campus in the west of Edinburgh. Originally established in 1821, as the world's first mechanics' institute, it was granted university status by royal charter in 1966. It has other campuses in the Scottish Borders, Orkney, United Arab Emirates and Putrajaya in Malaysia. It takes the name Heriot-Watt from Scottish inventor James Watt and Scottish philanthropist and goldsmith George Heriot. Heriot-Watt University has been named International University of the Year by The Times and Sunday Times Good University Guide 2018. In the latest Research Excellence Framework, it was ranked overall in the Top 25% of UK universities and 1st in Scotland for research impact. Edinburgh Napier University was originally founded as the Napier College, which was renamed Napier Polytechnic in 1986 and gained university status in 1992. Edinburgh Napier University has campuses in the south and west of the city, including the former Merchiston Tower and Craiglockhart Hydropathic. It is home to the Screen Academy Scotland. Queen Margaret University was located in Edinburgh before it moved to a new campus just outside the city boundary on the edge of Musselburgh in 2008. Until 2012, further education colleges in the city included Jewel and Esk College (incorporating Leith Nautical College founded in 1903), Telford College, opened in 1968, and Stevenson College, opened in 1970. These have now been amalgamated to form Edinburgh College. Scotland's Rural College also has a campus in south Edinburgh. Other institutions include the Royal College of Surgeons of Edinburgh and the Royal College of Physicians of Edinburgh which were established by royal charter in 1506 and 1681 respectively. The Trustees Drawing Academy of Edinburgh, founded in 1760, became the Edinburgh College of Art in 1907. There are 18 nursery, 94 primary and 23 secondary schools administered by the City of Edinburgh Council. Edinburgh is home to The Royal High School, one of the oldest schools in the country and the world. The city also has several independent, fee-paying schools including Edinburgh Academy, Fettes College, George Heriot's School, George Watson's College, Merchiston Castle School, Stewart's Melville College and The Mary Erskine School. In 2009, the proportion of pupils attending independent schools was 24.2%, far above the Scottish national average of just over 7% and higher than in any other region of Scotland. In August 2013, the City of Edinburgh Council opened the city's first stand-alone Gaelic primary school, Bun-sgoil Taobh na Pàirce. The main NHS Lothian hospitals serving the Edinburgh area are the Royal Infirmary of Edinburgh, which includes the University of Edinburgh Medical School, and the Western General Hospital, which has a large cancer treatment centre and nurse-led Minor Injuries Clinic. The Royal Edinburgh Hospital in Morningside specialises in mental health. The Royal Hospital for Children and Young People, colloquially referred to as the Sick Kids, is a specialist paediatrics hospital. There are two private hospitals: Murrayfield Hospital in the west of the city and Shawfair Hospital in the south; both are owned by Spire Healthcare. Edinburgh has three football clubs that play in the Scottish Professional Football League (SPFL): Heart of Midlothian, founded in 1874, Hibernian, founded in 1875 and Edinburgh City F.C., founded in 1966. Heart of Midlothian and Hibernian are known locally as "Hearts" and "Hibs", respectively. Both play in the Scottish Premiership. They are the oldest city rivals in Scotland and the Edinburgh derby is one of the oldest derby matches in world football. Both clubs have won the Scottish league championship four times. Hearts have won the Scottish Cup eight times and the Scottish League Cup four times. Hibs have won the Scottish Cup and the Scottish League Cup three times each. Edinburgh City were promoted to Scottish League Two in the 2015–16 season, becoming the first club to win promotion to the SPFL via the pyramid system playoffs. Edinburgh was also home to four other former Scottish Football League clubs: the original Edinburgh City (founded in 1928), Leith Athletic, Meadowbank Thistle and St Bernard's. Meadowbank Thistle played at Meadowbank Stadium until 1995, when the club moved to Livingston and became Livingston F.C. The Scottish national team has very occasionally played at Easter Road and Tynecastle, although its normal home stadium is Hampden Park in Glasgow. St Bernard's' New Logie Green was used to host the 1896 Scottish Cup Final, the only time the match has been played outside Glasgow. The city also plays host to Lowland Football League clubs Civil Service Strollers, Edinburgh University and Spartans, as well as East of Scotland League clubs Craigroyston, Edinburgh United, Heriot-Watt University, Leith Athletic, Lothian Thistle Hutchison Vale, and Tynecastle. In women's football, Hearts, Hibs and Spartans play in the SWPL 1. Hutchison Vale and Boroughmuir Thistle play in the SWPL 2. The Scotland national rugby union team play at Murrayfield Stadium, and the professional Edinburgh Rugby team play at the nextdoor Edinburgh Rugby Stadium; both are owned by the Scottish Rugby Union and are also used for other events, including music concerts. Murrayfield is the largest capacity stadium in Scotland, seating 67,144 spectators. Edinburgh is also home to Scottish Premiership teams Boroughmuir RFC, Currie RFC, the Edinburgh Academicals, Heriot's Rugby Club and Watsonians RFC. The Edinburgh Academicals ground at Raeburn Place was the location of the world's first international rugby game on 27 March 1871, between Scotland and England. Rugby league is represented by the Edinburgh Eagles who play in the Rugby League Conference Scotland Division. Murrayfield Stadium has hosted the Magic Weekend where all Super League matches are played in the stadium over one weekend. The Scottish cricket team, which represents Scotland internationally, play their home matches at the Grange cricket club. The Edinburgh Capitals are the latest of a succession of ice hockey clubs in the Scottish capital. Previously Edinburgh was represented by the Murrayfield Racers (2018), the original Murrayfield Racers (who folded in 1996) and the Edinburgh Racers. The club play their home games at the Murrayfield Ice Rink and have competed in the eleven-team professional Scottish National League (SNL) since the 2018–19 season. Next door to Murrayfield Ice Rink is a 7-sheeter dedicated curling facility where curling is played from October to March each season. Caledonia Pride are the only women's professional basketball team in Scotland. Established in 2016, the team compete in the UK wide Women's British Basketball League and play their home matches at the Oriam National Performance Centre. Edinburgh also has several men's basketball teams within the Scottish National League. Boroughmuir Blaze, City of Edinburgh Kings and Edinburgh Lions all compete in Division 1 of the National League, and Pleasance B.C. compete in Division 2. The Edinburgh Diamond Devils is a baseball club which won its first Scottish Championship in 1991 as the "Reivers." 1992 saw the team repeat the achievement, becoming the first team to do so in league history. The same year saw the start of their first youth team, the Blue Jays. The club adopted its present name in 1999. Edinburgh has also hosted national and international sports events including the World Student Games, the 1970 British Commonwealth Games, the 1986 Commonwealth Games and the inaugural 2000 Commonwealth Youth Games. For the 1970 Games the city built Olympic standard venues and facilities including Meadowbank Stadium and the Royal Commonwealth Pool. The Pool underwent refurbishment in 2012 and hosted the Diving competition in the 2014 Commonwealth Games which were held in Glasgow. In American football, the Scottish Claymores played WLAF/NFL Europe games at Murrayfield, including their World Bowl 96 victory. From 1995 to 1997 they played all their games there, from 1998 to 2000 they split their home matches between Murrayfield and Glasgow's Hampden Park, then moved to Glasgow full-time, with one final Murrayfield appearance in 2002. The city's most successful non-professional team are the Edinburgh Wolves who play at Meadowbank Stadium. The Edinburgh Marathon has been held annually in the city since 2003 with more than 16,000 runners taking part on each occasion. Its organisers have called it "the fastest marathon in the UK" due to the elevation drop of 40 m (130 ft). The city also organises a half-marathon, as well as 10 km (6.2 mi) and 5 km (3.1 mi) races, including a 5 km (3 mi) race on 1 January each year. Edinburgh has a speedway team, the Edinburgh Monarchs, which, since the loss of its stadium in the city, has raced at the Lothian Arena in Armadale, West Lothian. The Monarchs have won the Premier League championship five times in their history, in 2003 and again in 2008, 2010, 2014 and 2015. For basketball, the city has a basketball club, Edinburgh Tigers. Edinburgh has a long literary tradition, which became especially evident during the Scottish Enlightenment. This heritage and the city's lively literary life in the present led to it being declared the first UNESCO City of Literature in 2004. Prominent authors who have lived in Edinburgh include the economist Adam Smith, born in Kirkcaldy and author of The Wealth of Nations, James Boswell, biographer of Samuel Johnson; Sir Walter Scott, creator of the historical novel and author of works such as Rob Roy, Ivanhoe, and Heart of Midlothian; James Hogg, author of The Private Memoirs and Confessions of a Justified Sinner; Robert Louis Stevenson, creator of Treasure Island, Kidnapped, and Strange Case of Dr Jekyll and Mr Hyde; Sir Arthur Conan Doyle, the creator of Sherlock Holmes; Muriel Spark, author of The Prime of Miss Jean Brodie; diarist Janet Harden; Irvine Welsh, author of Trainspotting, whose novels are mostly set in the city and often written in colloquial Scots; Ian Rankin, author of the Inspector Rebus series of crime thrillers, Alexander McCall Smith, author of the No. 1 Ladies' Detective Agency series, and J. K. Rowling, author of Harry Potter, who wrote much of her first book in Edinburgh coffee shops and now lives in the Cramond area of the city. Scotland has a rich history of science and engineering, with Edinburgh producing a number of leading figures. John Napier, inventor of logarithms, was born in Merchiston Tower and lived and died in the city. His house now forms part of the original campus of Napier University which was named in his honour. He lies buried under St. Cuthbert's Church. James Clerk Maxwell, founder of the modern theory of electromagnetism, was born at 14 India Street (now the home of the James Clerk Maxwell Foundation) and educated at the Edinburgh Academy and the University of Edinburgh, as was the engineer and telephone pioneer Alexander Graham Bell. James Braidwood, who organised Britain's first municipal fire brigade, was also born in the city and began his career there. Other names connected with the city include physicist Max Born, a principle founder of Quantum mechanics and Nobel laureate; Charles Darwin, the biologist who propounded the theory of natural selection; David Hume, philosopher, economist and historian; James Hutton, regarded as the "Father of Geology"; Joseph Black, the chemist who discovered Magnesium and Carbon Dioxide, and one of the founders of Thermodynamics; pioneering medical researchers Joseph Lister and James Young Simpson; chemist and discoverer of the element nitrogen Daniel Rutherford; Colin Maclaurin, mathematician and developer of the Maclaurin series, and Ian Wilmut, the geneticist involved in the cloning of Dolly the sheep just outside Edinburgh, at the Roslin Institute. The stuffed carcass of Dolly the sheep is now on display in the National Museum of Scotland. The latest in a long line of science celebrities associated with the city is theoretical physicist, Nobel laureate and professor emeritus at the University of Edinburgh Peter Higgs, born in Newcastle but resident in Edinburgh for most of his academic career, after whom the Higgs boson particle has been named. Edinburgh has been the birthplace of actors like Alastair Sim and Sir Sean Connery, known for being the first cinematic James Bond, the comedian and actor Ronnie Corbett, best known as one of The Two Ronnies, and the impressionist Rory Bremner. Famous artists from the city include the portrait painters Sir Henry Raeburn, Sir David Wilkie and Allan Ramsay. The city has produced or been home to some very successful musicians in recent decades, particularly Ian Anderson, front man of the band Jethro Tull, The Incredible String Band, the folk duo The Corries, Wattie Buchan, lead singer and founding member of punk band The Exploited, Shirley Manson, lead singer of the band Garbage, the Bay City Rollers, The Proclaimers, Boards of Canada and Idlewild. Edinburgh is the birthplace of former British Prime Minister Tony Blair who attended the city's Fettes College. Notorious criminals from Edinburgh's past include Deacon Brodie, head of a trades guild and Edinburgh city councillor by day but a burglar by night, who is said to have been the inspiration for Robert Louis Stevenson's story, the Strange Case of Dr Jekyll and Mr Hyde, and murderers Burke and Hare who delivered fresh corpses for dissection to the famous anatomist Robert Knox. Another well-known Edinburgh resident was Greyfriars Bobby. The small Skye Terrier reputedly kept vigil over his dead master's grave in Greyfriars Kirkyard for 14 years in the 1860s and 1870s, giving rise to a story of canine devotion which plays a part in attracting visitors to the city. The City of Edinburgh has entered into 14 international twinning arrangements since 1954. Most of the arrangements are styled as Twin Cities but the agreement with Kraków is designated as a Partner City, and the agreement with Kyoto Prefecture is officially styled as a Friendship Link, reflecting its status as the only region to be twinned with Edinburgh. For a list of consulates in Edinburgh, see List of diplomatic missions in Scotland.
{{multiple image | perrow = 3 | total_width = 300 | image1 = Emmanuel_Macron_in_2019.jpg | link1 = Emmanuel Macron | image2 = Prabowo Subianto 2024 official portrait.jpg | link2 = Prabowo Subianto | image3 = Joe Biden presidential portrait (cropped).jpg | link3 = Joe Biden | image4 = 윤석열 검찰총장.png | link4 = Yoon Suk-yeol | image5 = Charles, Prince of Wales in 2021 (cropped) (3).jpg | link5 = Charles III | image6 = Lula da Silva zu Gast bei der LINKEN - 49643838032 (cropped).jpg | link6 = Luiz Inácio Lula da Silva | image7 = Bola A Tinubu presidential portrait 2023.jpg | link7 = Bola Tinubu | image8 = Tharman Shanmugaratnam Finance Minister of Singapore - (4421243960) (cropped).jpg | link8 = Tharman Shanmugaratnam | image9 = Vladimir Putin (2017-07-08) (cropped).jpg | link9 = Vladimir Putin | image10 = Rodrigo Duterte meets with Salman of Saudi Arabia (2017-04-11) (cropped).jpg | link10 = Salman of Saudi Arabia | image11 = Emperor Naruhito at TICAD7 (cropped) (3).jpg | link11 = Naruhito | image12 = Ferdinand Marcos Jr. Inauguration RVTM.jpg | link12 = Bongbong Marcos | Image13 = Felipe VI 2023.jpg | Link 13 = Felipe VI | footer = Heads of state of various countries: * [[Emmanuel Macron]], [[President of France]] * [[Prabowo Subianto]], [[President of Indonesia]] * [[Joe Biden]], [[President of the United States]] ---- * [[Yoon Suk-yeol]], [[President of South Korea]] * [[Charles III]], [[Monarchy of the United Kingdom|King of the United Kingdom]] * [[Luiz Inácio Lula da Silva]], [[President of Brazil]] ---- * [[Bola Tinubu]], [[President of Nigeria]] * [[Tharman Shanmugaratnam]], [[President of Singapore]] * [[Vladimir Putin]], [[President of Russia]] ---- * [[Salman of Saudi Arabia|Salman]], [[King of Saudi Arabia]] * [[Naruhito]], [[Emperor of Japan]] * [[Bongbong Marcos]], [[President of the Philippines]] * [[Felipe VI of Spain |Felipe VI]], [[Monarchy of Spain|King of Spain]] | align = | direction = | alt1 = | caption1 = | caption2 = | width = 300 }} The '''head of state''' is the leader of a [[country]]/[[sovereign state]]. The powers of a head of state may vary depending on the country. The head of state is usually the highest-ranking official in the country. == Titles == In most [[Republic|republic countries]], the head of state is called the [[president]] and is usually elected by the people. In countries that are [[Monarchy|monarchies]], the head of state is a [[monarch]] (such as a [[king]]), who gets the position through [[inheritance]]. There are many titles that a monarch could have, but they all have the same role as a head of state. == Types of heads of state == === Executive === Sometimes a head of state has an [[Executive (government)|executive role]]. This means that they take part in leading the [[government]]. In these systems, the head of state is also the [[head of government]]. For example, the [[President of the United States]] is both the head of state and head of government of the [[United States|US]]. There are many republics that operate like this but some monarchies still have a king who runs the government. === Ceremonial === In other countries, the head of state may be a separate person from the [[head of government]]. In this system, the head of state has the role of representing the [[nation]] and being a neutral non-political leader while the head of government takes care of the executive. This system is used in many modern monarchies such as the [[United Kingdom]]. However, some republics such as the [[Republic of Ireland]] follow this system.<br /> [[Category:Heads of state| ]]
Executive head (presidential system) Executive head (semi-presidential system) Executive head (parliamentary system) Ceremonial head (parliamentary system) Constitutional monarch (parliamentary system) A head of state (or chief of state) is the public persona who officially embodies a sovereign state in its unity and legitimacy. Depending on the country's form of government and separation of powers, the head of state may be a ceremonial figurehead or concurrently the head of government and more. In a parliamentary system, such as the United Kingdom or India, the head of state usually has mostly ceremonial powers, with a separate head of government. However, in some parliamentary systems, like South Africa, there is an executive president that is both head of state and head of government. Likewise, in some parliamentary systems the head of state is not the head of government, but still has significant powers, for example Morocco. In contrast, a semi-presidential system, such as France, has both heads of state and government as the de facto leaders of the nation (in practice they divide the leadership of the nation between themselves). Meanwhile, in presidential systems, the head of state is also the head of government. In one-party ruling communist states, the position of president has no tangible powers by itself, however, since such a head of state, as a matter of custom, simultaneously holds the post of General Secretary of the Communist Party, they are the executive leader with their powers deriving from their status of being the party leader, rather than the office of president. Former French president Charles de Gaulle, while developing the current Constitution of France (1958), said that the head of state should embody l'esprit de la nation ("the spirit of the nation"). Some academic writers discuss states and governments in terms of "models". An independent nation state normally has a head of state, and determines the extent of its head's executive powers of government or formal representational functions. In terms of protocol: the head of a sovereign, independent state is usually identified as the person who, according to that state's constitution, is the reigning monarch, in the case of a monarchy; or the president, in the case of a republic. Among the state constitutions (fundamental laws) that establish different political systems, four major types of heads of state can be distinguished: In a federal constituent or a dependent territory, the same role is fulfilled by the holder of an office corresponding to that of a head of state. For example, in each Canadian province the role is fulfilled by the lieutenant governor, whereas in most British Overseas Territories the powers and duties are performed by the governor. The same applies to Australian states, Indian states, etc. Hong Kong's constitutional document, the Basic Law, for example, specifies the chief executive as the head of the special administrative region, in addition to their role as the head of government. These non-sovereign-state heads, nevertheless, have limited or no role in diplomatic affairs, depending on the status and the norms and practices of the territories concerned. In parliamentary systems the head of state may be merely the nominal chief executive officer, heading the executive branch of the state, and possessing limited executive power. In reality, however, following a process of constitutional evolution, powers are usually only exercised by direction of a cabinet, presided over by a head of government who is answerable to the legislature. This accountability and legitimacy requires that someone be chosen who has a majority support in the legislature (or, at least, not a majority opposition – a subtle but important difference). It also gives the legislature the right to vote down the head of government and their cabinet, forcing it either to resign or seek a parliamentary dissolution. The executive branch is thus said to be responsible (or answerable) to the legislature, with the head of government and cabinet in turn accepting constitutional responsibility for offering constitutional advice to the head of state. In parliamentary constitutional monarchies, the legitimacy of the unelected head of state typically derives from the tacit approval of the people via the elected representatives. Accordingly, at the time of the Glorious Revolution, the English parliament acted of its own authority to name a new king and queen (the joint monarchs Mary II and William III); likewise, Edward VIII's abdication required the approval of each of the six independent realms of which he was monarch. In monarchies with a written constitution, the position of monarch is a creature of the constitution and could quite properly be abolished through a democratic procedure of constitutional amendment, although there are often significant procedural hurdles imposed on such a procedure (as in the Constitution of Spain). In republics with a parliamentary system (such as India, Germany, Austria, Italy and Israel), the head of state is usually titled president and the principal functions of such presidents are mainly ceremonial and symbolic, as opposed to the presidents in a presidential or semi-presidential system. In reality, numerous variants exist to the position of a head of state within a parliamentary system. The older the constitution, the more constitutional leeway tends to exist for a head of state to exercise greater powers over government, as many older parliamentary system constitutions in fact give heads of state powers and functions akin to presidential or semi-presidential systems, in some cases without containing reference to modern democratic principles of accountability to parliament or even to modern governmental offices. Usually, the king had the power of declaring war without previous consent of the parliament. For example, under the 1848 constitution of the Kingdom of Sardinia, and then the Kingdom of Italy, the Statuto Albertino—the parliamentary approval to the government appointed by the king—was customary, but not required by law. So, Italy had a de facto parliamentary system, but a de jure "presidential" system. Examples of heads of state in parliamentary systems using greater powers than usual, either because of ambiguous constitutions or unprecedented national emergencies, include the decision by King Leopold III of the Belgians to surrender on behalf of his state to the invading German army in 1940, against the will of his government. Judging that his responsibility to the nation by virtue of his coronation oath required him to act, he believed that his government's decision to fight rather than surrender was mistaken and would damage Belgium. (Leopold's decision proved highly controversial. After World War II, Belgium voted in a referendum to allow him to resume his monarchical powers and duties, but because of the ongoing controversy he ultimately abdicated.) The Belgian constitutional crisis in 1990, when the head of state refused to sign into law a bill permitting abortion, was resolved by the cabinet assuming the power to promulgate the law while he was treated as "unable to reign" for twenty-four hours. These officials are excluded completely from the executive: they do not possess even theoretical executive powers or any role, even formal, within the government. Hence their states' governments are not referred to by the traditional parliamentary model head of state styles of His/Her Majesty's Government or His/Her Excellency's Government. Within this general category, variants in terms of powers and functions may exist. The Constitution of Japan (日本国憲法, Nihonkoku-Kenpō) was drawn up under the Allied occupation that followed World War II and was intended to replace the previous militaristic and quasi-absolute monarchy system with a form of liberal democracy parliamentary system. The constitution explicitly vests all executive power in the Cabinet, who is chaired by the prime minister (articles 65 and 66) and responsible to the Diet (articles 67 and 69). The emperor is defined in the constitution as "the symbol of the State and of the unity of the people" (article 1), and is generally recognised throughout the world as the Japanese head of state. Although the emperor formally appoints the prime minister to office, article 6 of the constitution requires him to appoint the candidate "as designated by the Diet", without any right to decline appointment. He is a ceremonial figurehead with no independent discretionary powers related to the governance of Japan. Since the passage in Sweden of the 1974 Instrument of Government, the Swedish monarch no longer has many of the standard parliamentary system head of state functions that had previously belonged to him or her, as was the case in the preceding 1809 Instrument of Government. Today, the speaker of the Riksdag appoints (following a vote in the Riksdag) the prime minister and terminates their commission following a vote of no confidence or voluntary resignation. Cabinet members are appointed and dismissed at the sole discretion of the prime minister. Laws and ordinances are promulgated by two Cabinet members in unison signing "On Behalf of the Government" and the government—not the monarch—is the high contracting party with respect to international treaties. The remaining official functions of the sovereign, by constitutional mandate or by unwritten convention, are to open the annual session of the Riksdag, receive foreign ambassadors and sign the letters of credence for Swedish ambassadors, chair the foreign advisory committee, preside at the special Cabinet council when a new prime minister takes office, and to be kept informed by the prime minister on matters of state. In contrast, the only contact the president of Ireland has with the Irish government is through a formal briefing session given by the taoiseach (head of government) to the president. However, the president has no access to documentation and all access to ministers goes through the Department of the Taoiseach. The president does, however, hold limited reserve powers, such as referring a bill to the Supreme Court to test its constitutionality, which are used under the president's discretion. The most extreme non-executive republican head of state is the President of Israel, which holds no reserve powers whatsoever. The least ceremonial powers held by the president are to provide a mandate to attempt to form a government, to approve the dissolution of the Knesset made by the prime minister, and to pardon criminals or to commute their sentence. Some parliamentary republics (like South Africa, Botswana and Kiribati) have fused the roles of the head of state with the head of government (like in a presidential system), while having the sole executive officer, often called a president, being dependent on the Parliament's confidence to rule (like in a parliamentary system). While also being the leading symbol of the nation, the president in this system acts mostly as a prime minister since the incumbent must be a member of the legislature at the time of the election, answer question sessions in Parliament, avoid motions of no confidence, etc. Semi-presidential systems combine features of presidential and parliamentary systems, notably (in the president-parliamentary subtype) a requirement that the government be answerable to both the president and the legislature. The constitution of the Fifth French Republic provides for a prime minister who is chosen by the president, but who nevertheless must be able to gain support in the National Assembly. Should a president be of one side of the political spectrum and the opposition be in control of the legislature, the president is usually obliged to select someone from the opposition to become prime minister, a process known as Cohabitation. President François Mitterrand, a Socialist, for example, was forced to cohabit with the neo-Gaullist (right wing) Jacques Chirac, who became his prime minister from 1986 to 1988. In the French system, in the event of cohabitation, the president is often allowed to set the policy agenda in security and foreign affairs and the prime minister runs the domestic and economic agenda. Other countries evolve into something akin to a semi-presidential system or indeed a full presidential system. Weimar Germany, for example, in its constitution provided for a popularly elected president with theoretically dominant executive powers that were intended to be exercised only in emergencies, and a cabinet appointed by him from the Reichstag, which was expected, in normal circumstances, to be answerable to the Reichstag. Initially, the president was merely a symbolic figure with the Reichstag dominant; however, persistent political instability, in which governments often lasted only a few months, led to a change in the power structure of the republic, with the president's emergency powers called increasingly into use to prop up governments challenged by critical or even hostile Reichstag votes. By 1932, power had shifted to such an extent that the German president, Paul von Hindenburg, was able to dismiss a chancellor and select his own person for the job, even though the outgoing chancellor possessed the confidence of the Reichstag while the new chancellor did not. Subsequently, President von Hindenburg used his power to appoint Adolf Hitler as Chancellor without consulting the Reichstag. Note: The head of state in a "presidential" system may not actually hold the title of "president" - the name of the system refers to any head of state who actually governs and is not directly dependent on the legislature to remain in office. Some constitutions or fundamental laws provide for a head of state who is not only in theory but in practice chief executive, operating separately from, and independent from, the legislature. This system is known as a "presidential system" and sometimes called the "imperial model", because the executive officials of the government are answerable solely and exclusively to a presiding, acting head of state, and is selected by and on occasion dismissed by the head of state without reference to the legislature. It is notable that some presidential systems, while not providing for collective executive accountability to the legislature, may require legislative approval for individuals prior to their assumption of cabinet office and empower the legislature to remove a president from office (for example, in the United States of America). In this case the debate centers on confirming them into office, not removing them from office, and does not involve the power to reject or approve proposed cabinet members en bloc, so accountability does not operate in the same sense understood as a parliamentary system. Presidential systems are a notable feature of constitutions in the Americas, including those of Argentina, Brazil, Colombia, El Salvador, Mexico and Venezuela; this is generally attributed to the strong influence of the United States in the region, and as the United States Constitution served as an inspiration and model for the Latin American wars of independence of the early 19th century. Most presidents in such countries are selected by democratic means (popular direct or indirect election); however, like all other systems, the presidential model also encompasses people who become head of state by other means, notably through military dictatorship or coup d'état, as often seen in Latin American, Middle Eastern and other presidential regimes. Some of the characteristics of a presidential system, such as a strong dominant political figure with an executive answerable to them, not the legislature can also be found among absolute monarchies, parliamentary monarchies and single party (e.g., Communist) regimes, but in most cases of dictatorship, their stated constitutional models are applied in name only and not in political theory or practice. In certain states under Marxist–Leninist constitutions of the constitutionally socialist state type inspired by the former Union of Soviet Socialist Republics (USSR) and its constitutive Soviet republics, real political power belonged to the sole legal party. In these states, there was no formal office of head of state, but rather the leader of the legislative branch was considered to be the closest common equivalent of a head of state as a natural person. In the Soviet Union this position carried such titles as Chairman of the Central Executive Committee of the USSR; Chairman of the Presidium of the Supreme Soviet; and in the case of the Soviet Russia Chairman of the Central Executive Committee of the All-Russian Congress of Soviets (pre-1922), and Chairman of the Bureau of the Central Committee of the Russian SFSR (1956–1966). This position may or may not have been held by the de facto Soviet leader at the moment. For example, Nikita Khrushchev never headed the Supreme Soviet but was First Secretary of the Central Committee of the Communist Party (party leader) and Chairman of the Council of Ministers (head of government). This may even lead to an institutional variability, as in North Korea, where, after the presidency of party leader Kim Il Sung, the office was vacant for years. The late president was granted the posthumous title (akin to some ancient Far Eastern traditions to give posthumous names and titles to royalty) of "Eternal President". All substantive power, as party leader, itself not formally created for four years, was inherited by his son Kim Jong Il. The post of president was formally replaced on 5 September 1998, for ceremonial purposes, by the office of President of the Presidium of the Supreme People's Assembly, while the party leader's post as chairman of the National Defense Commission was simultaneously declared "the highest post of the state", not unlike Deng Xiaoping earlier in the People's Republic of China. In China, under the current country's constitution, the Chinese President is a largely ceremonial office with limited power. However, since 1993, as a matter of convention, the presidency has been held simultaneously by the General Secretary of the Chinese Communist Party, the top leader in the one party system. The presidency is officially regarded as an institution of the state rather than an administrative post; theoretically, the President serves at the pleasure of the National People's Congress, the legislature, and is not legally vested to take executive action on its own prerogative. While clear categories do exist, it is sometimes difficult to choose which category some individual heads of state belong to. In reality, the category to which each head of state belongs is assessed not by theory but by practice. Constitutional change in Liechtenstein in 2003 gave its head of state, the Reigning Prince, constitutional powers that included a veto over legislation and power to dismiss the head of government and cabinet. It could be argued that the strengthening of the Prince's powers, vis-a-vis the Landtag (legislature), has moved Liechtenstein into the semi-presidential category. Similarly the original powers given to the Greek President under the 1974 Hellenic Republic constitution moved Greece closer to the French semi-presidential model. Another complication exists with South Africa, in which the president is in fact elected by the National Assembly (legislature) and is thus similar, in principle, to a head of government in a parliamentary system but is also, in addition, recognised as the head of state. The offices of president of Nauru and president of Botswana are similar in this respect to the South African presidency. Panama, during the military dictatorships of Omar Torrijos and Manuel Noriega, was nominally a presidential republic. However, the elected civilian presidents were effectively figureheads with real political power being exercised by the chief of the Panamanian Defense Forces. Historically, at the time of the League of Nations (1920–1946) and the founding of the United Nations (1945), India's head of state was the monarch of the United Kingdom, ruling directly or indirectly as Emperor of India through the Viceroy and Governor-General of India. Head of state is the highest-ranking constitutional position in a sovereign state. A head of state has some or all of the roles listed below, often depending on the constitutional category (above), and does not necessarily regularly exercise the most power or influence of governance. There is usually a formal public ceremony when a person becomes head of state, or some time after. This may be the swearing in at the inauguration of a president of a republic, or the coronation of a monarch. One of the most important roles of the modern head of state is being a living national symbol of the state; in hereditary monarchies this extends to the monarch being a symbol of the unbroken continuity of the state. For instance, the Canadian monarch is described by the government as being the personification of the Canadian state and is described by the Department of Canadian Heritage as the "personal symbol of allegiance, unity and authority for all Canadians". In many countries, official portraits of the head of state can be found in government offices, courts of law, or other public buildings. The idea, sometimes regulated by law, is to use these portraits to make the public aware of the symbolic connection to the government, a practice that dates back to medieval times. Sometimes this practice is taken to excess, and the head of state becomes the principal symbol of the nation, resulting in the emergence of a personality cult where the image of the head of state is the only visual representation of the country, surpassing other symbols such as the flag. Other common representations are on coins, postage and other stamps and banknotes, sometimes by no more than a mention or signature; and public places, streets, monuments and institutions such as schools are named for current or previous heads of state. In monarchies (e.g., Belgium) there can even be a practice to attribute the adjective "royal" on demand based on existence for a given number of years. However, such political techniques can also be used by leaders without the formal rank of head of state, even party - and other revolutionary leaders without formal state mandate. Heads of state often greet important foreign visitors, particularly visiting heads of state. They assume a host role during a state visit, and the programme may feature playing of the national anthems by a military band, inspection of military troops, official exchange of gifts, and attending a state dinner at the official residence of the host. At home, heads of state are expected to render lustre to various occasions by their presence, such as by attending artistic or sports performances or competitions (often in a theatrical honour box, on a platform, on the front row, at the honours table), expositions, national day celebrations, dedication events, military parades and war remembrances, prominent funerals, visiting different parts of the country and people from different walks of life, and at times performing symbolic acts such as cutting a ribbon, groundbreaking, ship christening, laying the first stone. Some parts of national life receive their regular attention, often on an annual basis, or even in the form of official patronage. The Olympic Charter (rule 55.3) of the International Olympic Committee states that the Olympic summer and winter games shall be opened by the head of state of the host nation, by uttering a single formulaic phrase as determined by the charter. As such invitations may be very numerous, such duties are often in part delegated to such persons as a spouse, a head of government or a cabinet minister or in other cases (possibly as a message, for instance, to distance themselves without rendering offence) just a military officer or civil servant. For non-executive heads of state there is often a degree of censorship by the politically responsible government (such as the head of government). This means that the government discreetly approves agenda and speeches, especially where the constitution (or customary law) assumes all political responsibility by granting the crown inviolability (in fact also imposing political emasculation) as in the Kingdom of Belgium from its very beginning; in a monarchy this may even be extended to some degree to other members of the dynasty, especially the heir to the throne. Below follows a list of examples from different countries of general provisions in law, which either designate an office as head of state or define its general purpose. In the majority of states, whether republics or monarchies, executive authority is vested, at least notionally, in the head of state. In presidential systems the head of state is the actual, de facto chief executive officer. Under parliamentary systems the executive authority is exercised by the head of state, but in practice is done so on the advice of the cabinet of ministers. This produces such terms as "Her Majesty's Government" and "His Excellency's Government." Examples of parliamentary systems in which the head of state is notional chief executive include Australia, Austria, Canada, Denmark, India, Italy, Norway, Spain and the United Kingdom. The few exceptions where the head of state is not even the nominal chief executive - and where supreme executive authority is according to the constitution explicitly vested in a cabinet - include the Czech Republic, Ireland, Israel, Japan and Sweden. The head of state usually appoints most or all the key officials in the government, including the head of government and other cabinet ministers, key judicial figures; and all major office holders in the civil service, foreign service and commissioned officers in the military. In many parliamentary systems, the head of government is appointed with the consent (in practice often decisive) of the legislature, and other figures are appointed on the head of government's advice. In practice, these decisions are often a formality. The last time the prime minister of the United Kingdom was unilaterally selected by the monarch was in 1963, when Queen Elizabeth II appointed Alec Douglas-Home on the advice of outgoing Prime Minister Harold Macmillan. In presidential systems, such as that of the United States, appointments are nominated by the president's sole discretion, but this nomination is often subject to confirmation by the legislature; and specifically in the US, the Senate has to approve senior executive branch and judicial appointments by a simple majority vote. The head of state may also dismiss office-holders. There are many variants on how this can be done. For example, members of the Irish Cabinet are dismissed by the president on the advice of the taoiseach; in other instances, the head of state may be able to dismiss an office holder unilaterally; other heads of state, or their representatives, have the theoretical power to dismiss any office-holder, while it is exceptionally rarely used. In France, while the president cannot force the prime minister to tender the resignation of the government, he can, in practice, request it if the prime minister is from his own majority. In presidential systems, the president often has the power to fire ministers at his sole discretion. In the United States, the unwritten convention calls for the heads of the executive departments to resign on their own initiative when called to do so. Some countries have alternative provisions for senior appointments: In Sweden, under the Instrument of Government of 1974, the Speaker of the Riksdag has the role of formally appointing the prime minister, following a vote in the Riksdag, and the prime minister in turn appoints and dismisses cabinet ministers at his/her sole discretion. Although many constitutions, particularly from the 19th century and earlier, make no explicit mention of a head of state in the generic sense of several present day international treaties, the officeholders corresponding to this position are recognised as such by other countries. In a monarchy, the monarch is generally understood to be the head of state. The Vienna Convention on Diplomatic Relations, which codified longstanding custom, operates under the presumption that the head of a diplomatic mission (i.e. ambassador or nuncio) of the sending state is accredited to the head of state of the receiving state. The head of state accredits (i.e. formally validates) their country's ambassadors (or rarer equivalent diplomatic mission chiefs, such as high commissioner or papal nuncio) through sending formal a Letter of Credence (and a Letter of Recall at the end of a tenure) to other heads of state and, conversely, receives the letters of their foreign counterparts. Without that accreditation, the chief of the diplomatic mission cannot take up their role and receive the highest diplomatic status. The role of a head of state in this regard, is codified in the Vienna Convention on Diplomatic Relations from 1961, which (as of 2017) 191 sovereign states has ratified. However, there are provisions in the Vienna Convention that a diplomatic agent of lesser rank, such as a chargé d'affaires, is accredited to the minister of foreign affairs (or equivalent). The head of state is often designated the high contracting party in international treaties on behalf of the state; signs them either personally or has them signed in his/her name by ministers (government members or diplomats); subsequent ratification, when necessary, may rest with the legislature. The treaties constituting the European Union and the European Communities are noteworthy contemporary cases of multilateral treaties cast in this traditional format, as are the accession agreements of new member states. However, rather than being invariably concluded between two heads of state, it has become common that bilateral treaties are in present times cast in an intergovernmental format, e.g., between the Government of X and the Government of Y, rather than between His Majesty the King of X and His Excellency the President of Y. In Canada, these head of state powers belong to the monarch as part of the royal prerogative, but the Governor General has been permitted to exercise them since 1947 and has done so since the 1970s. A head of state is often, by virtue of holding the highest executive powers, explicitly designated as the commander-in-chief of that nation's armed forces, holding the highest office in all military chains of command. In a constitutional monarchy or non-executive presidency, the head of state may de jure hold ultimate authority over the armed forces but will only normally, as per either written law or unwritten convention, exercise their authority on the advice of their responsible ministers: meaning that the de facto ultimate decision making on military manoeuvres is made elsewhere. The head of state will, regardless of actual authority, perform ceremonial duties related to the country's armed forces, and will sometimes appear in military uniform for these purposes; particularly in monarchies where also the monarch's consort and other members of a royal family may also appear in military garb. This is generally the only time a head of state of a stable, democratic country will appear dressed in such a manner, as statesmen and public are eager to assert the primacy of (civilian, elected) politics over the armed forces. In military dictatorships, or governments which have arisen from coups d'état, the position of commander-in-chief is obvious, as all authority in such a government derives from the application of military force; occasionally a power vacuum created by war is filled by a head of state stepping beyond the normal constitutional role, as King Albert I of Belgium did during World War I. In these and in revolutionary regimes, the head of state, and often executive ministers whose offices are legally civilian, will frequently appear in military uniform. Some countries with a parliamentary system designate officials other than the head of state with command-in-chief powers. The armed forces of the Communist states are under the absolute control of the Communist party. It is usual that the head of state, particularly in parliamentary systems as part of the symbolic role, is the one who opens the annual sessions of the legislature, e.g. the annual State Opening of Parliament with the Speech from the Throne in Britain. Even in presidential systems the head of state often formally reports to the legislature on the present national status, e.g. the State of the Union address in the United States of America, or the State of the Nation Address in South Africa. Most countries require that all bills passed by the house or houses of the legislature be signed into law by the head of state. In some states, such as the United Kingdom, Belgium and Ireland, the head of state is, in fact, formally considered a tier of the legislature. However, in most parliamentary systems, the head of state cannot refuse to sign a bill, and, in granting a bill their assent, indicate that it was passed in accordance with the correct procedures. The signing of a bill into law is formally known as promulgation. Some monarchical states call this procedure royal assent. In some parliamentary systems, the head of state retains certain discretionary powers in relation to bills to be exercised. They may have authority to veto a bill until the houses of the legislature have reconsidered it, and approved it a second time; reserve a bill to be signed later, or suspend it indefinitely (generally in states with royal prerogative; this power is rarely used); refer a bill to the courts to test its constitutionality; refer a bill to the people in a referendum. If the head of state also serves as the chief executive, the head of state can politically control the necessary executive measures without which a proclaimed law can remain dead letter, sometimes for years or even forever. A head of state is often empowered to summon and dissolve the country's legislature. In most parliamentary systems, this is often done on the advice of the head of government. In some parliamentary systems, and in some presidential systems, however, the head of state may do so on their own initiative. Some states have fixed term legislatures, with no option of bringing forward elections (e.g., Article II, Section 3, of the U.S. Constitution). In other systems there are usually fixed terms, but the head of state retains authority to dissolve the legislature in certain circumstances. Where a head of government has lost support in the legislature, some heads of state may refuse a dissolution, where one is requested, thereby forcing the head of government's resignation. In a republic, the head of state nowadays usually bears the title of President, but some have or have had other titles. Titles commonly used by monarchs are King/Queen or Emperor/Empress, but also many other; e.g., Grand Duke, Prince, Emir and Sultan. Though president and various monarchical titles are most commonly used for heads of state, in some nationalistic regimes, the leader adopts, formally or de facto, a unique style simply meaning leader in the national language, e.g., Germany's single national socialist party chief and combined head of state and government, Adolf Hitler, as the Führer between 1934 and 1945. In 1959, when former British crown colony Singapore gained self-government, it adopted the Malay style Yang di-Pertuan Negara (means "the one who own the country" in Malay) for its governor (the actual head of state remained the British monarch). The second and last incumbent of the office, Yusof bin Ishak, kept the style at 31 August 1963 unilateral declaration of independence and after 16 September 1963 accession to Malaysia as a state (so now as a constituent part of the federation, a non-sovereign level). After its expulsion from Malaysia on 9 August 1965, Singapore became a sovereign Commonwealth republic and installed Yusof bin Ishak as its first president. In 1959 after the resignation of Vice President Mohammad Hatta, President Sukarno abolished the position and title of vice-president, assuming the positions of Prime Minister and Head of Cabinet. He also proclaimed himself president for life (Indonesian: Presiden Seumur Hidup Panglima Tertinggi; "panglima" meaning "commander or martial figurehead", "tertinggi" meaning "highest"; roughly translated to English as "Supreme Commander of the Revolution"). He was praised as "Paduka Yang Mulia", a Malay honorific originally given to kings; Sukarno awarded himself titles in that fashion due to his noble ancestry. There are also a few nations in which the exact title and definition of the office of head of state have been vague. During the Chinese Cultural Revolution, following the downfall of Chinese President Liu Shaoqi, no successor was named, so the duties of the head of state were transferred collectively to the Vice Presidents Soong Ching-ling and Dong Biwu, then to the Standing Committee of the National People's Congress, but Chairman Mao Zedong was still the paramount leader. This situation was later changed: the President of the People's Republic of China is now the head of state. Although the presidency is a largely ceremonial office with limited power, the symbolic role of a head of state is now generally performed by Xi Jinping, who is also General Secretary of the Communist Party (Communist Party leader) and Chairman of the Central Military Commission (Supreme Military Command), making him the most powerful person in China. In North Korea, the late Kim Il Sung was named "Eternal President" four years after his death and the presidency was abolished. As a result, some of the duties previously held by the president were constitutionally delegated to the Chairman of the Standing Committee of the Supreme People's Assembly, who performs some of the roles of a head of state, such as accrediting foreign ambassadors and undertaking overseas visits. However, the de jure role of head of state lies within the President of the State Affairs Commission, currently Kim Jong Un, who as the General Secretary of the Workers' Party of Korea and Supreme Commander of the Armed Forces of North Korea, is the most powerful person in North Korea. There is debate as to whether Samoa was an elective monarchy or an aristocratic republic, given the comparative ambiguity of the title O le Ao o le Malo and the nature of the head of state's office. In some states the office of head of state is not expressed in a specific title reflecting that role, but constitutionally awarded to a post of another formal nature. Colonel Muammar Gaddafi initially ruled as combined head of state and briefly head of government of the Libyan Arab Republic, styled as Chairman of the Revolutionary Command Council. In 1977, the Libyan Jamahiriya ("state of the masses") replaced the previous republic, and in March 1979 the role of head of state was transferred to the Secretary-General of the General People's Congress (comparable to a Speaker); in practice however Gaddafi remained the de facto leader as "Guide of the Revolution" until his overthrow in 2011. Sometimes a head of state assumes office as a state becomes legal and political reality, before a formal title for the highest office is determined; thus in the since 1 January 1960 independent republic Cameroon (Cameroun, a former French colony), the first president, Ahmadou Babatoura Ahidjo, was at first not styled président but 'merely' known as chef d'état (French for "head of state") until 5 May 1960. In Uganda, Idi Amin the military leader after the coup of 25 January 1971 was formally styled military head of state till 21 February 1971, only from then on as regular (but unconstitutional, unelected) president. In certain cases a special style is needed to accommodate imperfect statehood, e.g., the title Sadr-i-Riyasat was used in Kashmir after its accession to India, and the Palestine Liberation Organization leader, Yasser Arafat, was styled the first "President of the Palestinian National Authority" in 1994. In 2008, the same office was restyled as "President of the State of Palestine". In medieval Catholic Europe, it was universally accepted that the Pope ranked first among all rulers and was followed by the Holy Roman Emperor. The Pope also had the sole right to determine the precedence of all others. This principle was first challenged by a Protestant ruler, Gustavus Adolphus of Sweden and was later maintained by his country at the Congress of Westphalia. Great Britain would later claim a break of the old principle for the Quadruple Alliance in 1718. However, it was not until the 1815 Congress of Vienna, when it was decided (due to the abolition of the Holy Roman Empire in 1806 and the weak position of France and other catholic states to assert themselves) and remains so to this day, that all sovereign states are treated as equals, whether monarchies or republics. On occasions when multiple heads of state or their representatives meet, precedence is by the host usually determined in alphabetical order (in whatever language the host determines, although French has for much of the 19th and 20th centuries been the lingua franca of diplomacy) or by date of accession. Contemporary international law on precedence, built upon the universally admitted principles since 1815, derives from the Vienna Convention on Diplomatic Relations (in particular, articles 13, 16.1 and Appendix iii). Niccolò Machiavelli used Prince (Italian: Principe) as a generic term for the ruler, similar to contemporary usage of head of state, in his classical treatise The Prince, originally published in 1532: in fact that particular literary genre it belongs to is known as Mirrors for princes. Thomas Hobbes in his Leviathan (1651) used the term Sovereign. In Europe the role of a monarchs has gradually transitioned from that of a sovereign ruler—in the sense of Divine Right of Kings as articulated by Jean Bodin, Absolutism and the "L'etat c'est moi"—to that of a constitutional monarch; parallel with the conceptual evolution of sovereignty from merely the personal rule of a single person, to Westphalian sovereignty (Peace of Westphalia ending both the Thirty Years' War & Eighty Years' War) and popular sovereignty as in consent of the governed; as shown in the Glorious Revolution of 1688 in England & Scotland, the French Revolution in 1789, and the German Revolution of 1918–1919. The monarchies who survived through this era were the ones who were willing to subject themselves to constitutional limitations. Whenever a head of state is not available for any reason, constitutional provisions may allow the role to fall temporarily to an assigned person or collective body. In a republic, this is - depending on provisions outlined by the constitution or improvised - a vice-president, the chief of government, the legislature or its presiding officer. In a monarchy, this is usually a regent or collegial regency (council). For example, in the United States the vice-president acts when the president is incapacitated, and in the United Kingdom the monarch's powers may be delegated to counselors of state when they are abroad or unavailable. Neither of the two co-princes of Andorra is resident in Andorra; each is represented in Andorra by a delegate, though these persons hold no formal title. There are also several methods of head of state succession in the event of the removal, disability or death of an incumbent head of state. In exceptional situations, such as war, occupation, revolution or a coup d'état, constitutional institutions, including the symbolically crucial head of state, may be reduced to a figurehead or be suspended in favour of an emergency office (such as the original Roman dictator) or eliminated by a new "provisionary" regime, such as a collective of the junta type, or removed by an occupying force, such as a military governor (an early example being the Spartan Harmost). In early modern Europe, a single person was often monarch simultaneously of separate states. A composite monarchy is a retrospective label for those cases where the states were governed entirely separately. Of contemporary terms, a personal union had less government co-ordination than a real union. One of the two co-princes of Andorra is the president of France. Such arrangements are not to be confused with supranational entities which are not states and are not defined by a common monarchy but may (or not) have a symbolic, essentially protocollary, titled highest office, e.g., Head of the Commonwealth (held by the British monarch, but not legally reserved for it) or 'Head of the Arab Union' (14 February - 14 July 1958, held by the King of Iraq, during its short-lived Hashemite federation with Jordan). The Commonwealth realms share a monarch, currently Charles III. In the realms other than the United Kingdom, a governor-general (governor general in Canada) is appointed by the sovereign, usually on the advice of the relevant prime minister (although sometimes it is based on the result of a vote in the relevant parliament, which is the case for Papua New Guinea and the Solomon Islands), as a representative and to exercise almost all the royal prerogative according to established constitutional authority. In Australia the present king is generally assumed to be head of state, since the governor-general and the state governors are defined as his "representatives". However, since the governor-general performs almost all national regal functions, the governor-general has occasionally been referred to as head of state in political and media discussion. To a lesser extent, uncertainty has been expressed in Canada as to which officeholder—the monarch, the governor general, or both—can be considered the head of state. New Zealand, Papua New Guinea, and Tuvalu explicitly name the monarch as their head of state (though Tuvalu's constitution states that "references in any law to the Head of State shall be read as including a reference to the governor-general"). Governors-general are frequently treated as heads of state on state and official visits; at the United Nations, they are accorded the status of head of state in addition to the sovereign. An example of a governor-general departing from constitutional convention by acting unilaterally (that is, without direction from ministers, parliament, or the monarch) occurred in 1926, when Canada's governor general refused the head of government's formal advice requesting a dissolution of parliament and a general election. In a letter informing the monarch after the event, the Governor General said: "I have to await the verdict of history to prove my having adopted a wrong course, and this I do with an easy conscience that, right or wrong, I have acted in the interests of Canada and implicated no one else in my decision." Another example occurred when, in the 1975 Australian constitutional crisis, the governor-general unexpectedly dismissed the prime minister in order to break a stalemate between the House of Representatives and Senate over money bills. The governor-general issued a public statement saying he felt it was the only solution consistent with the constitution, his oath of office, and his responsibilities, authority, and duty as governor-general. A letter from the queen's private secretary at the time, Martin Charteris, confirmed that the only person competent to commission an Australian prime minister was the governor-general and it would not be proper for the monarch to personally intervene in matters that the Constitution Act so clearly places within the governor-general's jurisdiction. Other Commonwealth realms that are now constituted with a governor-general as the viceregal representative of Charles III are: Antigua and Barbuda, the Bahamas, Belize, Grenada, Jamaica, New Zealand, Saint Kitts and Nevis, Saint Lucia, and Saint Vincent and the Grenadines. Since antiquity, various dynasties or individual rulers have claimed the right to rule by divine authority, such as the Mandate of Heaven and the divine right of kings. Some monarchs even claimed divine ancestry, such as Egyptian pharaohs and Sapa Incas, who claimed descent from their respective sun gods and often sought to maintain this bloodline by practising incestuous marriage. In Ancient Rome, during the Principate, the title divus ('divine') was conferred (notably posthumously) on the emperor, a symbolic, legitimating element in establishing a de facto dynasty. In Roman Catholicism, the pope was once sovereign pontiff and head of state, first, of the politically important Papal States. After Italian unification, the pope remains head of state of Vatican City. Furthermore, the bishop of Urgell is ex officio one of the two co-princes of Andorra. In the Church of England, the reigning monarch holds the title Defender of the Faith and acts as supreme governor of the Church of England, although this is purely a symbolic role. During the early period of Islam, caliphs were spiritual and temporal absolute successors of the prophet Muhammad. Various political Muslim leaders since have styled themselves Caliph and served as dynastic heads of state, sometimes in addition to another title, such as the Ottoman Sultan. Historically, some theocratic Islamic states known as imamates have been led by imams as head of state, such as in what is now Oman, Yemen, and Saudi Arabia. In the Islamic Republic of Iran, the Supreme Leader, at present Ali Khamenei serves as head of state. The Aga Khans, a unique dynasty of temporal/religious leadership, leading the Nizari offshoot of Shia Islam in Central and South Asia, once ranking among British India's princely states, continue to the present day. In Hinduism, certain dynasties adopted a title expressing their positions as "servant" of a patron deity of the state, but in the sense of a viceroy under an absentee god-king, ruling "in the name of" the patron god(ess), such as Patmanabha Dasa (servant of Vishnu) in the case of the Maharaja of Travancore. From the time of the 5th Dalai Lama until the political retirement of the 14th Dalai Lama in 2011, Dalai Lamas were both political and spiritual leaders ("god-king") of Tibet. Outer Mongolia, the former homeland of the imperial dynasty of Genghis Khan, was another lamaist theocracy from 1585, using various styles, such as tulku. The establishment of the Communist Mongolian People's Republic replaced this regime in 1924. Sometimes multiple individuals are co-equal heads of state, or a corporate person embodies the functions of head of state. In some cases precedence rotates among the members of the collective as the term of office progresses. Of multiple royal systems, a diarchy, in which two rulers is the constitutional norm, may be distinguished from a coregency, in which a monarchy experiences an exceptional period of multiple rulers. Examples of collective republican systems include nominal triumvirates; the French Directory of the 1790s; the seven-member Swiss Federal Council, where each member acts in turn as President for one year; the Presidency of Bosnia and Herzegovina with members from three nations; the two Captains Regent of San Marino, which maintains the tradition of Italian medieval republics that had always had an even number of consuls. In the Roman Republic there were two heads of state, styled consul, both of whom alternated months of authority during their year in office, similarly there was an even number of supreme magistrates in the Italic republics of Ancient Age. In the Athenian Republic there were nine supreme magistrates, styled archons. In Carthage there were two supreme magistrates, styled kings or suffetes (judges). In ancient Sparta there were two hereditary kings, belonging to two dynasties. In the Soviet Union, the Presidium of the Supreme Soviet (between 1938 and 1989) served as the collective head of state. After World War II the Soviet model was subsequently adopted by almost all countries belonged to its sphere of influence. Czechoslovakia remained the only country among them that retained an office of president as a form of a single head of state throughout this period; Poland and Hungary, which initially had western-style constitutions (and therefore, western-style presidencies), switched to the presidium model with the adoption of new Soviet-influenced constitutions; Romania, which was a monarchy before the Soviet takeover, was the only country to move to a unitary presidency from a collective head of state, a move done by dictator Nicolae Ceausescu in 1974. A modern example of a collective head of state is the Sovereignty Council of Sudan, the interim ruling council of Sudan. The Sovereignty Council comprises 11 ministers, who together have exercised all governmental functions for Sudan since the fall of President Omar Al-Bashir. Decisions are made either by consensus or by a super majority vote (8 members). The National Government of the Republic of China, established in 1928, had a panel of about 40 people as collective head of state. Though beginning that year, a provisional constitution made the Kuomintang the sole government party and the National Government bound to the instructions of the Central Executive Committee of that party. The position of head of state can be established in different ways, and with different sources of legitimacy. Power can come from force, but formal legitimacy is often established, even if only by fictitious claims of continuity (e.g., a forged claim of descent from a previous dynasty). There have been cases of sovereignty granted by deliberate act, even when accompanied by orders of succession (as may be the case in a dynastic split). Such grants of sovereignty are usually forced, as is common with self-determination granted after nationalist revolts. This occurred with the last Attalid king of Hellenistic Pergamon, who by testament left his realm to Rome to avoid a disastrous conquest. Under a theocracy, perceived divine status translated into earthly authority under divine law. This can take the form of supreme divine authority above the state's, granting a tool for political influence to a priesthood. In this way, the Amun priesthood reversed the reforms of Pharaoh Akhenaten after his death. The division of theocratic power can be disputed, as happened between the Pope and Holy Roman Emperor in the investiture conflict when the temporal power sought to control key clergy nominations in order to guarantee popular support, and thereby his own legitimacy, by incorporating the formal ceremony of unction during coronation. The notion of a social contract holds that the nation—either the whole people or the electorate—gives a mandate, through acclamation or election. Individual heads of state may acquire their position by virtue of a constitution. An example is the Seychelles, as the 1976 Independence Constitution of the Seychelles, Article 31, stated that James Mancham would be the first President of the Republic by name, rather than by the fact he was the Prime Minister of colonial Seychelles immediately before independence. The position of a monarch is usually hereditary, but in constitutional monarchies, there are usually restrictions on the incumbent's exercise of powers and prohibitions on the possibility of choosing a successor by other means than by birth. In a hereditary monarchy, the position of monarch is inherited according to a statutory or customary order of succession, usually within one royal family tracing its origin through a historical dynasty or bloodline. This usually means that the heir to the throne is known well in advance of becoming monarch to ensure a smooth succession. However, many cases of uncertain succession in European history have often led to wars of succession. Primogeniture, in which the eldest child of the monarch is first in line to become monarch, is the most common system in hereditary monarchy. The order of succession is usually affected by rules on gender. Historically "agnatic primogeniture" or "patrilineal primogeniture" was favoured, that is inheritance according to seniority of birth among the sons of a monarch or head of family, with sons and their male issue inheriting before brothers and their issue, and male-line males inheriting before females of the male line. This is the same as semi-Salic primogeniture. Complete exclusion of females from dynastic succession is commonly referred to as application of the Salic law (see Terra salica). Before primogeniture was enshrined in European law and tradition, kings would often secure the succession by having their successor (usually their eldest son) crowned during their own lifetime, so for a time there would be two kings in coregency – a senior king and a junior king. Examples include Henry the Young King of England and the early Direct Capetians in France. Sometimes, however, primogeniture can operate through the female line. In some systems a female may rule as monarch only when the male line dating back to a common ancestor is exhausted. In 1980, Sweden, by rewriting its 1810 Act of Succession, became the first European monarchy to declare equal (full cognatic) primogeniture, meaning that the eldest child of the monarch, whether female or male, ascends to the throne. Other European monarchies (such as the Netherlands in 1983, Norway in 1990 and Belgium in 1991) have since followed suit. Similar reforms were proposed in 2011 for the United Kingdom and the other Commonwealth realms, which came into effect in 2015 after having been approved by all of the affected nations. Sometimes religion is affected; under the Act of Settlement 1701 all Roman Catholics and all persons who have married Roman Catholics are ineligible to be the British monarch and are skipped in the order of succession. In some monarchies there may be liberty for the incumbent, or some body convening after the death of the monarch, to choose from eligible members of the ruling house, often limited to legitimate descendants of the dynasty's founder. Rules of succession may be further limited by state religion, residency, equal marriage or even permission from the legislature. Other hereditary systems of succession included tanistry, which is semi-elective and gives weight to merit and Agnatic seniority. In some monarchies, such as Saudi Arabia, succession to the throne usually first passes to the monarch's next eldest brother, and only after that to the monarch's children (agnatic seniority). Election usually is the constitutional way to choose the head of state of a republic, and some monarchies, either directly through popular election, indirectly by members of the legislature or of a special college of electors (such as the Electoral College in the United States), or as an exclusive prerogative. Exclusive prerogative allows the heads of states of constituent monarchies of a federation to choose the head of state for the federation among themselves, as in the United Arab Emirates and Malaysia. The Pope, head of state of Vatican City, is chosen by previously appointed cardinals under 80 years of age from among themselves in a papal conclave. A head of state can be empowered to designate his successor, such as Lord Protector of the Commonwealth Oliver Cromwell, who was succeeded by his son Richard. A head of state may seize power by force or revolution. This is not the same as the use of force to maintain power, as is practised by authoritarian or totalitarian rulers. Dictators often use democratic titles, though some proclaim themselves monarchs. Examples of the latter include Emperor Napoleon I of France and King Zog of Albania. In Spain, general Francisco Franco adopted the formal title Jefe del Estado, or Chief of State, and established himself as regent for a vacant monarchy. Uganda's Idi Amin was one of several who named themselves President for Life. A foreign power can establishing a branch of their own dynasty, or one friendly to their interests. This was the outcome of the Russo-Swedish War from 1741 to 1743 where the Russian Empress made the imposition of her relative Adolf Frederick as the heir to the Swedish Throne, to succeed Frederick I who lacked legitimate issue, as a peace condition. Apart from violent overthrow, a head of state's position can be lost in several ways, including death, another by expiration of the constitutional term of office, abdication, or resignation. In some cases, an abdication cannot occur unilaterally, but comes into effect only when approved by an act of parliament, as in the case of British King Edward VIII. The post can also be abolished by constitutional change; in such cases, an incumbent may be allowed to finish their term. Of course, a head of state position will cease to exist if the state itself does. Heads of state generally enjoy widest inviolability, although some states allow impeachment, or a similar constitutional procedure by which the highest legislative or judicial authorities are empowered to revoke the head of state's mandate on exceptional grounds. This may be a common crime, a political sin, or an act by which the head of state violates such provisions as an established religion mandatory for the monarch. By similar procedure, an original mandate may be declared invalid. Effigies, memorials and monuments of former heads of state can be designed to represent the history or aspirations of a state or its people, such as the equestrian bronze sculpture of Kaiser Wilhelm I, first Emperor of a unified Germany erected in Berlin at the end of the nineteenth century; or the Victoria Memorial erected in front of Buckingham Palace London, commemorating Queen Victoria and her reign (1837–1901), and unveiled in 1911 by her grandson, King George V; or the monument, placed in front of the Victoria Memorial Hall, Kolkata (Calcutta) (1921), commemorating Queen Victoria's reign as Empress of India from 1876. Another, twentieth century, example is the Mount Rushmore National Memorial, a group sculpture constructed (1927–1941) on a conspicuous skyline in the Black Hills of South Dakota (40th state of the Union, 1889), in the midwestern United States, representing the territorial expansion of the United States in the first 130 years from its founding, which is promoted as the "Shrine of Democracy". Former presidents of the United States, while holding no political powers per se, sometimes continue to exert influence in national and world affairs. A monarch may retain his style and certain prerogatives after abdication, as did King Leopold III of Belgium, who left the throne to his son after winning a referendum which allowed him to retain a full royal household deprived him of a constitutional or representative role. Napoleon transformed the Italian principality of Elba, where he was imprisoned, into a miniature version of his First Empire, with most trappings of a sovereign monarchy, until his Cent Jours escape and reseizure of power in France convinced his opponents, reconvening the Vienna Congress in 1815, to revoke his gratuitous privileges and send him to die in exile on barren Saint Helena. By tradition, deposed monarchs who have not freely abdicated continue to use their monarchical titles as a courtesy for the rest of their lives. Hence, even after Constantine II ceased to be King of the Hellenes, it is still common to refer to the deposed king and his family as if Constantine II were still on the throne, as many European royal courts and households do in guest lists at royal weddings, as in Sweden in 2010, Britain in 2011 and Luxembourg in 2012. The current Hellenic Republic opposes the right of their deposed monarch and former royal family members to be referred to by their former titles or bearing a surname indicating royal status, and has enacted legislation which hinders acquisition of Greek citizenship unless those terms are met. The former king brought this issue, along with property ownership issues, before the European Court of Human Rights for alleged violations of the European Convention on Human Rights, but lost with respect to the name issue. However, some other states have no problem with deposed monarchs being referred to by their former title, and even allow them to travel internationally on the state's diplomatic passport. The Italian constitution provides that a former president of the Republic takes the title President Emeritus of the Italian Republic and he or she is also a senator for life, and enjoys immunity, flight status and official residences certain privileges.
[[File:Poumon artificiel.jpg|thumb|Examining a patient in a tank respirator]] A '''disease''' or '''medical condition''' is an unhealthy state where something bad happens to the [[human body|body]] or [[mind]].<ref>{{DorlandsDict|three/000030493|Disease}}</ref> Diseases can cause [[pain]], parts of the body to stop working the right way, or death. The word ''disease'' is sometimes used to include: * parts of the body being hurt, * not having the usual abilities, * medical problems or [[syndrome]]s, * [[infection]]s by [[microorganism]]s, * feeling unhealthy, such as having pain or feeling hot (called [[symptom|'symptoms']]), * unusual shapes of body parts. == Causes == A disease can be caused by many things. Sometimes [[germ]]s enter our body through food, water or air. A person can be [[Infection|infected]] by infectious agents like [[bacteria]], [[virus]]es or [[fungus]]. Disease can also be caused by eating bad or old [[food]]s. There are small [[Pathogen|germs]] in old foods that can cause diseases. Sometimes the [[germs]] produce [[chemicals]] or [[toxin]]s which causes the disease. One of the most common causes of disease is poor [[sanitation]] and lack of clean water. Some deadly diseases like [[malaria]] in [[tropical]] parts of the world are spread by a [[mosquito]]. Animals that spread disease are called ''vectors''. There are many vectors, including snails, ticks, and fleas. Some people are born with '[[Genetics|genetic]] diseases'. These are diseases because of an error or [[mutation]] in a person's [[DNA]]. An example of a mutation is [[cancer]]. Living or working in an unhealthy [[environment]] can also be a cause for diseases. Diseases are more common in older people. == Treatments == Some diseases can be helped with [[medicine]]. [[Infection]]s can often be cured by [[antibiotics]], though [[resistance to antibiotics]] is a problem. Some disease may be helped by [[surgery]]. Not every disease can be helped with medicine or surgery, though. Some diseases must be treated during the whole life; they are ''chronic'' (long-lasting) diseases. An example of a chronic disease is [[diabetes mellitus]]. Diabetes can be treated (made better) but it can not yet be cured (made to totally go away). People who usually treat diseases are called [[Medical doctor|doctors]] or [[physician]]s. == Prevention == Some diseases that are common or very bad are tested for even in people who do not show any [[symptoms]]. If these diseases are found early they can be treated before they cause problems. An example would be checking a woman for [[cervical cancer]] with a test called a [[pap smear]]. If cervical cancer is found early it can be cured. If it is found later it usually causes death. Another example is [[immunization]]. The basic idea is to make the body ready for a disease. The body has its own defense against disease called the [[immune system]]. One special characteristic of the [[immune system]] is its ability to remember some diseases. If a person is sick and recovers, the immune system will produce a substance called [[antibodies]] which fight the disease if it comes back to the person. The antibody is [[specific]] to a particular disease or [[antigen]]. An example of this is [[measles]] which is a [[virus]]. A person (usually a child) who had never been sick with [[measles]] is given a milder form of the [[virus]], this causes the [[immune system]] to produce antibodies against the virus. If this person is exposed to the same virus in the future, the person's immune system will remember and will fight the virus. For general prevention to be useful: *The disease must be found and stopped in early stage. *The disease should be common or be easy to recognize. *The test for the disease should be easy, work all the time, and not hurt people. *The society is well-trained and can recognize most common symptoms on some diseases. *The treatment for the disease should be safe and be easy for people to get. ==Epidemiology== {{Main|Epidemiology}} Epidemiology is the study of the cause of disease. Some diseases are more popular for people with common characteristics, like similar origins, sociological background, food or nationality. Without good epidemiological research some diseases can be hard to track and to name. Some diseases can be taken for something else. This is why epidemiology takes a huge part in understanding how to protect ourselves against viruses, toxins and bacteria.<ref>{{cite news| url = https://www.nytimes.com/2011/03/08/health/research/08fatigue.html| title = Defining an illness is fodder for debate| author = Tuller, David| newspaper = The New York Times| date = 4 March 2011}}</ref> ==Related pages== *[[Health]] *[[Healthy lifestyle]] *[[Viruses]] ==References== <references /> [[Category:Basic English 850 words]] [[Category:Diseases| ]]
A disease is a particular abnormal condition that adversely affects the structure or function of all or part of an organism and is not immediately due to any external injury. Diseases are often known to be medical conditions that are associated with specific signs and symptoms. A disease may be caused by external factors such as pathogens or by internal dysfunctions. For example, internal dysfunctions of the immune system can produce a variety of different diseases, including various forms of immunodeficiency, hypersensitivity, allergies, and autoimmune disorders. In humans, disease is often used more broadly to refer to any condition that causes pain, dysfunction, distress, social problems, or death to the person affected, or similar problems for those in contact with the person. In this broader sense, it sometimes includes injuries, disabilities, disorders, syndromes, infections, isolated symptoms, deviant behaviors, and atypical variations of structure and function, while in other contexts and for other purposes these may be considered distinguishable categories. Diseases can affect people not only physically but also mentally, as contracting and living with a disease can alter the affected person's perspective on life. Death due to disease is called death by natural causes. There are four main types of disease: infectious diseases, deficiency diseases, hereditary diseases (including both genetic and non-genetic hereditary diseases), and physiological diseases. Diseases can also be classified in other ways, such as communicable versus non-communicable diseases. The deadliest diseases in humans are coronary artery disease (blood flow obstruction), followed by cerebrovascular disease and lower respiratory infections. In developed countries, the diseases that cause the most sickness overall are neuropsychiatric conditions, such as depression and anxiety. The study of disease is called pathology, which includes the study of etiology, or cause. In many cases, terms such as disease, disorder, morbidity, sickness and illness are used interchangeably; however, there are situations when specific terms are considered preferable. In an infectious disease, the incubation period is the time between infection and the appearance of symptoms. The latency period is the time between infection and the ability of the disease to spread to another person, which may precede, follow, or be simultaneous with the appearance of symptoms. Some viruses also exhibit a dormant phase, called viral latency, in which the virus hides in the body in an inactive state. For example, varicella zoster virus causes chickenpox in the acute phase; after recovery from chickenpox, the virus may remain dormant in nerve cells for many years, and later cause herpes zoster (shingles). Diseases may be classified by cause, pathogenesis (mechanism by which the disease is caused), or by symptoms. Alternatively, diseases may be classified according to the organ system involved, though this is often complicated since many diseases affect more than one organ. A chief difficulty in nosology is that diseases often cannot be defined and classified clearly, especially when cause or pathogenesis are unknown. Thus diagnostic terms often only reflect a symptom or set of symptoms (syndrome). Classical classification of human disease derives from the observational correlation between pathological analysis and clinical syndromes. Today it is preferred to classify them by their cause if it is known. The most known and used classification of diseases is the World Health Organization's ICD. This is periodically updated. Currently, the last publication is the ICD-11. Diseases can be caused by any number of factors and may be acquired or congenital. Microorganisms, genetics, the environment or a combination of these can contribute to a diseased state. Only some diseases such as influenza are contagious and commonly believed infectious. The microorganisms that cause these diseases are known as pathogens and include varieties of bacteria, viruses, protozoa, and fungi. Infectious diseases can be transmitted, e.g. by hand-to-mouth contact with infectious material on surfaces, by bites of insects or other carriers of the disease, and from contaminated water or food (often via fecal contamination), etc. Also, there are sexually transmitted diseases. In some cases, microorganisms that are not readily spread from person to person play a role, while other diseases can be prevented or ameliorated with appropriate nutrition or other lifestyle changes. Some diseases, such as most (but not all) forms of cancer, heart disease, and mental disorders, are non-infectious diseases. Many non-infectious diseases have a partly or completely genetic basis (see genetic disorder) and may thus be transmitted from one generation to another. Social determinants of health are the social conditions in which people live that determine their health. Illnesses are generally related to social, economic, political, and environmental circumstances. Social determinants of health have been recognized by several health organizations such as the Public Health Agency of Canada and the World Health Organization to greatly influence collective and personal well-being. The World Health Organization's Social Determinants Council also recognizes Social determinants of health in poverty. When the cause of a disease is poorly understood, societies tend to mythologize the disease or use it as a metaphor or symbol of whatever that culture considers evil. For example, until the bacterial cause of tuberculosis was discovered in 1882, experts variously ascribed the disease to heredity, a sedentary lifestyle, depressed mood, and overindulgence in sex, rich food, or alcohol, all of which were social ills at the time. When a disease is caused by a pathogenic organism (e.g., when malaria is caused by Plasmodium), one should not confuse the pathogen (the cause of the disease) with disease itself. For example, West Nile virus (the pathogen) causes West Nile fever (the disease). The misuse of basic definitions in epidemiology is frequent in scientific publications. Many diseases and disorders can be prevented through a variety of means. These include sanitation, proper nutrition, adequate exercise, vaccinations and other self-care and public health measures, such as obligatory face mask mandates. Medical therapies or treatments are efforts to cure or improve a disease or other health problems. In the medical field, therapy is synonymous with the word treatment. Among psychologists, the term may refer specifically to psychotherapy or "talk therapy". Common treatments include medications, surgery, medical devices, and self-care. Treatments may be provided by an organized health care system, or informally, by the patient or family members. Preventive healthcare is a way to avoid an injury, sickness, or disease in the first place. A treatment or cure is applied after a medical problem has already started. A treatment attempts to improve or remove a problem, but treatments may not produce permanent cures, especially in chronic diseases. Cures are a subset of treatments that reverse diseases completely or end medical problems permanently. Many diseases that cannot be completely cured are still treatable. Pain management (also called pain medicine) is that branch of medicine employing an interdisciplinary approach to the relief of pain and improvement in the quality of life of those living with pain. Treatment for medical emergencies must be provided promptly, often through an emergency department or, in less critical situations, through an urgent care facility. Epidemiology is the study of the factors that cause or encourage diseases. Some diseases are more common in certain geographic areas, among people with certain genetic or socioeconomic characteristics, or at different times of the year. Epidemiology is considered a cornerstone methodology of public health research and is highly regarded in evidence-based medicine for identifying risk factors for diseases. In the study of communicable and non-communicable diseases, the work of epidemiologists ranges from outbreak investigation to study design, data collection, and analysis including the development of statistical models to test hypotheses and the documentation of results for submission to peer-reviewed journals. Epidemiologists also study the interaction of diseases in a population, a condition known as a syndemic. Epidemiologists rely on a number of other scientific disciplines such as biology (to better understand disease processes), biostatistics (the current raw information available), Geographic Information Science (to store data and map disease patterns) and social science disciplines (to better understand proximate and distal risk factors). Epidemiology can help identify causes as well as guide prevention efforts. In studying diseases, epidemiology faces the challenge of defining them. Especially for poorly understood diseases, different groups might use significantly different definitions. Without an agreed-on definition, different researchers may report different numbers of cases and characteristics of the disease. Some morbidity databases are compiled with data supplied by states and territories health authorities, at national levels or larger scale (such as European Hospital Morbidity Database (HMDB)) which may contain hospital discharge data by detailed diagnosis, age and sex. The European HMDB data was submitted by European countries to the World Health Organization Regional Office for Europe. Disease burden is the impact of a health problem in an area measured by financial cost, mortality, morbidity, or other indicators. There are several measures used to quantify the burden imposed by diseases on people. The years of potential life lost (YPLL) is a simple estimate of the number of years that a person's life was shortened due to a disease. For example, if a person dies at the age of 65 from a disease, and would probably have lived until age 80 without that disease, then that disease has caused a loss of 15 years of potential life. YPLL measurements do not account for how disabled a person is before dying, so the measurement treats a person who dies suddenly and a person who died at the same age after decades of illness as equivalent. In 2004, the World Health Organization calculated that 932 million years of potential life were lost to premature death. The quality-adjusted life year (QALY) and disability-adjusted life year (DALY) metrics are similar but take into account whether the person was healthy after diagnosis. In addition to the number of years lost due to premature death, these measurements add part of the years lost to being sick. Unlike YPLL, these measurements show the burden imposed on people who are very sick, but who live a normal lifespan. A disease that has high morbidity, but low mortality, has a high DALY and a low YPLL. In 2004, the World Health Organization calculated that 1.5 billion disability-adjusted life years were lost to disease and injury. In the developed world, heart disease and stroke cause the most loss of life, but neuropsychiatric conditions like major depressive disorder cause the most years lost to being sick. How a society responds to diseases is the subject of medical sociology. A condition may be considered a disease in some cultures or eras but not in others. For example, obesity can represent wealth and abundance, and is a status symbol in famine-prone areas and some places hard-hit by HIV/AIDS. Epilepsy is considered a sign of spiritual gifts among the Hmong people. Sickness confers the social legitimization of certain benefits, such as illness benefits, work avoidance, and being looked after by others. The person who is sick takes on a social role called the sick role. A person who responds to a dreaded disease, such as cancer, in a culturally acceptable fashion may be publicly and privately honored with higher social status. In return for these benefits, the sick person is obligated to seek treatment and work to become well once more. As a comparison, consider pregnancy, which is not interpreted as a disease or sickness, even if the mother and baby may both benefit from medical care. Most religions grant exceptions from religious duties to people who are sick. For example, one whose life would be endangered by fasting on Yom Kippur or during Ramadan is exempted from the requirement, or even forbidden from participating. People who are sick are also exempted from social duties. For example, ill health is the only socially acceptable reason for an American to refuse an invitation to the White House. The identification of a condition as a disease, rather than as simply a variation of human structure or function, can have significant social or economic implications. The controversial recognition of diseases such as repetitive stress injury (RSI) and post-traumatic stress disorder (PTSD) has had a number of positive and negative effects on the financial and other responsibilities of governments, corporations, and institutions towards individuals, as well as on the individuals themselves. The social implication of viewing aging as a disease could be profound, though this classification is not yet widespread. Lepers were people who were historically shunned because they had an infectious disease, and the term "leper" still evokes social stigma. Fear of disease can still be a widespread social phenomenon, though not all diseases evoke extreme social stigma. Social standing and economic status affect health. Diseases of poverty are diseases that are associated with poverty and low social status; diseases of affluence are diseases that are associated with high social and economic status. Which diseases are associated with which states vary according to time, place, and technology. Some diseases, such as diabetes mellitus, may be associated with both poverty (poor food choices) and affluence (long lifespans and sedentary lifestyles), through different mechanisms. The term lifestyle diseases describes diseases associated with longevity and that are more common among older people. For example, cancer is far more common in societies in which most members live until they reach the age of 80 than in societies in which most members die before they reach the age of 50. An illness narrative is a way of organizing a medical experience into a coherent story that illustrates the sick individual's personal experience. People use metaphors to make sense of their experiences with disease. The metaphors move disease from an objective thing that exists to an affective experience. The most popular metaphors draw on military concepts: Disease is an enemy that must be feared, fought, battled, and routed. The patient or the healthcare provider is a warrior, rather than a passive victim or bystander. The agents of communicable diseases are invaders; non-communicable diseases constitute internal insurrection or civil war. Because the threat is urgent, perhaps a matter of life and death, unthinkably radical, even oppressive, measures are society's and the patient's moral duty as they courageously mobilize to struggle against destruction. The War on Cancer is an example of this metaphorical use of language. This language is empowering to some patients, but leaves others feeling like they are failures. Another class of metaphors describes the experience of illness as a journey: The person travels to or from a place of disease, and changes himself, discovers new information, or increases his experience along the way. He may travel "on the road to recovery" or make changes to "get on the right track" or choose "pathways". Some are explicitly immigration-themed: the patient has been exiled from the home territory of health to the land of the ill, changing identity and relationships in the process. This language is more common among British healthcare professionals than the language of physical aggression. Some metaphors are disease-specific. Slavery is a common metaphor for addictions: The alcoholic is enslaved by drink, and the smoker is captive to nicotine. Some cancer patients treat the loss of their hair from chemotherapy as a metonymy or metaphor for all the losses caused by the disease. Some diseases are used as metaphors for social ills: "Cancer" is a common description for anything that is endemic and destructive in society, such as poverty, injustice, or racism. AIDS was seen as a divine judgment for moral decadence, and only by purging itself from the "pollution" of the "invader" could society become healthy again. More recently, when AIDS seemed less threatening, this type of emotive language was applied to avian flu and type 2 diabetes mellitus. Authors in the 19th century commonly used tuberculosis as a symbol and a metaphor for transcendence. People with the disease were portrayed in literature as having risen above daily life to become ephemeral objects of spiritual or artistic achievement. In the 20th century, after its cause was better understood, the same disease became the emblem of poverty, squalor, and other social problems.
{{Infobox carbon}} '''Carbon''' is a very important [[chemical element]], with a [[chemical symbol]] of '''C'''. All known [[life]] on [[Earth]] needs it to survive. Carbon has [[atomic mass]] 12 and [[atomic number]] 6. It is a [[nonmetal]], meaning that it is not a [[metal]]. When [[iron]] is [[alloy]]ed with carbon, hard [[steel]] is formed. Carbon in the form of [[coal]] is an important [[fuel]]. == Chemistry of carbon == A whole type of chemistry, called [[organic chemistry]], is about carbon and its [[chemical compound|compounds]]. Carbon makes many types of compounds. ''Hydrocarbons'' are molecules with carbon and hydrogen. [[Methane]], [[Propane]], and many other [[fuel]]s are hydrocarbons. Many of the substances that people use daily are organic compounds. Carbon, [[hydrogen]], [[nitrogen]], [[oxygen]], and some other elements like [[sulfur]] and [[phosphorus]] together form most life on [[earth]] (see [[List of biologically important elements]]). Carbon forms a very large number of [[organic compound]]s because it can form strong [[Chemical bond|bonds]] with itself and with other elements. Because of the amounts of carbon living things have, all organic things are considered "carbon-based". Each carbon atom usually forms four [[chemical bond]]s, which are strong connections to other atoms to form [[molecule]]s. The kind of bond that carbon makes is called a [[covalent bond]]. These bonds allow carbon to form many kinds of small and large molecules. A molecule of [[methane]] is the smallest; it has four [[hydrogen]] atoms bonded to carbon. The bonds can be [[double bond]]s, meaning that two bonds form between carbon and another atom to make a stronger connection. For example, [[carbon dioxide]] has two [[oxygen]] atoms, and each one is double bonded to carbon. Carbon can even form three bonds with another atom, called a [[triple bond]]. For example, in the gas [[acetylene]] carbon forms a triple bond with another carbon atom. By bonding to other carbon atoms, carbon can form long chain-shaped [[molecule]]s, called [[polymer]]s, such as [[plastic]]s and [[protein]]s. Atoms of other elements can be part of the long polymer chains, often nitrogen or oxygen. Pure carbon forms [[diamond]] by bonding to four other carbon atoms in a [[3D|three dimensional]] [[crystal]]. It forms [[graphite]] by bonding to three other carbon atoms to form thin flat layers. == Etymology == The name of carbon comes from [[Latin]] ''carbo'', meaning [[charcoal]]. In many foreign languages the words for carbon, [[coal]] and charcoal are synonyms. == Types of carbon == {{main|Allotropes of carbon}} Carbon in [[nature]] is found in three forms called [[allotrope]]s: [[diamond]], [[graphite]], and [[fullerenes]]. Graphite, with [[clay]], is in [[pencil]]s. It is very soft. The carbon atoms in it make rings, which are on top of each other and slide very easily. Diamonds are the hardest natural [[mineral]]. Fullerenes are a "[[soccer]] ball" shape of carbon. They are mostly of interest to [[science]]. A special, man-made, tube-shaped allotrope of carbon is the [[carbon nanotube]]. Carbon nanotubes are very hard, so they might be used in [[armor]]. Nanotubes might be useful in [[nanotechnology]]. There are 10 million known carbon [[chemical compound|compounds]]. [[File:Eight Allotropes of Carbon.png|thumb|350px|right|Some forms of carbon: a) [[diamond]]; b) [[graphite]]; c) [[lonsdaleite]]; d-f) [[fullerene]]s (C60, C540, C70); g) [[amorphous carbon]]; h) [[carbon nanotube]].]] == Radiocarbon dating == {{main|Radiocarbon dating}} A [[radioactive]] [[isotope]] of carbon, carbon-14, can be used to figure out how old some objects are or when something died. As long as something is on the surface of the earth and taking in carbon, the amount of carbon-14 stays the same. When an object stops taking in carbon, the carbon-14 amount goes down. Because the ''[[Half-life (element)|half-life]]'' (how long it takes for half of a radioactive isotope to go away) of carbon-14 is 5730 years,<ref name="NaturesBB">{{cite book|author=Emsley, John|year=2001|title=Nature's Building blocks|publisher=Oxford University Press|isbn=0-19-850341-5}}</ref> [[scientist]]s can see how old the object is by how much carbon-14 is left. == Where carbon is == Carbon is in many places in the universe. It was first made in old [[star]]s. Carbon is the fourth most common element in the [[sun]].<ref name="NaturesBB"/> The [[atmosphere]]s of [[Venus]] and [[Mars (planet)|Mars]] are mostly [[carbon dioxide]].<ref name="Webelements">{{cite web|author=University of Sheffield and Webelements Ltd.|year=2007|title=Chemistry : Periodic Table : carbon : key information|url=http://www.webelements.com/webelements/elements/text/C/key.html}}</ref> Carbon is important to the [[human body]] and other [[living thing]]s, and it is the second most common element in the human body, at 23% of all body weight.<ref name="NaturesBB"/> It is also a key part of many biological molecules ([[molecules]] used in life). Most of the carbon on Earth is [[coal]]. [[Graphite]] is in many (typically [[desert]]) areas, including [[Sri Lanka]], [[Madagascar]], and [[Russia]]. Diamonds are rare and are found largely in [[Africa]]. Carbon is also in some meteorites. ==Related pages== * [[List of common elements]] * [[Carbon cycle]] ** [[Carbon sequestration]] == References == {{reflist}} == Other websites == * [http://www.webelements.com/webelements/elements/text/C/key.html Carbon at Webelements] {{Periodic Table}} {{authority control}} [[Category:Carbon]] {{chem-stub}} {{sci-stub}}
Carbon (from Latin carbo 'coal') is a chemical element; it has symbol C and atomic number 6. It is nonmetallic and tetravalent—meaning that its atoms are able to form up to four covalent bonds due to its valence shell exhibiting 4 electrons. It belongs to group 14 of the periodic table. Carbon makes up about 0.025 percent of Earth's crust. Three isotopes occur naturally, C and C being stable, while C is a radionuclide, decaying with a half-life of about 5,730 years. Carbon is one of the few elements known since antiquity. Carbon is the 15th most abundant element in the Earth's crust, and the fourth most abundant element in the universe by mass after hydrogen, helium, and oxygen. Carbon's abundance, its unique diversity of organic compounds, and its unusual ability to form polymers at the temperatures commonly encountered on Earth, enables this element to serve as a common element of all known life. It is the second most abundant element in the human body by mass (about 18.5%) after oxygen. The atoms of carbon can bond together in diverse ways, resulting in various allotropes of carbon. Well-known allotropes include graphite, diamond, amorphous carbon, and fullerenes. The physical properties of carbon vary widely with the allotropic form. For example, graphite is opaque and black, while diamond is highly transparent. Graphite is soft enough to form a streak on paper (hence its name, from the Greek verb "γράφειν" which means "to write"), while diamond is the hardest naturally occurring material known. Graphite is a good electrical conductor while diamond has a low electrical conductivity. Under normal conditions, diamond, carbon nanotubes, and graphene have the highest thermal conductivities of all known materials. All carbon allotropes are solids under normal conditions, with graphite being the most thermodynamically stable form at standard temperature and pressure. They are chemically resistant and require high temperature to react even with oxygen. The most common oxidation state of carbon in inorganic compounds is +4, while +2 is found in carbon monoxide and transition metal carbonyl complexes. The largest sources of inorganic carbon are limestones, dolomites and carbon dioxide, but significant quantities occur in organic deposits of coal, peat, oil, and methane clathrates. Carbon forms a vast number of compounds, with about two hundred million having been described and indexed; and yet that number is but a fraction of the number of theoretically possible compounds under standard conditions. The allotropes of carbon include graphite, one of the softest known substances, and diamond, the hardest naturally occurring substance. It bonds readily with other small atoms, including other carbon atoms, and is capable of forming multiple stable covalent bonds with suitable multivalent atoms. Carbon is a component element in the large majority of all chemical compounds, with about two hundred million examples having been described in the published chemical literature. Carbon also has the highest sublimation point of all elements. At atmospheric pressure it has no melting point, as its triple point is at 10.8 ± 0.2 megapascals (106.6 ± 2.0 atm; 1,566 ± 29 psi) and 4,600 ± 300 K (4,330 ± 300 °C; 7,820 ± 540 °F), so it sublimes at about 3,900 K (3,630 °C; 6,560 °F). Graphite is much more reactive than diamond at standard conditions, despite being more thermodynamically stable, as its delocalised pi system is much more vulnerable to attack. For example, graphite can be oxidised by hot concentrated nitric acid at standard conditions to mellitic acid, C6(CO2H)6, which preserves the hexagonal units of graphite while breaking up the larger structure. Carbon sublimes in a carbon arc, which has a temperature of about 5800 K (5,530 °C or 9,980 °F). Thus, irrespective of its allotropic form, carbon remains solid at higher temperatures than the highest-melting-point metals such as tungsten or rhenium. Although thermodynamically prone to oxidation, carbon resists oxidation more effectively than elements such as iron and copper, which are weaker reducing agents at room temperature. Carbon is the sixth element, with a ground-state electron configuration of 1s2s2p, of which the four outer electrons are valence electrons. Its first four ionisation energies, 1086.5, 2352.6, 4620.5 and 6222.7 kJ/mol, are much higher than those of the heavier group-14 elements. The electronegativity of carbon is 2.5, significantly higher than the heavier group-14 elements (1.8–1.9), but close to most of the nearby nonmetals, as well as some of the second- and third-row transition metals. Carbon's covalent radii are normally taken as 77.2 pm (C−C), 66.7 pm (C=C) and 60.3 pm (C≡C), although these may vary depending on coordination number and what the carbon is bonded to. In general, covalent radius decreases with lower coordination number and higher bond order. Carbon-based compounds form the basis of all known life on Earth, and the carbon-nitrogen-oxygen cycle provides a small portion of the energy produced by the Sun, and most of the energy in larger stars (e.g. Sirius). Although it forms an extraordinary variety of compounds, most forms of carbon are comparatively unreactive under normal conditions. At standard temperature and pressure, it resists all but the strongest oxidizers. It does not react with sulfuric acid, hydrochloric acid, chlorine or any alkalis. At elevated temperatures, carbon reacts with oxygen to form carbon oxides and will rob oxygen from metal oxides to leave the elemental metal. This exothermic reaction is used in the iron and steel industry to smelt iron and to control the carbon content of steel: Carbon reacts with sulfur to form carbon disulfide, and it reacts with steam in the coal-gas reaction used in coal gasification: Carbon combines with some metals at high temperatures to form metallic carbides, such as the iron carbide cementite in steel and tungsten carbide, widely used as an abrasive and for making hard tips for cutting tools. The system of carbon allotropes spans a range of extremes: Atomic carbon is a very short-lived species and, therefore, carbon is stabilized in various multi-atomic structures with diverse molecular configurations called allotropes. The three relatively well-known allotropes of carbon are amorphous carbon, graphite, and diamond. Once considered exotic, fullerenes are nowadays commonly synthesized and used in research; they include buckyballs, carbon nanotubes, carbon nanobuds and nanofibers. Several other exotic allotropes have also been discovered, such as lonsdaleite, glassy carbon, carbon nanofoam and linear acetylenic carbon (carbyne). Graphene is a two-dimensional sheet of carbon with the atoms arranged in a hexagonal lattice. As of 2009, graphene appears to be the strongest material ever tested. The process of separating it from graphite will require some further technological development before it is economical for industrial processes. If successful, graphene could be used in the construction of a space elevator. It could also be used to safely store hydrogen for use in a hydrogen based engine in cars. The amorphous form is an assortment of carbon atoms in a non-crystalline, irregular, glassy state, not held in a crystalline macrostructure. It is present as a powder, and is the main constituent of substances such as charcoal, lampblack (soot), and activated carbon. At normal pressures, carbon takes the form of graphite, in which each atom is bonded trigonally to three others in a plane composed of fused hexagonal rings, just like those in aromatic hydrocarbons. The resulting network is 2-dimensional, and the resulting flat sheets are stacked and loosely bonded through weak van der Waals forces. This gives graphite its softness and its cleaving properties (the sheets slip easily past one another). Because of the delocalization of one of the outer electrons of each atom to form a π-cloud, graphite conducts electricity, but only in the plane of each covalently bonded sheet. This results in a lower bulk electrical conductivity for carbon than for most metals. The delocalization also accounts for the energetic stability of graphite over diamond at room temperature. At very high pressures, carbon forms the more compact allotrope, diamond, having nearly twice the density of graphite. Here, each atom is bonded tetrahedrally to four others, forming a 3-dimensional network of puckered six-membered rings of atoms. Diamond has the same cubic structure as silicon and germanium, and because of the strength of the carbon-carbon bonds, it is the hardest naturally occurring substance measured by resistance to scratching. Contrary to the popular belief that "diamonds are forever", they are thermodynamically unstable (ΔfG°(diamond, 298 K) = 2.9 kJ/mol) under normal conditions (298 K, 10 Pa) and should theoretically transform into graphite. But due to a high activation energy barrier, the transition into graphite is so slow at normal temperature that it is unnoticeable. However, at very high temperatures diamond will turn into graphite, and diamonds can burn up in a house fire. The bottom left corner of the phase diagram for carbon has not been scrutinized experimentally. Although a computational study employing density functional theory methods reached the conclusion that as T → 0 K and p → 0 Pa, diamond becomes more stable than graphite by approximately 1.1 kJ/mol, more recent and definitive experimental and computational studies show that graphite is more stable than diamond for T < 400 K, without applied pressure, by 2.7 kJ/mol at T = 0 K and 3.2 kJ/mol at T = 298.15 K. Under some conditions, carbon crystallizes as lonsdaleite, a hexagonal crystal lattice with all atoms covalently bonded and properties similar to those of diamond. Fullerenes are a synthetic crystalline formation with a graphite-like structure, but in place of flat hexagonal cells only, some of the cells of which fullerenes are formed may be pentagons, nonplanar hexagons, or even heptagons of carbon atoms. The sheets are thus warped into spheres, ellipses, or cylinders. The properties of fullerenes (split into buckyballs, buckytubes, and nanobuds) have not yet been fully analyzed and represent an intense area of research in nanomaterials. The names fullerene and buckyball are given after Richard Buckminster Fuller, popularizer of geodesic domes, which resemble the structure of fullerenes. The buckyballs are fairly large molecules formed completely of carbon bonded trigonally, forming spheroids (the best-known and simplest is the soccerball-shaped C60 buckminsterfullerene). Carbon nanotubes (buckytubes) are structurally similar to buckyballs, except that each atom is bonded trigonally in a curved sheet that forms a hollow cylinder. Nanobuds were first reported in 2007 and are hybrid buckytube/buckyball materials (buckyballs are covalently bonded to the outer wall of a nanotube) that combine the properties of both in a single structure. Of the other discovered allotropes, carbon nanofoam is a ferromagnetic allotrope discovered in 1997. It consists of a low-density cluster-assembly of carbon atoms strung together in a loose three-dimensional web, in which the atoms are bonded trigonally in six- and seven-membered rings. It is among the lightest known solids, with a density of about 2 kg/m. Similarly, glassy carbon contains a high proportion of closed porosity, but contrary to normal graphite, the graphitic layers are not stacked like pages in a book, but have a more random arrangement. Linear acetylenic carbon has the chemical structure −(C≡C)n− . Carbon in this modification is linear with sp orbital hybridization, and is a polymer with alternating single and triple bonds. This carbyne is of considerable interest to nanotechnology as its Young's modulus is 40 times that of the hardest known material – diamond. In 2015, a team at the North Carolina State University announced the development of another allotrope they have dubbed Q-carbon, created by a high-energy low-duration laser pulse on amorphous carbon dust. Q-carbon is reported to exhibit ferromagnetism, fluorescence, and a hardness superior to diamonds. In the vapor phase, some of the carbon is in the form of highly reactive diatomic carbon dicarbon (C2). When excited, this gas glows green. Carbon is the fourth most abundant chemical element in the observable universe by mass after hydrogen, helium, and oxygen. Carbon is abundant in the Sun, stars, comets, and in the atmospheres of most planets. Some meteorites contain microscopic diamonds that were formed when the Solar System was still a protoplanetary disk. Microscopic diamonds may also be formed by the intense pressure and high temperature at the sites of meteorite impacts. In 2014 NASA announced a greatly upgraded database for tracking polycyclic aromatic hydrocarbons (PAHs) in the universe. More than 20% of the carbon in the universe may be associated with PAHs, complex compounds of carbon and hydrogen without oxygen. These compounds figure in the PAH world hypothesis where they are hypothesized to have a role in abiogenesis and formation of life. PAHs seem to have been formed "a couple of billion years" after the Big Bang, are widespread throughout the universe, and are associated with new stars and exoplanets. It has been estimated that the solid earth as a whole contains 730 ppm of carbon, with 2000 ppm in the core and 120 ppm in the combined mantle and crust. Since the mass of the earth is 5.972×10 kg, this would imply 4360 million gigatonnes of carbon. This is much more than the amount of carbon in the oceans or atmosphere (below). In combination with oxygen in carbon dioxide, carbon is found in the Earth's atmosphere (approximately 900 gigatonnes of carbon — each ppm corresponds to 2.13 Gt) and dissolved in all water bodies (approximately 36,000 gigatonnes of carbon). Carbon in the biosphere has been estimated at 550 gigatonnes but with a large uncertainty, due mostly to a huge uncertainty in the amount of terrestrial deep subsurface bacteria. Hydrocarbons (such as coal, petroleum, and natural gas) contain carbon as well. Coal "reserves" (not "resources") amount to around 900 gigatonnes with perhaps 18,000 Gt of resources. Oil reserves are around 150 gigatonnes. Proven sources of natural gas are about 175×10 cubic metres (containing about 105 gigatonnes of carbon), but studies estimate another 900×10 cubic metres of "unconventional" deposits such as shale gas, representing about 540 gigatonnes of carbon. Carbon is also found in methane hydrates in polar regions and under the seas. Various estimates put this carbon between 500, 2500, or 3,000 Gt. According to one source, in the period from 1751 to 2008 about 347 gigatonnes of carbon were released as carbon dioxide to the atmosphere from burning of fossil fuels. Another source puts the amount added to the atmosphere for the period since 1750 at 879 Gt, and the total going to the atmosphere, sea, and land (such as peat bogs) at almost 2,000 Gt. Carbon is a constituent (about 12% by mass) of the very large masses of carbonate rock (limestone, dolomite, marble, and others). Coal is very rich in carbon (anthracite contains 92–98%) and is the largest commercial source of mineral carbon, accounting for 4,000 gigatonnes or 80% of fossil fuel. As for individual carbon allotropes, graphite is found in large quantities in the United States (mostly in New York and Texas), Russia, Mexico, Greenland, and India. Natural diamonds occur in the rock kimberlite, found in ancient volcanic "necks", or "pipes". Most diamond deposits are in Africa, notably in South Africa, Namibia, Botswana, the Republic of the Congo, and Sierra Leone. Diamond deposits have also been found in Arkansas, Canada, the Russian Arctic, Brazil, and in Northern and Western Australia. Diamonds are now also being recovered from the ocean floor off the Cape of Good Hope. Diamonds are found naturally, but about 30% of all industrial diamonds used in the U.S. are now manufactured. Carbon-14 is formed in upper layers of the troposphere and the stratosphere at altitudes of 9–15 km by a reaction that is precipitated by cosmic rays. Thermal neutrons are produced that collide with the nuclei of nitrogen-14, forming carbon-14 and a proton. As such, 1.5%×10 of atmospheric carbon dioxide contains carbon-14. Carbon-rich asteroids are relatively preponderant in the outer parts of the asteroid belt in the Solar System. These asteroids have not yet been directly sampled by scientists. The asteroids can be used in hypothetical space-based carbon mining, which may be possible in the future, but is currently technologically impossible. Isotopes of carbon are atomic nuclei that contain six protons plus a number of neutrons (varying from 2 to 16). Carbon has two stable, naturally occurring isotopes. The isotope carbon-12 (C) forms 98.93% of the carbon on Earth, while carbon-13 (C) forms the remaining 1.07%. The concentration of C is further increased in biological materials because biochemical reactions discriminate against C. In 1961, the International Union of Pure and Applied Chemistry (IUPAC) adopted the isotope carbon-12 as the basis for atomic weights. Identification of carbon in nuclear magnetic resonance (NMR) experiments is done with the isotope C. Carbon-14 (C) is a naturally occurring radioisotope, created in the upper atmosphere (lower stratosphere and upper troposphere) by interaction of nitrogen with cosmic rays. It is found in trace amounts on Earth of 1 part per trillion (0.0000000001%) or more, mostly confined to the atmosphere and superficial deposits, particularly of peat and other organic materials. This isotope decays by 0.158 MeV β emission. Because of its relatively short half-life of 5730 years, C is virtually absent in ancient rocks. The amount of C in the atmosphere and in living organisms is almost constant, but decreases predictably in their bodies after death. This principle is used in radiocarbon dating, invented in 1949, which has been used extensively to determine the age of carbonaceous materials with ages up to about 40,000 years. There are 15 known isotopes of carbon and the shortest-lived of these is C which decays through proton emission and alpha decay and has a half-life of 1.98739 × 10 s. The exotic C exhibits a nuclear halo, which means its radius is appreciably larger than would be expected if the nucleus were a sphere of constant density. Formation of the carbon atomic nucleus occurs within a giant or supergiant star through the triple-alpha process. This requires a nearly simultaneous collision of three alpha particles (helium nuclei), as the products of further nuclear fusion reactions of helium with hydrogen or another helium nucleus produce lithium-5 and beryllium-8 respectively, both of which are highly unstable and decay almost instantly back into smaller nuclei. The triple-alpha process happens in conditions of temperatures over 100 megakelvins and helium concentration that the rapid expansion and cooling of the early universe prohibited, and therefore no significant carbon was created during the Big Bang. According to current physical cosmology theory, carbon is formed in the interiors of stars on the horizontal branch. When massive stars die as supernova, the carbon is scattered into space as dust. This dust becomes component material for the formation of the next-generation star systems with accreted planets. The Solar System is one such star system with an abundance of carbon, enabling the existence of life as we know it. It is the opinion of most scholars that all the carbon in the Solar System and the Milky Way comes from dying stars. The CNO cycle is an additional hydrogen fusion mechanism that powers stars, wherein carbon operates as a catalyst. Rotational transitions of various isotopic forms of carbon monoxide (for example, CO, CO, and CO) are detectable in the submillimeter wavelength range, and are used in the study of newly forming stars in molecular clouds. Under terrestrial conditions, conversion of one element to another is very rare. Therefore, the amount of carbon on Earth is effectively constant. Thus, processes that use carbon must obtain it from somewhere and dispose of it somewhere else. The paths of carbon in the environment form the carbon cycle. For example, photosynthetic plants draw carbon dioxide from the atmosphere (or seawater) and build it into biomass, as in the Calvin cycle, a process of carbon fixation. Some of this biomass is eaten by animals, while some carbon is exhaled by animals as carbon dioxide. The carbon cycle is considerably more complicated than this short loop; for example, some carbon dioxide is dissolved in the oceans; if bacteria do not consume it, dead plant or animal matter may become petroleum or coal, which releases carbon when burned. Carbon can form very long chains of interconnecting carbon–carbon bonds, a property that is called catenation. Carbon-carbon bonds are strong and stable. Through catenation, carbon forms a countless number of compounds. A tally of unique compounds shows that more contain carbon than do not. A similar claim can be made for hydrogen because most organic compounds contain hydrogen chemically bonded to carbon or another common element like oxygen or nitrogen. The simplest form of an organic molecule is the hydrocarbon—a large family of organic molecules that are composed of hydrogen atoms bonded to a chain of carbon atoms. A hydrocarbon backbone can be substituted by other atoms, known as heteroatoms. Common heteroatoms that appear in organic compounds include oxygen, nitrogen, sulfur, phosphorus, and the nonradioactive halogens, as well as the metals lithium and magnesium. Organic compounds containing bonds to metal are known as organometallic compounds (see below). Certain groupings of atoms, often including heteroatoms, recur in large numbers of organic compounds. These collections, known as functional groups, confer common reactivity patterns and allow for the systematic study and categorization of organic compounds. Chain length, shape and functional groups all affect the properties of organic molecules. In most stable compounds of carbon (and nearly all stable organic compounds), carbon obeys the octet rule and is tetravalent, meaning that a carbon atom forms a total of four covalent bonds (which may include double and triple bonds). Exceptions include a small number of stabilized carbocations (three bonds, positive charge), radicals (three bonds, neutral), carbanions (three bonds, negative charge) and carbenes (two bonds, neutral), although these species are much more likely to be encountered as unstable, reactive intermediates. Carbon occurs in all known organic life and is the basis of organic chemistry. When united with hydrogen, it forms various hydrocarbons that are important to industry as refrigerants, lubricants, solvents, as chemical feedstock for the manufacture of plastics and petrochemicals, and as fossil fuels. When combined with oxygen and hydrogen, carbon can form many groups of important biological compounds including sugars, lignans, chitins, alcohols, fats, aromatic esters, carotenoids and terpenes. With nitrogen it forms alkaloids, and with the addition of sulfur also it forms antibiotics, amino acids, and rubber products. With the addition of phosphorus to these other elements, it forms DNA and RNA, the chemical-code carriers of life, and adenosine triphosphate (ATP), the most important energy-transfer molecule in all living cells. Norman Horowitz, head of the Mariner and Viking missions to Mars (1965-1976), considered that the unique characteristics of carbon made it unlikely that any other element could replace carbon, even on another planet, to generate the biochemistry necessary for life. Commonly carbon-containing compounds which are associated with minerals or which do not contain bonds to the other carbon atoms, halogens, or hydrogen, are treated separately from classical organic compounds; the definition is not rigid, and the classification of some compounds can vary from author to author (see reference articles above). Among these are the simple oxides of carbon. The most prominent oxide is carbon dioxide (CO2). This was once the principal constituent of the paleoatmosphere, but is a minor component of the Earth's atmosphere today. Dissolved in water, it forms carbonic acid (H2CO3), but as most compounds with multiple single-bonded oxygens on a single carbon it is unstable. Through this intermediate, though, resonance-stabilized carbonate ions are produced. Some important minerals are carbonates, notably calcite. Carbon disulfide (CS2) is similar. Nevertheless, due to its physical properties and its association with organic synthesis, carbon disulfide is sometimes classified as an organic solvent. The other common oxide is carbon monoxide (CO). It is formed by incomplete combustion, and is a colorless, odorless gas. The molecules each contain a triple bond and are fairly polar, resulting in a tendency to bind permanently to hemoglobin molecules, displacing oxygen, which has a lower binding affinity. Cyanide (CN), has a similar structure, but behaves much like a halide ion (pseudohalogen). For example, it can form the nitride cyanogen molecule ((CN)2), similar to diatomic halides. Likewise, the heavier analog of cyanide, cyaphide (CP), is also considered inorganic, though most simple derivatives are highly unstable. Other uncommon oxides are carbon suboxide (C3O2), the unstable dicarbon monoxide (C2O), carbon trioxide (CO3), cyclopentanepentone (C5O5), cyclohexanehexone (C6O6), and mellitic anhydride (C12O9). However, mellitic anhydride is the triple acyl anhydride of mellitic acid; moreover, it contains a benzene ring. Thus, many chemists consider it to be organic. With reactive metals, such as tungsten, carbon forms either carbides (C) or acetylides (C2) to form alloys with high melting points. These anions are also associated with methane and acetylene, both very weak acids. With an electronegativity of 2.5, carbon prefers to form covalent bonds. A few carbides are covalent lattices, like carborundum (SiC), which resembles diamond. Nevertheless, even the most polar and salt-like of carbides are not completely ionic compounds. Organometallic compounds by definition contain at least one carbon-metal covalent bond. A wide range of such compounds exist; major classes include simple alkyl-metal compounds (for example, tetraethyllead), η-alkene compounds (for example, Zeise's salt), and η-allyl compounds (for example, allylpalladium chloride dimer); metallocenes containing cyclopentadienyl ligands (for example, ferrocene); and transition metal carbene complexes. Many metal carbonyls and metal cyanides exist (for example, tetracarbonylnickel and potassium ferricyanide); some workers consider metal carbonyl and cyanide complexes without other carbon ligands to be purely inorganic, and not organometallic. However, most organometallic chemists consider metal complexes with any carbon ligand, even 'inorganic carbon' (e.g., carbonyls, cyanides, and certain types of carbides and acetylides) to be organometallic in nature. Metal complexes containing organic ligands without a carbon-metal covalent bond (e.g., metal carboxylates) are termed metalorganic compounds. While carbon is understood to strongly prefer formation of four covalent bonds, other exotic bonding schemes are also known. Carboranes are highly stable dodecahedral derivatives of the [B12H12] unit, with one BH replaced with a CH. Thus, the carbon is bonded to five boron atoms and one hydrogen atom. The cation [(Ph3PAu)6C] contains an octahedral carbon bound to six phosphine-gold fragments. This phenomenon has been attributed to the aurophilicity of the gold ligands, which provide additional stabilization of an otherwise labile species. In nature, the iron-molybdenum cofactor (FeMoco) responsible for microbial nitrogen fixation likewise has an octahedral carbon center (formally a carbide, C(-IV)) bonded to six iron atoms. In 2016, it was confirmed that, in line with earlier theoretical predictions, the hexamethylbenzene dication contains a carbon atom with six bonds. More specifically, the dication could be described structurally by the formulation [MeC(η-C5Me5)], making it an "organic metallocene" in which a MeC fragment is bonded to a η-C5Me5 fragment through all five of the carbons of the ring. It is important to note that in the cases above, each of the bonds to carbon contain less than two formal electron pairs. Thus, the formal electron count of these species does not exceed an octet. This makes them hypercoordinate but not hypervalent. Even in cases of alleged 10-C-5 species (that is, a carbon with five ligands and a formal electron count of ten), as reported by Akiba and co-workers, electronic structure calculations conclude that the electron population around carbon is still less than eight, as is true for other compounds featuring four-electron three-center bonding. The English name carbon comes from the Latin carbo for coal and charcoal, whence also comes the French charbon, meaning charcoal. In German, Dutch and Danish, the names for carbon are Kohlenstoff, koolstof, and kulstof respectively, all literally meaning coal-substance. Carbon was discovered in prehistory and was known in the forms of soot and charcoal to the earliest human civilizations. Diamonds were known probably as early as 2500 BCE in China, while carbon in the form of charcoal was made around Roman times by the same chemistry as it is today, by heating wood in a pyramid covered with clay to exclude air. In 1722, René Antoine Ferchault de Réaumur demonstrated that iron was transformed into steel through the absorption of some substance, now known to be carbon. In 1772, Antoine Lavoisier showed that diamonds are a form of carbon; when he burned samples of charcoal and diamond and found that neither produced any water and that both released the same amount of carbon dioxide per gram. In 1779, Carl Wilhelm Scheele showed that graphite, which had been thought of as a form of lead, was instead identical with charcoal but with a small admixture of iron, and that it gave "aerial acid" (his name for carbon dioxide) when oxidized with nitric acid. In 1786, the French scientists Claude Louis Berthollet, Gaspard Monge and C. A. Vandermonde confirmed that graphite was mostly carbon by oxidizing it in oxygen in much the same way Lavoisier had done with diamond. Some iron again was left, which the French scientists thought was necessary to the graphite structure. In their publication they proposed the name carbone (Latin carbonum) for the element in graphite which was given off as a gas upon burning graphite. Antoine Lavoisier then listed carbon as an element in his 1789 textbook. A new allotrope of carbon, fullerene, that was discovered in 1985 includes nanostructured forms such as buckyballs and nanotubes. Their discoverers – Robert Curl, Harold Kroto, and Richard Smalley – received the Nobel Prize in Chemistry in 1996. The resulting renewed interest in new forms led to the discovery of further exotic allotropes, including glassy carbon, and the realization that "amorphous carbon" is not strictly amorphous. Commercially viable natural deposits of graphite occur in many parts of the world, but the most important sources economically are in China, India, Brazil, and North Korea. Graphite deposits are of metamorphic origin, found in association with quartz, mica, and feldspars in schists, gneisses, and metamorphosed sandstones and limestone as lenses or veins, sometimes of a metre or more in thickness. Deposits of graphite in Borrowdale, Cumberland, England were at first of sufficient size and purity that, until the 19th century, pencils were made by sawing blocks of natural graphite into strips before encasing the strips in wood. Today, smaller deposits of graphite are obtained by crushing the parent rock and floating the lighter graphite out on water. There are three types of natural graphite—amorphous, flake or crystalline flake, and vein or lump. Amorphous graphite is the lowest quality and most abundant. Contrary to science, in industry "amorphous" refers to very small crystal size rather than complete lack of crystal structure. Amorphous is used for lower value graphite products and is the lowest priced graphite. Large amorphous graphite deposits are found in China, Europe, Mexico and the United States. Flake graphite is less common and of higher quality than amorphous; it occurs as separate plates that crystallized in metamorphic rock. Flake graphite can be four times the price of amorphous. Good quality flakes can be processed into expandable graphite for many uses, such as flame retardants. The foremost deposits are found in Austria, Brazil, Canada, China, Germany and Madagascar. Vein or lump graphite is the rarest, most valuable, and highest quality type of natural graphite. It occurs in veins along intrusive contacts in solid lumps, and it is only commercially mined in Sri Lanka. According to the USGS, world production of natural graphite was 1.1 million tonnes in 2010, to which China contributed 800,000 t, India 130,000 t, Brazil 76,000 t, North Korea 30,000 t and Canada 25,000 t. No natural graphite was reported mined in the United States, but 118,000 t of synthetic graphite with an estimated value of $998 million was produced in 2009. The diamond supply chain is controlled by a limited number of powerful businesses, and is also highly concentrated in a small number of locations around the world (see figure). Only a very small fraction of the diamond ore consists of actual diamonds. The ore is crushed, during which care has to be taken in order to prevent larger diamonds from being destroyed in this process and subsequently the particles are sorted by density. Today, diamonds are located in the diamond-rich density fraction with the help of X-ray fluorescence, after which the final sorting steps are done by hand. Before the use of X-rays became commonplace, the separation was done with grease belts; diamonds have a stronger tendency to stick to grease than the other minerals in the ore. Historically diamonds were known to be found only in alluvial deposits in southern India. India led the world in diamond production from the time of their discovery in approximately the 9th century BC to the mid-18th century AD, but the commercial potential of these sources had been exhausted by the late 18th century and at that time India was eclipsed by Brazil where the first non-Indian diamonds were found in 1725. Diamond production of primary deposits (kimberlites and lamproites) only started in the 1870s after the discovery of the diamond fields in South Africa. Production has increased over time and an accumulated total of over 4.5 billion carats have been mined since that date. Most commercially viable diamond deposits were in Russia, Botswana, Australia and the Democratic Republic of Congo. By 2005, Russia produced almost one-fifth of the global diamond output (mostly in Yakutia territory; for example, Mir pipe and Udachnaya pipe) but the Argyle mine in Australia became the single largest source, producing 14 million carats in 2018. New finds, the Canadian mines at Diavik and Ekati, are expected to become even more valuable owing to their production of gem quality stones. In the United States, diamonds have been found in Arkansas, Colorado, and Montana. In 2004, a startling discovery of a microscopic diamond in the United States led to the January 2008 bulk-sampling of kimberlite pipes in a remote part of Montana. Carbon is essential to all known living systems, and without it life as we know it could not exist (see alternative biochemistry). The major economic use of carbon other than food and wood is in the form of hydrocarbons, most notably the fossil fuel methane gas and crude oil (petroleum). Crude oil is distilled in refineries by the petrochemical industry to produce gasoline, kerosene, and other products. Cellulose is a natural, carbon-containing polymer produced by plants in the form of wood, cotton, linen, and hemp. Cellulose is used primarily for maintaining structure in plants. Commercially valuable carbon polymers of animal origin include wool, cashmere, and silk. Plastics are made from synthetic carbon polymers, often with oxygen and nitrogen atoms included at regular intervals in the main polymer chain. The raw materials for many of these synthetic substances come from crude oil. The uses of carbon and its compounds are extremely varied. It can form alloys with iron, of which the most common is carbon steel. Graphite is combined with clays to form the 'lead' used in pencils used for writing and drawing. It is also used as a lubricant and a pigment, as a molding material in glass manufacture, in electrodes for dry batteries and in electroplating and electroforming, in brushes for electric motors, and as a neutron moderator in nuclear reactors. Charcoal is used as a drawing material in artwork, barbecue grilling, iron smelting, and in many other applications. Wood, coal and oil are used as fuel for production of energy and heating. Gem quality diamond is used in jewelry, and industrial diamonds are used in drilling, cutting and polishing tools for machining metals and stone. Plastics are made from fossil hydrocarbons, and carbon fiber, made by pyrolysis of synthetic polyester fibers is used to reinforce plastics to form advanced, lightweight composite materials. Carbon fiber is made by pyrolysis of extruded and stretched filaments of polyacrylonitrile (PAN) and other organic substances. The crystallographic structure and mechanical properties of the fiber depend on the type of starting material, and on the subsequent processing. Carbon fibers made from PAN have structure resembling narrow filaments of graphite, but thermal processing may re-order the structure into a continuous rolled sheet. The result is fibers with higher specific tensile strength than steel. Carbon black is used as the black pigment in printing ink, artist's oil paint, and water colours, carbon paper, automotive finishes, India ink and laser printer toner. Carbon black is also used as a filler in rubber products such as tyres and in plastic compounds. Activated charcoal is used as an absorbent and adsorbent in filter material in applications as diverse as gas masks, water purification, and kitchen extractor hoods, and in medicine to absorb toxins, poisons, or gases from the digestive system. Carbon is used in chemical reduction at high temperatures. Coke is used to reduce iron ore into iron (smelting). Case hardening of steel is achieved by heating finished steel components in carbon powder. Carbides of silicon, tungsten, boron, and titanium are among the hardest known materials, and are used as abrasives in cutting and grinding tools. Carbon compounds make up most of the materials used in clothing, such as natural and synthetic textiles and leather, and almost all of the interior surfaces in the built environment other than glass, stone, drywall and metal. The diamond industry falls into two categories: one dealing with gem-grade diamonds and the other, with industrial-grade diamonds. While a large trade in both types of diamonds exists, the two markets function dramatically differently. Unlike precious metals such as gold or platinum, gem diamonds do not trade as a commodity: there is a substantial mark-up in the sale of diamonds, and there is not a very active market for resale of diamonds. Industrial diamonds are valued mostly for their hardness and heat conductivity, with the gemological qualities of clarity and color being mostly irrelevant. About 80% of mined diamonds (equal to about 100 million carats or 20 tonnes annually) are unsuitable for use as gemstones and relegated for industrial use (known as bort). Synthetic diamonds, invented in the 1950s, found almost immediate industrial applications; 3 billion carats (600 tonnes) of synthetic diamond is produced annually. The dominant industrial use of diamond is in cutting, drilling, grinding, and polishing. Most of these applications do not require large diamonds; in fact, most diamonds of gem-quality except for their small size can be used industrially. Diamonds are embedded in drill tips or saw blades, or ground into a powder for use in grinding and polishing applications. Specialized applications include use in laboratories as containment for high-pressure experiments (see diamond anvil cell), high-performance bearings, and limited use in specialized windows. With the continuing advances in the production of synthetic diamonds, new applications are becoming feasible. Garnering much excitement is the possible use of diamond as a semiconductor suitable for microchips, and because of its exceptional heat conductance property, as a heat sink in electronics. Pure carbon has extremely low toxicity to humans and can be handled safely in the form of graphite or charcoal. It is resistant to dissolution or chemical attack, even in the acidic contents of the digestive tract. Consequently, once it enters into the body's tissues it is likely to remain there indefinitely. Carbon black was probably one of the first pigments to be used for tattooing, and Ötzi the Iceman was found to have carbon tattoos that survived during his life and for 5200 years after his death. Inhalation of coal dust or soot (carbon black) in large quantities can be dangerous, irritating lung tissues and causing the congestive lung disease, coalworker's pneumoconiosis. Diamond dust used as an abrasive can be harmful if ingested or inhaled. Microparticles of carbon are produced in diesel engine exhaust fumes, and may accumulate in the lungs. In these examples, the harm may result from contaminants (e.g., organic chemicals, heavy metals) rather than from the carbon itself. Carbon generally has low toxicity to life on Earth; but carbon nanoparticles are deadly to Drosophila. Carbon may burn vigorously and brightly in the presence of air at high temperatures. Large accumulations of coal, which have remained inert for hundreds of millions of years in the absence of oxygen, may spontaneously combust when exposed to air in coal mine waste tips, ship cargo holds and coal bunkers, and storage dumps. In nuclear applications where graphite is used as a neutron moderator, accumulation of Wigner energy followed by a sudden, spontaneous release may occur. Annealing to at least 250 °C can release the energy safely, although in the Windscale fire the procedure went wrong, causing other reactor materials to combust. The great variety of carbon compounds include such lethal poisons as tetrodotoxin, the lectin ricin from seeds of the castor oil plant Ricinus communis, cyanide (CN), and carbon monoxide; and such essentials to life as glucose and protein.
[[File:Rust03102006.JPG|thumb|200px|Rusting iron]] [[File:Large bonfire.jpg|thumb|200px|A bonfire is an example for [[redox]]]] {{TopicTOC-Chemistry}} A '''chemical reaction''' happens when one or more [[chemical]]s are changed into one or more other chemicals. Examples: * [[iron]] and [[oxygen]] combining to make [[rust]] * [[vinegar]] and [[baking soda]] combining to make [[sodium]] [[acetate]], [[carbon dioxide]] and water * things burning or exploding * many reactions that happen inside living things, such as [[photosynthesis]] * [[Electrochemistry|electrochemical]] reactions when discharging or recharging [[Battery|batteries]] Some reactions are fast, and others are slow. Some happen at different speeds, depending on [[temperature]] or other things. For example, [[wood]] does not react with [[air]] when it is cold, but if it is made hot enough, it will start to burn. Some reactions give out [[energy]]. These are [[exothermic reaction]]s. In other reactions, energy is taken in. These are [[endothermic reaction]]s. [[Nuclear reaction]]s are ''not'' chemical reactions. Chemical reactions involve only the [[electrons]] of atoms; nuclear reactions involve the [[protons]] and [[neutrons]] in the [[atomic nucleus|atomic nuclei]]. == Four basic types == [[File:Chemical reactions.svg|thumb|center|500px|The four basic chemical reactions types: synthesis, decomposition, single replacement and double replacement]] === Synthesis === In a synthesis reaction, two or more simple substances combine to form a more complex substance. :<math>A + B \longrightarrow AB</math> "Two or more reactants giving one product" is another way to identify a synthesis reaction. One example of a synthesis reaction is the combination of [[iron]] and [[sulfur]] to form [[iron(II) sulfide]]: :<math>8Fe + S_8 \longrightarrow 8FeS</math> Another example is simple hydrogen gas combined with simple oxygen gas to produce a more complex substance, such as water.<ref name="to react or not to react">[http://utahscience.oremjr.alpine.k12.ut.us/sciber99/8th/matter/sciber/chemtype.htm To react or not to react?] {{Webarchive|url=https://web.archive.org/web/20150110214558/http://utahscience.oremjr.alpine.k12.ut.us/sciber99/8th/matter/sciber/chemtype.htm |date=2015-01-10 }} Utah State Office of Education. Retrieved 4 June 2011</ref> === Decomposition === A decomposition reaction is when a more complex substance breaks down into its more simple parts. It is thus the opposite of a synthesis reaction, and can be written as:<ref name="to react or not to react"/><ref name="chemical reactions">[http://misterguch.brinkster.net/6typesofchemicalrxn.html Six types of chemical reactions] – MrGuch ChemFiesta.</ref> :<math>AB \longrightarrow A + B</math> One example of a decomposition reaction is the [[electrolysis]] of water to make [[oxygen]] and [[hydrogen]] gas: :<math>2H_2O \longrightarrow 2H_2 + O_2</math> Another example of a decomposition reaction is [[calcium carbonate]] breaking down into [[calcium oxide]] and [[carbon dioxide]] under high temperatures: CaCO<sub>3</sub> —> CaO + CO<sub>2</sub> === Single replacement === In a single replacement reaction, a single uncombined element replaces another in a [[chemical compound|compound]]; in other words, one element trades places with another element in a compound<ref name="to react or not to react"/> These reactions come in the general form of: :<math>A + BC \longrightarrow AC + B</math> One example of a single displacement reaction is when [[magnesium]] replaces [[hydrogen]] in [[water]] to make [[magnesium hydroxide]] and hydrogen gas: :<math>Mg + 2H_2O \longrightarrow Mg(OH)_2 + H_2</math> === Double replacement === In a double replacement reaction, the anions and cations of two compounds switch places and form two entirely different compounds.<ref name="to react or not to react"/> These reactions are in the general form:<ref name="chemical reactions"/> :<math>AB + CD \longrightarrow AD + CB</math> For example, when [[barium chloride]] (BaCl<sub>2</sub>) and [[magnesium sulfate]] (MgSO<sub>4</sub>) react, the SO<sub>4</sub><sup>2−</sup> anion switches places with the 2Cl<sup>−</sup> anion, giving the compounds BaSO<sub>4</sub> and MgCl<sub>2</sub>. Another example of a double displacement reaction is the reaction of [[lead(II) nitrate]] with [[potassium iodide]] to form [[lead(II) iodide]] and [[potassium nitrate]]: :<math>Pb(NO_3)_2 + 2 KI \longrightarrow PbI_2 + 2 KNO_3</math> == Equations == A chemical reaction is being displayed by an equation: :<math>\mathrm{ A + B \longrightarrow C + D}</math> :<small>Here, A and B react to C and D in a chemical reaction.</small> : This is an example of a '''combustion reaction'''. <math>\mathrm{C + O_2 \longrightarrow CO_2}</math> : <small>[[carbon]] + [[oxygen]] → [[carbon dioxide]] </small> == Related pages == * [[Catalysis]] * [[Chemical kinetics]] * [[Chemical synthesis]] * [[Nuclear reaction]] * [[Organic reaction]] * [[Redox]] == Other websites == * [http://www.purchon.com/chemistry/rates.htm Rates of reaction] {{Webarchive|url=https://web.archive.org/web/20070305222603/http://www.purchon.com/chemistry/rates.htm |date=2007-03-05 }} * [http://www.webqc.org/balance.php Online chemical equation balancer] Balances equation of any chemical reaction (full or half-cell) in one click. ==References== {{Reflist}} [[Category:Chemistry]]
A chemical reaction is a process that leads to the chemical transformation of one set of chemical substances to another. Classically, chemical reactions encompass changes that only involve the positions of electrons in the forming and breaking of chemical bonds between atoms, with no change to the nuclei (no change to the elements present), and can often be described by a chemical equation. Nuclear chemistry is a sub-discipline of chemistry that involves the chemical reactions of unstable and radioactive elements where both electronic and nuclear changes can occur. The substance (or substances) initially involved in a chemical reaction are called reactants or reagents. Chemical reactions are usually characterized by a chemical change, and they yield one or more products, which usually have properties different from the reactants. Reactions often consist of a sequence of individual sub-steps, the so-called elementary reactions, and the information on the precise course of action is part of the reaction mechanism. Chemical reactions are described with chemical equations, which symbolically present the starting materials, end products, and sometimes intermediate products and reaction conditions. Chemical reactions happen at a characteristic reaction rate at a given temperature and chemical concentration. Typically, reaction rates increase with increasing temperature because there is more thermal energy available to reach the activation energy necessary for breaking bonds between atoms. A reaction may be classified as redox in which oxidation and reduction occur or non-redox in which there is no oxidation and reduction occurring. Most simple redox reactions may be classified as a combination, decomposition, or single displacement reaction. Different chemical reactions are used during chemical synthesis in order to obtain the desired product. In biochemistry, a consecutive series of chemical reactions (where the product of one reaction is the reactant of the next reaction) form metabolic pathways. These reactions are often catalyzed by protein enzymes. Enzymes increase the rates of biochemical reactions, so that metabolic syntheses and decompositions impossible under ordinary conditions can occur at the temperature and concentrations present within a cell. The general concept of a chemical reaction has been extended to reactions between entities smaller than atoms, including nuclear reactions, radioactive decays and reactions between elementary particles, as described by quantum field theory. Chemical reactions such as combustion in fire, fermentation and the reduction of ores to metals were known since antiquity. Initial theories of transformation of materials were developed by Greek philosophers, such as the Four-Element Theory of Empedocles stating that any substance is composed of the four basic elements – fire, water, air and earth. In the Middle Ages, chemical transformations were studied by alchemists. They attempted, in particular, to convert lead into gold, for which purpose they used reactions of lead and lead-copper alloys with sulfur. The artificial production of chemical substances already was a central goal for medieval alchemists. Examples include the synthesis of ammonium chloride from organic substances as described in the works (c. 850–950) attributed to Jābir ibn Ḥayyān, or the production of mineral acids such as sulfuric and nitric acids by later alchemists, starting from c. 1300. The production of mineral acids involved the heating of sulfate and nitrate minerals such as copper sulfate, alum and saltpeter. In the 17th century, Johann Rudolph Glauber produced hydrochloric acid and sodium sulfate by reacting sulfuric acid and sodium chloride. With the development of the lead chamber process in 1746 and the Leblanc process, allowing large-scale production of sulfuric acid and sodium carbonate, respectively, chemical reactions became implemented into the industry. Further optimization of sulfuric acid technology resulted in the contact process in the 1880s, and the Haber process was developed in 1909–1910 for ammonia synthesis. From the 16th century, researchers including Jan Baptist van Helmont, Robert Boyle, and Isaac Newton tried to establish theories of experimentally observed chemical transformations. The phlogiston theory was proposed in 1667 by Johann Joachim Becher. It postulated the existence of a fire-like element called "phlogiston", which was contained within combustible bodies and released during combustion. This proved to be false in 1785 by Antoine Lavoisier who found the correct explanation of the combustion as a reaction with oxygen from the air. Joseph Louis Gay-Lussac recognized in 1808 that gases always react in a certain relationship with each other. Based on this idea and the atomic theory of John Dalton, Joseph Proust had developed the law of definite proportions, which later resulted in the concepts of stoichiometry and chemical equations. Regarding the organic chemistry, it was long believed that compounds obtained from living organisms were too complex to be obtained synthetically. According to the concept of vitalism, organic matter was endowed with a "vital force" and distinguished from inorganic materials. This separation was ended however by the synthesis of urea from inorganic precursors by Friedrich Wöhler in 1828. Other chemists who brought major contributions to organic chemistry include Alexander William Williamson with his synthesis of ethers and Christopher Kelk Ingold, who, among many discoveries, established the mechanisms of substitution reactions. The general characteristics of chemical reactions are: Chemical equations are used to graphically illustrate chemical reactions. They consist of chemical or structural formulas of the reactants on the left and those of the products on the right. They are separated by an arrow (→) which indicates the direction and type of the reaction; the arrow is read as the word "yields". The tip of the arrow points in the direction in which the reaction proceeds. A double arrow (⇌) pointing in opposite directions is used for equilibrium reactions. Equations should be balanced according to the stoichiometry, the number of atoms of each species should be the same on both sides of the equation. This is achieved by scaling the number of involved molecules (A, B, C and D in a schematic example below) by the appropriate integers a, b, c and d. More elaborate reactions are represented by reaction schemes, which in addition to starting materials and products show important intermediates or transition states. Also, some relatively minor additions to the reaction can be indicated above the reaction arrow; examples of such additions are water, heat, illumination, a catalyst, etc. Similarly, some minor products can be placed below the arrow, often with a minus sign. Retrosynthetic analysis can be applied to design a complex synthesis reaction. Here the analysis starts from the products, for example by splitting selected chemical bonds, to arrive at plausible initial reagents. A special arrow (⇒) is used in retro reactions. The elementary reaction is the smallest division into which a chemical reaction can be decomposed, it has no intermediate products. Most experimentally observed reactions are built up from many elementary reactions that occur in parallel or sequentially. The actual sequence of the individual elementary reactions is known as reaction mechanism. An elementary reaction involves a few molecules, usually one or two, because of the low probability for several molecules to meet at a certain time. The most important elementary reactions are unimolecular and bimolecular reactions. Only one molecule is involved in a unimolecular reaction; it is transformed by isomerization or a dissociation into one or more other molecules. Such reactions require the addition of energy in the form of heat or light. A typical example of a unimolecular reaction is the cis–trans isomerization, in which the cis-form of a compound converts to the trans-form or vice versa. In a typical dissociation reaction, a bond in a molecule splits (ruptures) resulting in two molecular fragments. The splitting can be homolytic or heterolytic. In the first case, the bond is divided so that each product retains an electron and becomes a neutral radical. In the second case, both electrons of the chemical bond remain with one of the products, resulting in charged ions. Dissociation plays an important role in triggering chain reactions, such as hydrogen–oxygen or polymerization reactions. For bimolecular reactions, two molecules collide and react with each other. Their merger is called chemical synthesis or an addition reaction. Another possibility is that only a portion of one molecule is transferred to the other molecule. This type of reaction occurs, for example, in redox and acid-base reactions. In redox reactions, the transferred particle is an electron, whereas in acid-base reactions it is a proton. This type of reaction is also called metathesis. for example Most chemical reactions are reversible; that is, they can and do run in both directions. The forward and reverse reactions are competing with each other and differ in reaction rates. These rates depend on the concentration and therefore change with the time of the reaction: the reverse rate gradually increases and becomes equal to the rate of the forward reaction, establishing the so-called chemical equilibrium. The time to reach equilibrium depends on parameters such as temperature, pressure, and the materials involved, and is determined by the minimum free energy. In equilibrium, the Gibbs free energy must be zero. The pressure dependence can be explained with the Le Chatelier's principle. For example, an increase in pressure due to decreasing volume causes the reaction to shift to the side with fewer moles of gas. The reaction yield stabilizes at equilibrium but can be increased by removing the product from the reaction mixture or changed by increasing the temperature or pressure. A change in the concentrations of the reactants does not affect the equilibrium constant but does affect the equilibrium position. Chemical reactions are determined by the laws of thermodynamics. Reactions can proceed by themselves if they are exergonic, that is if they release free energy. The associated free energy change of the reaction is composed of the changes of two different thermodynamic quantities, enthalpy and entropy: Reactions can be exothermic, where ΔH is negative and energy is released. Typical examples of exothermic reactions are combustion, precipitation and crystallization, in which ordered solids are formed from disordered gaseous or liquid phases. In contrast, in endothermic reactions, heat is consumed from the environment. This can occur by increasing the entropy of the system, often through the formation of gaseous or dissolved reaction products, which have higher entropy. Since the entropy term in the free-energy change increases with temperature, many endothermic reactions preferably take place at high temperatures. On the contrary, many exothermic reactions such as crystallization occur preferably at lower temperatures. A change in temperature can sometimes reverse the sign of the enthalpy of a reaction, as for the carbon monoxide reduction of molybdenum dioxide: This reaction to form carbon dioxide and molybdenum is endothermic at low temperatures, becoming less so with increasing temperature. ΔH° is zero at 1855 K, and the reaction becomes exothermic above that temperature. Changes in temperature can also reverse the direction tendency of a reaction. For example, the water gas shift reaction is favored by low temperatures, but its reverse is favored by high temperatures. The shift in reaction direction tendency occurs at 1100 K. Reactions can also be characterized by their internal energy change, which takes into account changes in the entropy, volume and chemical potentials. The latter depends, among other things, on the activities of the involved substances. The speed at which reactions take place is studied by reaction kinetics. The rate depends on various parameters, such as: Several theories allow calculating the reaction rates at the molecular level. This field is referred to as reaction dynamics. The rate v of a first-order reaction, which could be the disintegration of a substance A, is given by: Its integration yields: Here k is the first-order rate constant, having dimension 1/time, [A](t) is the concentration at a time t and [A]0 is the initial concentration. The rate of a first-order reaction depends only on the concentration and the properties of the involved substance, and the reaction itself can be described with a characteristic half-life. More than one time constant is needed when describing reactions of higher order. The temperature dependence of the rate constant usually follows the Arrhenius equation: where Ea is the activation energy and kB is the Boltzmann constant. One of the simplest models of reaction rate is the collision theory. More realistic models are tailored to a specific problem and include the transition state theory, the calculation of the potential energy surface, the Marcus theory and the Rice–Ramsperger–Kassel–Marcus (RRKM) theory. In a synthesis reaction, two or more simple substances combine to form a more complex substance. These reactions are in the general form: Two or more reactants yielding one product is another way to identify a synthesis reaction. One example of a synthesis reaction is the combination of iron and sulfur to form iron(II) sulfide: Another example is simple hydrogen gas combined with simple oxygen gas to produce a more complex substance, such as water. A decomposition reaction is when a more complex substance breaks down into its more simple parts. It is thus the opposite of a synthesis reaction and can be written as One example of a decomposition reaction is the electrolysis of water to make oxygen and hydrogen gas: In a single displacement reaction, a single uncombined element replaces another in a compound; in other words, one element trades places with another element in a compound These reactions come in the general form of: One example of a single displacement reaction is when magnesium replaces hydrogen in water to make magnesium hydroxide and hydrogen gas: In a double displacement reaction, the anions and cations of two compounds switch places and form two entirely different compounds. These reactions are in the general form: For example, when barium chloride (BaCl2) and magnesium sulfate (MgSO4) react, the SO4 anion switches places with the 2Cl anion, giving the compounds BaSO4 and MgCl2. Another example of a double displacement reaction is the reaction of lead(II) nitrate with potassium iodide to form lead(II) iodide and potassium nitrate: According to Le Châtelier's Principle, reactions may proceed in the forward or reverse direction until they end or reach equilibrium. Reactions that proceed in the forward direction to approach equilibrium are often called spontaneous reactions, that is, Δ G {\displaystyle \Delta G} is negative, which means that if they occur at constant temperature and pressure, they decrease the Gibbs free energy of the reaction. They don't require much energy to proceed in the forward direction. Most reactions are forward reactions. Examples: Reactions that proceed in the backward direction to approach equilibrium are often called non-spontaneous reactions, that is, Δ G {\displaystyle \Delta G} is positive, which means that if they occur at constant temperature and pressure, they increase the Gibbs free energy of the reaction. They require input of energy to proceed in the forward direction. Examples include: In a combustion reaction, an element or compound reacts with an oxidant, usually oxygen, often producing energy in the form of heat or light. Combustion reactions frequently involve a hydrocarbon. For instance, the combustion of 1 mole (114 g) of octane in oxygen releases 5500 kJ. A combustion reaction can also result from carbon, magnesium or sulfur reacting with oxygen. Redox reactions can be understood in terms of the transfer of electrons from one involved species (reducing agent) to another (oxidizing agent). In this process, the former species is oxidized and the latter is reduced. Though sufficient for many purposes, these descriptions are not precisely correct. Oxidation is better defined as an increase in oxidation state of atoms and reduction as a decrease in oxidation state. In practice, the transfer of electrons will always change the oxidation state, but there are many reactions that are classed as "redox" even though no electron transfer occurs (such as those involving covalent bonds). In the following redox reaction, hazardous sodium metal reacts with toxic chlorine gas to form the ionic compound sodium chloride, or common table salt: In the reaction, sodium metal goes from an oxidation state of 0 (as it is a pure element) to +1: in other words, the sodium lost one electron and is said to have been oxidized. On the other hand, the chlorine gas goes from an oxidation of 0 (it is also a pure element) to −1: the chlorine gains one electron and is said to have been reduced. Because the chlorine is the one reduced, it is considered the electron acceptor, or in other words, induces oxidation in the sodium – thus the chlorine gas is considered the oxidizing agent. Conversely, the sodium is oxidized or is the electron donor, and thus induces a reduction in the other species and is considered the reducing agent. Which of the involved reactants would be a reducing or oxidizing agent can be predicted from the electronegativity of their elements. Elements with low electronegativities, such as most metals, easily donate electrons and oxidize – they are reducing agents. On the contrary, many oxides or ions with high oxidation numbers of their non-oxygen atoms, such as H2O2, MnO4, CrO3, Cr2O7, or OsO4, can gain one or two extra electrons and are strong oxidizing agents. For some main-group elements the number of electrons donated or accepted in a redox reaction can be predicted from the electron configuration of the reactant element. Elements try to reach the low-energy noble gas configuration, and therefore alkali metals and halogens will donate and accept one electron, respectively. Noble gases themselves are chemically inactive. The overall redox reaction can be balanced by combining the oxidation and reduction half-reactions multiplied by coefficients such that the number of electrons lost in the oxidation equals the number of electrons gained in the reduction. An important class of redox reactions are the electrolytic electrochemical reactions, where electrons from the power supply at the negative electrode are used as the reducing agent and electron withdrawal at the positive electrode as the oxidizing agent. These reactions are particularly important for the production of chemical elements, such as chlorine or aluminium. The reverse process, in which electrons are released in redox reactions and chemical energy is converted to electrical energy, is possible and used in batteries. In complexation reactions, several ligands react with a metal atom to form a coordination complex. This is achieved by providing lone pairs of the ligand into empty orbitals of the metal atom and forming dipolar bonds. The ligands are Lewis bases, they can be both ions and neutral molecules, such as carbon monoxide, ammonia or water. The number of ligands that react with a central metal atom can be found using the 18-electron rule, saying that the valence shells of a transition metal will collectively accommodate 18 electrons, whereas the symmetry of the resulting complex can be predicted with the crystal field theory and ligand field theory. Complexation reactions also include ligand exchange, in which one or more ligands are replaced by another, and redox processes which change the oxidation state of the central metal atom. In the Brønsted–Lowry acid–base theory, an acid–base reaction involves a transfer of protons (H) from one species (the acid) to another (the base). When a proton is removed from an acid, the resulting species is termed that acid's conjugate base. When the proton is accepted by a base, the resulting species is termed that base's conjugate acid. In other words, acids act as proton donors and bases act as proton acceptors according to the following equation: The reverse reaction is possible, and thus the acid/base and conjugated base/acid are always in equilibrium. The equilibrium is determined by the acid and base dissociation constants (Ka and Kb) of the involved substances. A special case of the acid-base reaction is the neutralization where an acid and a base, taken at the exact same amounts, form a neutral salt. Acid-base reactions can have different definitions depending on the acid-base concept employed. Some of the most common are: Precipitation is the formation of a solid in a solution or inside another solid during a chemical reaction. It usually takes place when the concentration of dissolved ions exceeds the solubility limit and forms an insoluble salt. This process can be assisted by adding a precipitating agent or by the removal of the solvent. Rapid precipitation results in an amorphous or microcrystalline residue and a slow process can yield single crystals. The latter can also be obtained by recrystallization from microcrystalline salts. Reactions can take place between two solids. However, because of the relatively small diffusion rates in solids, the corresponding chemical reactions are very slow in comparison to liquid and gas phase reactions. They are accelerated by increasing the reaction temperature and finely dividing the reactant to increase the contacting surface area. The reaction can take place at the solid|gas interface, surfaces at very low pressure such as ultra-high vacuum. Via scanning tunneling microscopy, it is possible to observe reactions at the solid|gas interface in real space, if the time scale of the reaction is in the correct range. Reactions at the solid|gas interface are in some cases related to catalysis. In photochemical reactions, atoms and molecules absorb energy (photons) of the illumination light and convert it into an excited state. They can then release this energy by breaking chemical bonds, thereby producing radicals. Photochemical reactions include hydrogen–oxygen reactions, radical polymerization, chain reactions and rearrangement reactions. Many important processes involve photochemistry. The premier example is photosynthesis, in which most plants use solar energy to convert carbon dioxide and water into glucose, disposing of oxygen as a side-product. Humans rely on photochemistry for the formation of vitamin D, and vision is initiated by a photochemical reaction of rhodopsin. In fireflies, an enzyme in the abdomen catalyzes a reaction that results in bioluminescence. Many significant photochemical reactions, such as ozone formation, occur in the Earth atmosphere and constitute atmospheric chemistry. In catalysis, the reaction does not proceed directly, but through a reaction with a third substance known as catalyst. Although the catalyst takes part in the reaction, forming weak bonds with reactants or intermediates, it is returned to its original state by the end of the reaction and so is not consumed. However, it can be inhibited, deactivated or destroyed by secondary processes. Catalysts can be used in a different phase (heterogeneous) or in the same phase (homogeneous) as the reactants. In heterogeneous catalysis, typical secondary processes include coking where the catalyst becomes covered by polymeric side products. Additionally, heterogeneous catalysts can dissolve into the solution in a solid-liquid system or evaporate in a solid–gas system. Catalysts can only speed up the reaction – chemicals that slow down the reaction are called inhibitors. Substances that increase the activity of catalysts are called promoters, and substances that deactivate catalysts are called catalytic poisons. With a catalyst, a reaction that is kinetically inhibited by high activation energy can take place in the circumvention of this activation energy. Heterogeneous catalysts are usually solids, powdered in order to maximize their surface area. Of particular importance in heterogeneous catalysis are the platinum group metals and other transition metals, which are used in hydrogenations, catalytic reforming and in the synthesis of commodity chemicals such as nitric acid and ammonia. Acids are an example of a homogeneous catalyst, they increase the nucleophilicity of carbonyls, allowing a reaction that would not otherwise proceed with electrophiles. The advantage of homogeneous catalysts is the ease of mixing them with the reactants, but they may also be difficult to separate from the products. Therefore, heterogeneous catalysts are preferred in many industrial processes. In organic chemistry, in addition to oxidation, reduction or acid-base reactions, a number of other reactions can take place which involves covalent bonds between carbon atoms or carbon and heteroatoms (such as oxygen, nitrogen, halogens, etc.). Many specific reactions in organic chemistry are name reactions designated after their discoverers. In a substitution reaction, a functional group in a particular chemical compound is replaced by another group. These reactions can be distinguished by the type of substituting species into a nucleophilic, electrophilic or radical substitution. In the first type, a nucleophile, an atom or molecule with an excess of electrons and thus a negative charge or partial charge, replaces another atom or part of the "substrate" molecule. The electron pair from the nucleophile attacks the substrate forming a new bond, while the leaving group departs with an electron pair. The nucleophile may be electrically neutral or negatively charged, whereas the substrate is typically neutral or positively charged. Examples of nucleophiles are hydroxide ion, alkoxides, amines and halides. This type of reaction is found mainly in aliphatic hydrocarbons, and rarely in aromatic hydrocarbon. The latter have high electron density and enter nucleophilic aromatic substitution only with very strong electron withdrawing groups. Nucleophilic substitution can take place by two different mechanisms, SN1 and SN2. In their names, S stands for substitution, N for nucleophilic, and the number represents the kinetic order of the reaction, unimolecular or bimolecular. The SN1 reaction proceeds in two steps. First, the leaving group is eliminated creating a carbocation. This is followed by a rapid reaction with the nucleophile. In the SN2 mechanisms, the nucleophile forms a transition state with the attacked molecule, and only then the leaving group is cleaved. These two mechanisms differ in the stereochemistry of the products. SN1 leads to the non-stereospecific addition and does not result in a chiral center, but rather in a set of geometric isomers (cis/trans). In contrast, a reversal (Walden inversion) of the previously existing stereochemistry is observed in the SN2 mechanism. Electrophilic substitution is the counterpart of the nucleophilic substitution in that the attacking atom or molecule, an electrophile, has low electron density and thus a positive charge. Typical electrophiles are the carbon atom of carbonyl groups, carbocations or sulfur or nitronium cations. This reaction takes place almost exclusively in aromatic hydrocarbons, where it is called electrophilic aromatic substitution. The electrophile attack results in the so-called σ-complex, a transition state in which the aromatic system is abolished. Then, the leaving group, usually a proton, is split off and the aromaticity is restored. An alternative to aromatic substitution is electrophilic aliphatic substitution. It is similar to the nucleophilic aliphatic substitution and also has two major types, SE1 and SE2 In the third type of substitution reaction, radical substitution, the attacking particle is a radical. This process usually takes the form of a chain reaction, for example in the reaction of alkanes with halogens. In the first step, light or heat disintegrates the halogen-containing molecules producing radicals. Then the reaction proceeds as an avalanche until two radicals meet and recombine. The addition and its counterpart, the elimination, are reactions that change the number of substituents on the carbon atom, and form or cleave multiple bonds. Double and triple bonds can be produced by eliminating a suitable leaving group. Similar to the nucleophilic substitution, there are several possible reaction mechanisms that are named after the respective reaction order. In the E1 mechanism, the leaving group is ejected first, forming a carbocation. The next step, the formation of the double bond, takes place with the elimination of a proton (deprotonation). The leaving order is reversed in the E1cb mechanism, that is the proton is split off first. This mechanism requires the participation of a base. Because of the similar conditions, both reactions in the E1 or E1cb elimination always compete with the SN1 substitution. The E2 mechanism also requires a base, but there the attack of the base and the elimination of the leaving group proceed simultaneously and produce no ionic intermediate. In contrast to the E1 eliminations, different stereochemical configurations are possible for the reaction product in the E2 mechanism, because the attack of the base preferentially occurs in the anti-position with respect to the leaving group. Because of the similar conditions and reagents, the E2 elimination is always in competition with the SN2-substitution. The counterpart of elimination is an addition where double or triple bonds are converted into single bonds. Similar to substitution reactions, there are several types of additions distinguished by the type of the attacking particle. For example, in the electrophilic addition of hydrogen bromide, an electrophile (proton) attacks the double bond forming a carbocation, which then reacts with the nucleophile (bromine). The carbocation can be formed on either side of the double bond depending on the groups attached to its ends, and the preferred configuration can be predicted with the Markovnikov's rule. This rule states that "In the heterolytic addition of a polar molecule to an alkene or alkyne, the more electronegative (nucleophilic) atom (or part) of the polar molecule becomes attached to the carbon atom bearing the smaller number of hydrogen atoms." If the addition of a functional group takes place at the less substituted carbon atom of the double bond, then the electrophilic substitution with acids is not possible. In this case, one has to use the hydroboration–oxidation reaction, wherein the first step, the boron atom acts as electrophile and adds to the less substituted carbon atom. In the second step, the nucleophilic hydroperoxide or halogen anion attacks the boron atom. While the addition to the electron-rich alkenes and alkynes is mainly electrophilic, the nucleophilic addition plays an important role in the carbon-heteroatom multiple bonds, and especially its most important representative, the carbonyl group. This process is often associated with elimination so that after the reaction the carbonyl group is present again. It is, therefore, called an addition-elimination reaction and may occur in carboxylic acid derivatives such as chlorides, esters or anhydrides. This reaction is often catalyzed by acids or bases, where the acids increase the electrophilicity of the carbonyl group by binding to the oxygen atom, whereas the bases enhance the nucleophilicity of the attacking nucleophile. Nucleophilic addition of a carbanion or another nucleophile to the double bond of an alpha, beta-unsaturated carbonyl compound can proceed via the Michael reaction, which belongs to the larger class of conjugate additions. This is one of the most useful methods for the mild formation of C–C bonds. Some additions which can not be executed with nucleophiles and electrophiles can be succeeded with free radicals. As with the free-radical substitution, the radical addition proceeds as a chain reaction, and such reactions are the basis of the free-radical polymerization. In a rearrangement reaction, the carbon skeleton of a molecule is rearranged to give a structural isomer of the original molecule. These include hydride shift reactions such as the Wagner-Meerwein rearrangement, where a hydrogen, alkyl or aryl group migrates from one carbon to a neighboring carbon. Most rearrangements are associated with the breaking and formation of new carbon-carbon bonds. Other examples are sigmatropic reaction such as the Cope rearrangement. Cyclic rearrangements include cycloadditions and, more generally, pericyclic reactions, wherein two or more double bond-containing molecules form a cyclic molecule. An important example of cycloaddition reaction is the Diels–Alder reaction (the so-called [4+2] cycloaddition) between a conjugated diene and a substituted alkene to form a substituted cyclohexene system. Whether a certain cycloaddition would proceed depends on the electronic orbitals of the participating species, as only orbitals with the same sign of wave function will overlap and interact constructively to form new bonds. Cycloaddition is usually assisted by light or heat. These perturbations result in a different arrangement of electrons in the excited state of the involved molecules and therefore in different effects. For example, the [4+2] Diels-Alder reactions can be assisted by heat whereas the [2+2] cycloaddition is selectively induced by light. Because of the orbital character, the potential for developing stereoisomeric products upon cycloaddition is limited, as described by the Woodward–Hoffmann rules. Biochemical reactions are mainly controlled by enzymes. These proteins can specifically catalyze a single reaction so that reactions can be controlled very precisely. The reaction takes place in the active site, a small part of the enzyme which is usually found in a cleft or pocket lined by amino acid residues, and the rest of the enzyme is used mainly for stabilization. The catalytic action of enzymes relies on several mechanisms including the molecular shape ("induced fit"), bond strain, proximity and orientation of molecules relative to the enzyme, proton donation or withdrawal (acid/base catalysis), electrostatic interactions and many others. The biochemical reactions that occur in living organisms are collectively known as metabolism. Among the most important of its mechanisms is the anabolism, in which different DNA and enzyme-controlled processes result in the production of large molecules such as proteins and carbohydrates from smaller units. Bioenergetics studies the sources of energy for such reactions. Important energy sources are glucose and oxygen, which can be produced by plants via photosynthesis or assimilated from food and air, respectively. All organisms use this energy to produce adenosine triphosphate (ATP), which can then be used to energize other reactions. Chemical reactions are central to chemical engineering, where they are used for the synthesis of new compounds from natural raw materials such as petroleum, mineral ores, and oxygen in air. It is essential to make the reaction as efficient as possible, maximizing the yield and minimizing the number of reagents, energy inputs and waste. Catalysts are especially helpful for reducing the energy required for the reaction and increasing its reaction rate. Some specific reactions have their niche applications. For example, the thermite reaction is used to generate light and heat in pyrotechnics and welding. Although it is less controllable than the more conventional oxy-fuel welding, arc welding and flash welding, it requires much less equipment and is still used to mend rails, especially in remote areas. Mechanisms of monitoring chemical reactions depend strongly on the reaction rate. Relatively slow processes can be analyzed in situ for the concentrations and identities of the individual ingredients. Important tools of real-time analysis are the measurement of pH and analysis of optical absorption (color) and emission spectra. A less accessible but rather efficient method is the introduction of a radioactive isotope into the reaction and monitoring how it changes over time and where it moves to; this method is often used to analyze the redistribution of substances in the human body. Faster reactions are usually studied with ultrafast laser spectroscopy where utilization of femtosecond lasers allows short-lived transition states to be monitored at a time scaled down to a few femtoseconds.
[[File:Geometry Rectangle.svg|thumb|A rectangle]] In [[geometry]], a '''rectangle''' is a [[shape]] with four [[side]]s and four corners. The corners are all [[right angle]]s. It follows that the pairs of sides opposite each other must be [[Parallel (geometry)|parallel]] and of the same length. People make many rectangular things, including most [[Table (furniture)|tables]], boxes, books, and papers. The word comes from Latin words meaning "right" and angle". A rectangle whose four sides have the same length is called a [[square (geometry)|square]]. == Formulas == [[File:PerimeterRectangle.svg|thumb|150px|The formula for the perimeter of a rectangle.]] If a rectangle has length <math>\ell</math> and width ''w'', then'':'' * It has [[area]] <math>K = \ell w</math>. * it has [[perimeter]] <math>P = 2\ell + 2w = 2(\ell + w)\,</math>.<ref>{{Cite web|last=|first=|date=2020-04-17|title=List of Geometry and Trigonometry Symbols|url=https://mathvault.ca/hub/higher-math/math-symbols/geometry-trigonometry-symbols/|archive-url=|archive-date=|access-date=2020-09-25|website=Math Vault|language=en-US}}</ref> * Each diagonal has length <math>d=\sqrt{\ell^2 + w^2}</math>.<ref name=":0">{{Cite web|title=Rectangle|url=https://www.mathsisfun.com/geometry/rectangle.html|access-date=2020-09-25|website=www.mathsisfun.com}}</ref><ref>{{Cite web|last=Weisstein|first=Eric W.|title=Rectangle|url=https://mathworld.wolfram.com/Rectangle.html|access-date=2020-09-25|website=mathworld.wolfram.com|language=en}}</ref> * it has an aspect ratio of <math>\ell</math> : w. * When <math>\ell = w\,</math>, the rectangle is a [[Square (geometry)|square]]. * When <math>\ell / w</math> is equal to the [[golden ratio]], the rectangle is called a [[golden rectangle]].<ref name=":0" /> == Related pages == * [[Matrix (mathematics)]] * [[Quadrilateral]] * [[Rhombus]] == References == <references /> {{Commons category|Rectangles}} {{Math-stub}} {{shapes}} [[Category:Polygons]]
In Euclidean plane geometry, a rectangle is a quadrilateral with four right angles. It can also be defined as: an equiangular quadrilateral, since equiangular means that all of its angles are equal (360°/4 = 90°); or a parallelogram containing a right angle. A rectangle with four sides of equal length is a square. The term "oblong" is occasionally used to refer to a non-square rectangle. A rectangle with vertices ABCD would be denoted as ABCD. The word rectangle comes from the Latin rectangulus, which is a combination of rectus (as an adjective, right, proper) and angulus (angle). A crossed rectangle is a crossed (self-intersecting) quadrilateral which consists of two opposite sides of a rectangle along with the two diagonals (therefore only two sides are parallel). It is a special case of an antiparallelogram, and its angles are not right angles and not all equal, though opposite angles are equal. Other geometries, such as spherical, elliptic, and hyperbolic, have so-called rectangles with opposite sides equal in length and equal angles that are not right angles. Rectangles are involved in many tiling problems, such as tiling the plane by rectangles or tiling a rectangle by polygons. A convex quadrilateral is a rectangle if and only if it is any one of the following: A rectangle is a special case of a parallelogram in which each pair of adjacent sides is perpendicular. A parallelogram is a special case of a trapezium (known as a trapezoid in North America) in which both pairs of opposite sides are parallel and equal in length. A trapezium is a convex quadrilateral which has at least one pair of parallel opposite sides. A convex quadrilateral is De Villiers defines a rectangle more generally as any quadrilateral with axes of symmetry through each pair of opposite sides. This definition includes both right-angled rectangles and crossed rectangles. Each has an axis of symmetry parallel to and equidistant from a pair of opposite sides, and another which is the perpendicular bisector of those sides, but, in the case of the crossed rectangle, the first axis is not an axis of symmetry for either side that it bisects. Quadrilaterals with two axes of symmetry, each through a pair of opposite sides, belong to the larger class of quadrilaterals with at least one axis of symmetry through a pair of opposite sides. These quadrilaterals comprise isosceles trapezia and crossed isosceles trapezia (crossed quadrilaterals with the same vertex arrangement as isosceles trapezia). A rectangle is cyclic: all corners lie on a single circle. It is equiangular: all its corner angles are equal (each of 90 degrees). It is isogonal or vertex-transitive: all corners lie within the same symmetry orbit. It has two lines of reflectional symmetry and rotational symmetry of order 2 (through 180°). The dual polygon of a rectangle is a rhombus, as shown in the table below. A rectangle is a rectilinear polygon: its sides meet at right angles. A rectangle in the plane can be defined by five independent degrees of freedom consisting, for example, of three for position (comprising two of translation and one of rotation), one for shape (aspect ratio), and one for overall size (area). Two rectangles, neither of which will fit inside the other, are said to be incomparable. If a rectangle has length ℓ {\displaystyle \ell } and width w {\displaystyle w} The isoperimetric theorem for rectangles states that among all rectangles of a given perimeter, the square has the largest area. The midpoints of the sides of any quadrilateral with perpendicular diagonals form a rectangle. A parallelogram with equal diagonals is a rectangle. The Japanese theorem for cyclic quadrilaterals states that the incentres of the four triangles determined by the vertices of a cyclic quadrilateral taken three at a time form a rectangle. The British flag theorem states that with vertices denoted A, B, C, and D, for any point P on the same plane of a rectangle: For every convex body C in the plane, we can inscribe a rectangle r in C such that a homothetic copy R of r is circumscribed about C and the positive homothety ratio is at most 2 and 0.5 × Area ( R ) ≤ Area ( C ) ≤ 2 × Area ( r ) {\displaystyle 0.5{\text{ × Area}}(R)\leq {\text{Area}}(C)\leq 2{\text{ × Area}}(r)} . A crossed quadrilateral (self-intersecting) consists of two opposite sides of a non-self-intersecting quadrilateral along with the two diagonals. Similarly, a crossed rectangle is a crossed quadrilateral which consists of two opposite sides of a rectangle along with the two diagonals. It has the same vertex arrangement as the rectangle. It appears as two identical triangles with a common vertex, but the geometric intersection is not considered a vertex. A crossed quadrilateral is sometimes likened to a bow tie or butterfly, sometimes called an "angular eight". A three-dimensional rectangular wire frame that is twisted can take the shape of a bow tie. The interior of a crossed rectangle can have a polygon density of ±1 in each triangle, dependent upon the winding orientation as clockwise or counterclockwise. A crossed rectangle may be considered equiangular if right and left turns are allowed. As with any crossed quadrilateral, the sum of its interior angles is 720°, allowing for internal angles to appear on the outside and exceed 180°. A rectangle and a crossed rectangle are quadrilaterals with the following properties in common: In spherical geometry, a spherical rectangle is a figure whose four edges are great circle arcs which meet at equal angles greater than 90°. Opposite arcs are equal in length. The surface of a sphere in Euclidean solid geometry is a non-Euclidean surface in the sense of elliptic geometry. Spherical geometry is the simplest form of elliptic geometry. In elliptic geometry, an elliptic rectangle is a figure in the elliptic plane whose four edges are elliptic arcs which meet at equal angles greater than 90°. Opposite arcs are equal in length. In hyperbolic geometry, a hyperbolic rectangle is a figure in the hyperbolic plane whose four edges are hyperbolic arcs which meet at equal angles less than 90°. Opposite arcs are equal in length. The rectangle is used in many periodic tessellation patterns, in brickwork, for example, these tilings: A rectangle tiled by squares, rectangles, or triangles is said to be a "squared", "rectangled", or "triangulated" (or "triangled") rectangle respectively. The tiled rectangle is perfect if the tiles are similar and finite in number and no two tiles are the same size. If two such tiles are the same size, the tiling is imperfect. In a perfect (or imperfect) triangled rectangle the triangles must be right triangles. A database of all known perfect rectangles, perfect squares and related shapes can be found at squaring.net. The lowest number of squares need for a perfect tiling of a rectangle is 9 and the lowest number needed for a perfect tilling a square is 21, found in 1978 by computer search. A rectangle has commensurable sides if and only if it is tileable by a finite number of unequal squares. The same is true if the tiles are unequal isosceles right triangles. The tilings of rectangles by other tiles which have attracted the most attention are those by congruent non-rectangular polyominoes, allowing all rotations and reflections. There are also tilings by congruent polyaboloes.
[[File:Trafalgar Fountain July 06.jpg|thumb|right|230px|The Trafalgar fountain]] '''Trafalgar Square''' is in the middle of [[London]] and is a large pedestrian square that is bounded on three sides by roads. It serves as a refuge and a major [[traffic]] intersection. Important roads go from the square: [[Whitehall]] goes to [[Palace of Westminster|Parliament]], the [[Mall]] goes to [[Buckingham Palace]] and the [[Strand, London|Strand]] goes to the [[City of London]]. The square is also close to [[Covent Garden]] and [[Charing Cross tube station|Charing Cross station]] and [[Charing Cross railway station|Charing Cross station]].<ref>Mace, Rodney 2005. ''Trafalgar Square: Emblem of Empire'' (2nd ed). London: Lawrence and Wishart. ISBN 978-1-905007-11-0</ref> More than 15 million people go to visit there every year. It contains a large statue of Admiral [[Horatio Nelson]]. The square celebrates the [[Battle of Trafalgar]], which was fought in 1805, and contains [[Nelson's Column]], a statue of Nelson mounted on a tall column, with four statues of [[lion]]s around it. The column is 56 m tall, and the statue is 5 m tall. The [[National Gallery, London|National Art Gallery]] is one of several important buildings facing the Square. A point in Trafalgar Square is regarded as the official centre of London in [[legislation]], and when measuring distances from the capital.<ref>[https://www.bbc.co.uk/london/content/articles/2005/08/15/charingcross_feature.shtml Where Is The Centre Of London?] {{Webarchive|url=https://web.archive.org/web/20100817141948/http://www.bbc.co.uk/london/content/articles/2005/08/15/charingcross_feature.shtml |date=17 August 2010 }} BBC</ref> == Uses == The square is visited by many tourists, mainly for pleasure and relaxation. There are, however, meetings and [[demonstration]]s in the square. When the square was first built, demonstrations were banned. The ban lasted until the 1880s, when the new [[labour movement]] started to hold demonstrations. One group that did so was the Social Democratic Federation. On "Black Monday", on 6 February 1886, there was a major demonstration about [[unemployment]] that led to a [[riot]] in [[Pall Mall]]. There were demonstrations in the 1980s against [[South Africa]]n [[apartheid]]. In 1990, there were riots against the [[poll tax]]. In the 2000s, there were demonstrations against the [[Iraq war|Iraq War]]. In recent years the square has become a gathering place for celebrations. When [[England]] won the [[Rugby World Cup]] in 2003, thousands of fans gathered in the square. Public festivities again happened when London won its bid to hold the 2012 [[Summer Olympic Games]]. The square was also the scene of a large gathering after the [[7 July 2005 London bombings|terrorist bombings in London]] on 7 July 2005. == Past == The Square was last dug up in the 19th century in preparation to its present format. Discovered were the skeletons of [[lion]], [[woolly mammoth]], [[hippopotamus]], and [[Hyaena|hyaenas]]. Very similar species were found in [[East London]] when [[Ilford]] was dug up. The warm stage lasted from 125,000 to 24,000 years ago before the colder climate prevailed until modern times.<ref>National Geographic 2022. ''Hidden London'' issue, p8/9. </ref> ==References== {{reflist}} {{authority control}} <!-- Interwiki -->[[Category:National squares]] [[Category:Town squares in Europe]] [[Category:Buildings and structures in the City of Westminster]]
Trafalgar Square (/trəˈfælɡər/ trə-FAL-gər) is a public square in the City of Westminster, Central London, established in the early 19th century around the area formerly known as Charing Cross. The square's name commemorates the Battle of Trafalgar, the British naval victory in the Napoleonic Wars over France and Spain that took place on 21 October 1805 off the coast of Cape Trafalgar. The site around Trafalgar Square had been a significant landmark since the 1200s. For centuries, distances measured from Charing Cross have served as location markers. The site of the present square formerly contained the elaborately designed, enclosed courtyard, King's Mews. After George IV moved the mews to Buckingham Palace, the area was redeveloped by John Nash, but progress was slow after his death, and the square did not open until 1844. The 169-foot (52 m) Nelson's Column at its centre is guarded by four lion statues. A number of commemorative statues and sculptures occupy the square, but the Fourth Plinth, left empty since 1840, has been host to contemporary art since 1999. Prominent buildings facing the square include the National Gallery, St Martin-in-the-Fields, Canada House, and South Africa House. The square has been used for community gatherings and political demonstrations, including Bloody Sunday in 1887, the culmination of the first Aldermaston March, anti-war protests, and campaigns against climate change. A Christmas tree has been donated to the square by Norway since 1947 and is erected for twelve days before and after Christmas Day. The square is a centre of annual celebrations on New Year's Eve. It was well known for its feral pigeons until their removal in the early 21st century. The square is named after the Battle of Trafalgar, a British naval victory in the Napoleonic Wars with France and Spain that took place on 21 October 1805 off the coast of Cape Trafalgar, southwest Spain, although it was not named as such until 1835. The name "Trafalgar" is a Spanish word of Arabic origin, derived from either Taraf al-Ghar (طرف الغار 'cape of the cave/laurel') or Taraf al-Gharb (طرف الغرب 'extremity of the west'). Trafalgar Square is owned by the King in Right of the Crown and managed by the Greater London Authority, while Westminster City Council owns the roads around the square, including the pedestrianised area of the North Terrace. The square contains a large central area with roadways on three sides and a terrace to the north, in front of the National Gallery. The roads around the square form part of the A4, a major road running west of the City of London. Originally having roadways on all four sides, traffic travelled in both directions around the square until a one-way clockwise gyratory system was introduced on 26 April 1926. Works completed in 2003 reduced the width of the roads and closed the northern side to traffic. Nelson's Column is in the centre of the square, flanked by fountains designed by Sir Edwin Lutyens between 1937 and 1939 (replacements for two of Peterhead granite, now in Canada) and guarded by four monumental bronze lions sculpted by Sir Edwin Landseer. At the top of the column is a statue of Horatio Nelson, who commanded the British Navy at the Battle of Trafalgar. Surrounding the square are the National Gallery on the north side and St Martin-in-the-Fields Church to the east. Also on the east is South Africa House, and facing it across the square is Canada House. To the south west is The Mall, which leads towards Buckingham Palace via Admiralty Arch, while Whitehall is to the south and the Strand to the east. Charing Cross Road passes between the National Gallery and the church. London Underground's Charing Cross station on the Northern and Bakerloo lines has an exit in the square. The lines had separate stations, of which the Bakerloo line one was called Trafalgar Square until they were linked and renamed in 1979 as part of the construction of the Jubilee line, which was rerouted to Westminster in 1999. Other nearby tube stations are Embankment connecting the District, Circle, Northern and Bakerloo lines, and Leicester Square on the Northern and Piccadilly lines. London bus routes 3, 6, 9, 11, 12, 13, 15, 23, 24, 29, 53, 87, 88, 91, 139, 159, 176, 453 are only some among the bus routes that pass through Trafalgar Square. A point in Trafalgar Square is regarded as the official centre of London in legislation and when measuring distances from the capital. Building work on the south side of the square in the late 1950s revealed deposits from the last interglacial period. Among the findings were the remains of cave lions, rhinoceroses, straight-tusked elephants and hippopotami. The site has been significant since the 13th century. During Edward I's reign it hosted the King's Mews, running north from the T-junction in the south, Charing Cross, where the Strand from the City meets Whitehall coming north from Westminster. From the reign of Richard II to that of Henry VII, the mews was at the western end of the Strand. The name "Royal Mews" comes from the practice of keeping hawks here for moulting; "mew" is an old word for this. After a fire in 1534, the mews were rebuilt as stables, and remained here until George IV moved them to Buckingham Palace. After 1732, the King's Mews were divided into the Great Mews and the smaller Green Mews to the north by the Crown Stables, a large block, built to the designs of William Kent. Its site is occupied by the National Gallery. In 1826 the Commissioners of Woods, Forests and Land Revenues instructed John Nash to draw up plans for clearing a large area south of Kent's stable block, and as far east as St Martin's Lane. His plans left open the whole area of what became Trafalgar Square, except for a block in the centre, which he reserved for a new building for the Royal Academy of Arts. The plans included the demolition and redevelopment of buildings between St Martin's Lane and the Strand and the construction of a road (now called Duncannon Street) across the churchyard of St Martin-in-the-Fields. The Charing Cross Act was passed in 1826 and clearance started soon after. Nash died soon after construction started, impeding its progress. The square was to be named for William IV commemorating his ascent to the throne in 1830. Around 1835, it was decided that the square would be named after the Battle of Trafalgar as suggested by architect George Ledwell Taylor, commemorating Nelson's victory over the French and Spanish in 1805 during the Napoleonic Wars. After the clearance, development progressed slowly. The National Gallery was built on the north side between 1832 and 1838 to a design by William Wilkins, and in 1837 the Treasury approved Wilkins' plan for the laying out of the square, but it was not put into effect. In April 1840, following Wilkins' death, new plans by Charles Barry were accepted, and construction started within weeks. For Barry, as for Wilkins, a major consideration was increasing the visual impact of the National Gallery, which had been widely criticised for its lack of grandeur. He dealt with the complex sloping site by excavating the main area to the level of the footway between Cockspur Street and the Strand, and constructing a 15-foot (4.6 m) high balustraded terrace with a roadway on the north side, and steps at each end leading to the main level. Wilkins had proposed a similar solution with a central flight of steps. All the stonework was of Aberdeen granite. In 1845, four Bude-Lights with octagonal glass lanterns were installed. Two, opposite the National Gallery, are on tall bronze columns, and two, in the south-west and south-east corners of the square, on shorter bronze columns on top of wider granite columns. They were designed by Barry and manufactured by Stevens and Son, of Southwark. In 1841 it was decided that two fountains should be included in the layout. The estimated budget, excluding paving and sculptures, was £11,000. The earth removed was used to level Green Park. The square was originally surfaced with tarmacadam, which was replaced with stone in the 1920s. Trafalgar Square was opened to the public on 1 May 1844. Nelson's Column was planned independently of Barry's work. In 1838 a Nelson Memorial Committee had approached the government proposing that a monument to the victory of Trafalgar, funded by public subscription, should be erected in the square. A competition was held and won by the architect William Railton, who proposed a 218-foot-3-inch (66.52 m) Corinthinan column topped by a statue of Nelson and guarded by four sculpted lions. The design was approved, but received widespread objections from the public. Construction went ahead beginning in 1840 but with the height reduced to 145 feet 3 inches (44.27 m). The column was completed and the statue raised in November 1843. The last of the bronze reliefs on the column's pedestals was not completed until May 1854, and the four lions, although part of the original design, were only added in 1867. Each lion weighs seven tons. A hoarding remained around the base of Nelson's Column for some years and some of its upper scaffolding remained in place. Landseer, the sculptor, had asked for a lion that had died at the London Zoo to be brought to his studio. He took so long to complete sketches that its corpse began to decompose and some parts had to be improvised. The statues have paws that resemble cats more than lions. Barry was unhappy about Nelson's Column being placed in the square. In July 1840, when its foundations had been laid, he told a parliamentary select committee that "it would in my opinion be desirable that the area should be wholly free from all insulated objects of art". In 1940 the Nazi SS developed secret plans to transfer Nelson's Column to Berlin after an expected German invasion, as related by Norman Longmate in If Britain Had Fallen (1972). The square has been Grade I listed on the Register of Historic Parks and Gardens since 1996. The square was the target of two suffragette bombings in 1913 and 1914. This was as part of the suffragette bombing and arson campaign of 1912–1914, in which suffragettes carried out a series of politically-motivated bombing and arson attacks nationwide as part of their campaign for women's suffrage. The first attack occurred on 15 May 1913. A bomb was planted in the public area outside the National Gallery, but failed to explode. A second attack occurred at St Martin-in-the-Fields church at the north-east corner of the square one 4 April 1914. A bomb exploded inside the church, blowing out the windows and showering passers-by with broken glass. The bomb then started a fire. In the aftermath, a mass of people rushed to the scene, many of whom aggressively expressed their anger towards the suffragettes. Churches were a particular target during the campaign, as it was believed that the Church of England was complicit in reinforcing opposition to women's suffrage. Between 1913 and 1914, 32 churches were attacked nationwide. In the weeks after the bombing, there were also attacks on Westminster Abbey and St Paul's Cathedral. A major 18-month redevelopment of the square led by W.S. Atkins with Foster and Partners as sub-consultants was completed in 2003. The work involved closing the eastbound road along the north side and diverting traffic around the other three sides of the square, demolishing the central section of the northern retaining wall and inserting a wide set of steps to the pedestrianised terrace in front of the National Gallery. The construction includes two lifts for disabled access, public toilets and a café. Access between the square and the gallery had been by two crossings at the northeast and northwest corners. Barry's scheme provided two plinths for sculptures on the north side of the square. A bronze equestrian statue of George IV was designed by Sir Francis Chantrey and Thomas Earle. It was originally intended to be placed on top of the Marble Arch, but instead was installed on the eastern plinth in 1843, while the other plinths remained empty until late in the 20th century. There are two other statues on plinths, both installed during the 19th century: General Sir Charles James Napier by George Cannon Adams in the south-west corner in 1855, and Major-General Sir Henry Havelock by William Behnes in the south-east in 1861. In 2000, the Mayor of London, Ken Livingstone, suggested replacing the statues with figures more familiar to the general public. In the 21st century, the empty plinth in the north-west corner of the square, the "Fourth Plinth", has been used to show specially commissioned temporary artworks. The scheme was initiated by the Royal Society of Arts and continued by the Fourth Plinth Commission, appointed by the Mayor of London. There are three busts of admirals against the north wall of the square. Those of John Jellicoe, 1st Earl Jellicoe (by Sir Charles Wheeler) and David Beatty, 1st Earl Beatty (by William MacMillan) were installed in 1948 in conjunction with the square's fountains, which also commemorate them. The third, of the Second World War First Sea Lord Andrew Cunningham, 1st Viscount Cunningham of Hyndhope (by Franta Belsky) was unveiled alongside them on 2 April 1967. On the south side of Trafalgar Square, on the site of the original Charing Cross, is a bronze equestrian statue of Charles I by Hubert Le Sueur. It was cast in 1633, and placed in its present position in 1678. The two statues on the lawn in front of the National Gallery are the statue of James II (designed by Peter van Dievoet and Laurens van der Meulen for the studio of Grinling Gibbons) to the west of the portico, and of one George Washington, a replica of a work by Jean-Antoine Houdon, to the east. The latter was a gift from the Commonwealth of Virginia, installed in 1921. Two statues erected in the 19th century have since been removed. One of Edward Jenner, pioneer of the smallpox vaccine, was set up in the south-west corner of the square in 1858, next to that of Napier. Sculpted by William Calder Marshall, it showed Jenner sitting in a chair in a relaxed pose, and was inaugurated at a ceremony presided over by Prince Albert. It was moved to Kensington Gardens in 1862. The other, of General Charles George Gordon by Hamo Thornycroft, was erected on an 18-foot high pedestal between the fountains in 1888. It was removed in 1943 and re-sited on the Victoria Embankment ten years later. In 1841, following suggestions from the local paving board, Barry agreed that two fountains should be installed to counteract the effects of reflected heat and glare from the asphalt surface. The First Commissioner of Woods and Forests welcomed the plan because the fountains reduced the open space available for public gatherings and reduced the risk of riotous assembly. The fountains were fed from two wells, one in front of the National Gallery and one behind it connected by a tunnel. Water was pumped to the fountains by a steam engine housed in a building behind the gallery. In the late-1930s it was decided to replace the pump and the centrepieces of the fountains. The new centrepieces, designed by Sir Edwin Lutyens, were memorials to Lord Jellicoe and Lord Beatty, although busts of the admirals, initially intended to be placed in the fountain surrounds were placed against the northern retaining wall when the project was completed after the Second World War. The fountains cost almost £50,000. The old ones were presented to the Canadian government and are now located in Ottawa's Confederation Park and Regina's Wascana Centre. A programme of restoration was completed by May 2009. The pump system was replaced with one capable of sending an 80-foot (24 m) jet of water into the air. A LED lighting system that can project different combinations of colours on to the fountains was installed to reduce the cost of lighting maintenance and to coincide with the 2012 Summer Olympics. The square was once famous for feral pigeons and feeding them was a popular activity. Pigeons began flocking to the square before construction was completed and feed sellers became well known in the Victorian era. The desirability of the birds' presence was contentious: their droppings disfigured the stonework and the flock, estimated at its peak to be 35,000, was considered a health hazard. A stall seller, Bernie Rayner, infamously sold bird seed to tourists at inflated prices. In February 2001, the sale of bird seed in the square was stopped and other measures were introduced to discourage the pigeons including the use of birds of prey. Supporters continued to feed the birds but in 2003 the mayor, Ken Livingstone, enacted bylaws to ban feeding them in the square. In September 2007 Westminster City Council passed further bylaws banning feeding birds on the pedestrianised North Terrace and other pavements in the area. Nelson's column was repaired from years of damage from pigeon droppings at a cost of £140,000. For many years, revellers celebrating the New Year have gathered in the square despite a lack of celebrations being arranged. The lack of official events was partly because the authorities were concerned that encouraging more partygoers would cause overcrowding. Since 2003, a firework display centred on the London Eye and South Bank of the Thames has been provided as an alternative. Since 2014, New Year celebrations have been organised by the Greater London Authority in conjunction with the charity Unicef, who began ticketing the event to control crowd numbers. The fireworks display has been cancelled during the COVID-19 pandemic, with an event due to take place in the Square to see in 2022. However the event was cancelled during the spread of the SARS-Cov-2 Omicron variant. A Christmas ceremony has been held in the square every year since 1947. A Norway spruce (or sometimes a fir) is presented by Norway's capital city, Oslo as London's Christmas tree, a token of gratitude for Britain's support during World War II. (Besides war-time support, Norway's Prince Olav and the country's government lived in exile in London throughout the war.) The Christmas tree is decorated with lights that are switched on at a seasonal ceremony. It is usually held twelve days before Christmas Day. The festivity is open to the public and attracts a large number of people. The switch-on is usually followed by several nights of Christmas carol singing and other performances and events. On the twelfth night of Christmas, the tree is taken down for recycling. Westminster City Council threatened to abandon the event to save £5,000 in 1980 but the decision was reversed. The tree is selected by the Head Forester from Oslo's municipal forest and shipped, across the North Sea to the Port of Felixstowe, then by road to Trafalgar Square. The first tree was 48 feet (15 m) tall, but more recently has been around 75 feet (23 m). In 1987, protesters chained themselves to the tree. In 1990, a man sawed into the tree with a chainsaw a few hours before a New Year's Eve party was scheduled to take place. He was arrested and the tree was repaired by tree surgeons who removed gouged sections from the trunk while the tree was suspended from a crane. The square has become a social and political focus for visitors and Londoners, developing over its history from "an esplanade peopled with figures of national heroes, into the country's foremost place politique", as historian Rodney Mace has written. Since its construction, it has been a venue for political demonstrations. The great Chartist rally in 1848, a campaign for social reform by the working class began in the square. A ban on political rallies remained in effect until the 1880s, when the emerging Labour movement, particularly the Social Democratic Federation, began holding protests. On 8 February 1886 (also known as "Black Monday"), protesters rallied against unemployment leading to a riot in Pall Mall. A larger riot ("Bloody Sunday") occurred in the square on 13 November 1887. The Campaign for Nuclear Disarmament's first Aldermaston March, protesting against the Atomic Weapons Establishment (AWE), began in the square in 1958. One of the first significant demonstrations of the modern era was held in the square on 19 September 1961 by the Committee of 100, which included the philosopher Bertrand Russell. The protesters rallied for peace and against war and nuclear weapons. In March 1968, a crowd of 10,000 demonstrated against US involvement in the Vietnam War before marching to the American Embassy in Grosvenor Square. Throughout the 1980s, a continuous anti-apartheid protest was held outside South Africa House. In 1990, the Poll Tax Riots began by a demonstration attended by 200,000 people and ultimately caused rioting in the surrounding area. More recently, there have been anti-war demonstrations opposing the Afghanistan War and the Iraq War. A large vigil was held shortly after the terrorist bombings in London on Thursday, 7 July 2005. In December 2009, participants from the Camp for Climate Action occupied the square for the two weeks during which the UN Conference on Climate Change took place in Copenhagen. It was billed as a UK base for direct action on climate change and saw various actions and protests stem from the occupation. In March 2011, the square was occupied by a crowd protesting against the UK Budget and proposed budget cuts. During the night the situation turned violent as the escalation by riot police and protesters damaged portions of the square. In November 2015 a vigil against the terrorist attacks in Paris was held. Crowds sang the French national anthem, La Marseillaise, and held banners in support of the city and country. Every year on the anniversary of the Battle of Trafalgar (21 October), the Sea Cadet Corps holds a parade in honour of Admiral Lord Nelson and the British victory over the combined fleets of Spain and France at Trafalgar. The Royal British Legion holds a Silence in the Square event on Armistice Day, 11 November, in remembrance of those who died in war. The event includes music and poetry readings, culminating in a bugler playing the Last Post and a two-minute silence at 11 am. In February 2019, hundreds of students participated in a protest against climate change as a part of the School strike for Climate campaign. The protest started in the nearby Parliament Square, and as the day went on, the demonstrators moved towards Trafalgar Square. In July 2020, two members of the protest group Animal Rebellion were arrested on suspicion for criminal damage after releasing red dye into the fountains. In September 2020 anti-lockdown protests opposed to the imposition of regulations relating to the coronavirus outbreak took place in the square. A police observation box has been in the Square since 1919, originally a wooden freestanding unit, it was replaced by hollowing out a lampstand at the southeastern corner of the Square into a permanent structure in 1928, but decommissioned in the 1970s. In the 21st century, Trafalgar Square has been the location for several sporting events and victory parades. In June 2002, 12,000 people gathered to watch England's FIFA World Cup quarter-final against Brazil on giant video screens which had been erected for the occasion. The square was used by England on 9 December 2003 to celebrate their victory in the Rugby World Cup, and on 13 September 2005 for England's victory in the Ashes series. On 6 July 2005 Trafalgar Square hosted the official watch party for London's bid to host the 2012 Summer Olympics at the 117th IOC Session in Singapore, hosted by Katy Hill and Margherita Taylor. A countdown clock was erected in March 2011, although engineering and weather-related faults caused it to stop a day later. In 2007, it hosted the opening ceremonies of the Tour de France and was part of the course for subsequent races. The Sea Cadets hold a yearly Battle of Trafalgar victory parade running the north of Whitehall, from Horse Guard's Parade to Nelson's Column. As an archetypal London location, Trafalgar Square featured in film and television productions during the Swinging London era of the late 1960s, including The Avengers, Casino Royale, Doctor Who, and The Ipcress File. It was used for filming several sketches and a cartoon backdrop in the BBC comedy series Monty Python's Flying Circus. In May 2007, the square was grassed over with 2,000 square metres of turf for two days in a campaign by London authorities to promote "green spaces" in the city. In July 2011, due to building works in Leicester Square, the world premiere of the final film in the Harry Potter series, Harry Potter and the Deathly Hallows – Part 2, was held in Trafalgar Square, with a 0.75-mile (1.21 km) red carpet linking the squares. Fans camped in Trafalgar Square for up to three days before the premiere, despite torrential rain. It was the first film premiere ever to be held there. A Lego architecture set based on Trafalgar Square was released in 2019. It contains models of the National Gallery and Nelson's Column alongside miniature lions, fountains and double-decker buses. Trafalgar Square is one of the squares on the standard British Monopoly Board. It is in the red set alongside the Strand and Fleet Street. Several scenes in the dystopian future of George Orwell's Nineteen Eighty Four take place in Trafalgar Square, which was renamed "Victory Square" by the story's totalitarian regime and dominated by the giant statue of Big Brother which replaced Nelson. The square has seen controversy over busking and street theatre, which have attracted complaints over noise and public safety. In 2012, the Greater London Authority created a bylaw for regulating busking and associated tourism. In 2016, the National Gallery proposed to introduce licensing for such performances. A Trafalgar Square in Stepney is recorded in Lockie's Topography of London, published in 1810. Trafalgar Square in Scarborough, North Yorkshire gives its name to the Trafalgar Square End at the town's North Marine Road cricket ground. The square known as Chelsea Square, London SW3 was at one time known as Trafalgar Square and predated the one in Westminster. National Heroes Square in Bridgetown, Barbados, was named Trafalgar Square in 1813, before its better-known British namesake. It was renamed in 1999 to commemorate national heroes of Barbados. There is a life scale replica of the square in Bahria Town, Lahore, Pakistan where it is a tourist attraction and centre for local residents. Notes Citations Sources
[[File:Meithei_manuscript,_a_Indian_language.jpg|thumb|upright=2.5|center|A [[Puya (Meitei texts)|Puya]] ([[Ancient Manipuri]] manuscript), dating back to the period of [[Ancient Manipur]]]] [[File:Ancient Tamil Script.jpg|thumb|right|Ancient [[Tamil language|Tamil]] inscription at the [[Brihadeeswara Temple]] in [[Thanjavur]]]] An '''official language''' is a [[language]] that has special [[wikt:status|status]] in a [[country]] or other [[organization]]. Usually, the [[government]] does its [[business]] in the official language. They are sometimes named in a country's [[constitution]]. Some countries, like [[Switzerland]] and [[Luxembourg]], have [[Multilingualism|more than one]] official language. Other countries, such as the [[United States]] and [[Australia]], may not have a ''[[de jure]]'' official language set by [[law]], but they may have one or more [[national language]]s that is considered ''[[de facto]]'' official because it is used by the government and people in the country. An official language does not have to be a written language. It can be a [[pidgin]] language (like in [[Papua New Guinea]]), or a sign language (like in [[New Zealand]]). == De jure, de facto, and minority languages == A [[de jure]] language is a language that is officially recognized by the government. For example, Persian is the de jure language of Tajikistan. However, a [[de facto]] language is a language that is generally accepted as the official language but has no legal status. For example, English is a de facto language of the United States as it has no official status. Other languages are not generally accepted as the official language, but are recognized as languages commonly spoken in the region. For example, Irish is a minority language of Great Britain, as it does not have official status but is commonly spoken there. == Related pages == * [[European Charter for Regional or Minority Languages]] *[[National language]] * [[Lingua franca]] [[Category:Languages]] [[Category:Vocabulary]] {{stub}}
An official language is a language having certain rights to be used in defined situations. These rights can be created in written form or by historic usage. 178 countries recognize an official language, 101 of them recognizing more than one. The government of Italy made Italian official only in 1999, and some nations (such as the United States, Mexico, and Australia) have never declared de jure official languages at the national level. Other nations have declared non-indigenous official languages. Many of the world's constitutions mention one or more official or national languages. Some countries use the official language designation to empower indigenous groups by giving them access to the government in their native languages. In countries that do not formally designate an official language, a de facto national language usually evolves. English is the most common official language, with recognized status in 51 countries. Arabic, French, and Spanish are also widely recognized. An official language that is also an indigenous language is called endoglossic, one that is not indigenous is exoglossic. An instance is Nigeria which has three endoglossic official languages. By this, the country aims to protect the indigenous languages although at the same time recognising the English language as its lingua franca. In spatial terms, indigenous (endoglossic) languages are mostly employed in the function of official languages in Eurasia, while mainly non-indigenous (exoglossic) imperial (European) languages fulfill this function in most of the "Rest of the World" (that is, in Africa, the Americas, Australia and Oceania). Lesotho, Madagascar, South Africa, East African countries, Greenland, New Zealand, Samoa and Paraguay are among the exceptions to this tendency. Around 500 BC, when Darius the Great annexed Mesopotamia to the Persian Empire, he chose a form of the Aramaic language (the so-called Official Aramaic or Imperial Aramaic) as the vehicle for written communication between the different regions of the vast empire with its different peoples and languages. Aramaic script was widely employed from Egypt in the southwest to Bactria and Sogdiana in the northeast. Texts were dictated in the native dialects and written down in Aramaic, and then read out again in the native language at the places they were received. The First Emperor of Qin standardized the written language of China after unifying the country in 221 BC. Classical Chinese would remain the standard written language for the next 2000 years. Standardization of the spoken language received less political attention, and Mandarin developed on an ad hoc basis from the dialects of the various imperial capitals until being officially standardized in the early twentieth century. The following languages are official (de jure or de facto) in three or more sovereign states. In some cases, a language may be defined as different languages in different countries. Examples are Hindi and Urdu, Malay and Indonesian, Serbian and Croatian, Persian and Tajik. Some countries—like Australia, the United Kingdom and the United States—have no official language recognized as such at a national level. On the other extreme, Bolivia officially recognizes 37 languages, the most of any country in the world. Second to Bolivia is India with 22 official languages. South Africa is the country with the third lead with 11 official languages that all have equal status; Bolivia gives primacy to Spanish, and India gives primacy to English and Hindi. The selection of an official language (or the lack thereof) is often contentious. An alternative to having a single official language is "official multilingualism", where a government recognizes multiple official languages. Under this system, all government services are available in all official languages. Each citizen may choose their preferred language when conducting business. Most countries are multilingual and many are officially multilingual. Taiwan, Canada, the Philippines, Belgium, Switzerland, and the European Union are examples of official multilingualism. This has been described as controversial and, in some other areas where it has been proposed, the idea has been rejected. It has also been described as necessary for the recognition of different groups or as an advantage for the country in presenting itself to outsiders. Following Chapter 1, Article 16 of the Constitution of Afghanistan, the Afghan government gives equal status to Pashto and Dari as official languages. English is the de facto national language of Australia, while Australia has no de jure official language, English is the first language of the majority of the population, and has been entrenched as the de facto national language since European settlement, being the only language spoken in the home for 72% of Australians. Article 21 of Azerbaijani Constitution designates the official language of the Republic of Azerbaijan as Azerbaijani Language. After the independence of Bangladesh in 1971, the then Head of the State Sheikh Mujibur Rahman adopted the policy of 'one state one language'. The de facto national language, Bengali, is the sole official language of Bangladesh according to the third article of the Constitution of Bangladesh. The government of Bangladesh introduced the Bengali Language Implementation Act, 1987 to ensure the mandatory use of Bengali in all government affairs. Belarusian and Russian have official status in the Republic of Belarus. Belgium has three official languages: Dutch, French and German. Bulgarian is the sole official language in Bulgaria. Following the Constitution Act, 1982 the (federal) Government of Canada gives equal status to English and French as official languages. The Province of New Brunswick is also officially bilingual, as is Yukon. Nunavut has four official languages: English, French, Inuktitut and Inuinnaqtun. The Northwest Territories has eleven official languages: Chipewyan/Dené, Cree, English, French, Gwich’in, Inuinnaqtun, Inuktitut, Inuvialuktun, North Slavey, South Slavey, and Tłı̨chǫ (Dogrib). All provinces, however, offer some necessary services in both English and French. The Province of Quebec with the Official Language Act (Quebec) and Charter of the French Language defines French, the language of the majority of the population, as the official language of the provincial government. Ethiopia has five official languages (Amharic alone until 2020) Amharic, Oromo, Somali, Tigrinya, and Afar, but Amharic is the de facto sole official language which is used by the government for issuing driving licenses, business licenses, passport, and foreign diplomacy with the addition that Court documents are in Amharic, and the constitution is written in Amharic, making Amharic a higher official language in the country. According to the Finnish constitution, Finnish and Swedish are the official languages of the republic. Citizens have the right to communicate in either language with government agencies. German is the official language of Germany. However, its minority languages include Sorbian (Upper Sorbian and Lower Sorbian), Romani, Danish and North Frisian, which are officially recognised. Migrant languages like Turkish, Russian and Spanish are widespread but are not officially recognised languages. According to the Basic Law of Hong Kong and the Official Languages Ordinance, both Chinese and English are the official languages of Hong Kong with equal status. The variety of Chinese is not stipulated; however, Cantonese, being the language most commonly used by the majority of Hongkongers, forms the de facto standard. Similarly, Traditional Chinese characters are most commonly used in Hong Kong and form the de facto standard for written Chinese, however, there is an increasing presence of Simplified Chinese characters particularly in areas related to tourism. In government use, documents written using Traditional Chinese characters are authoritative over ones written with Simplified Chinese characters. The Constitution of India (part 17) designates the official language of the Government of India as Hindi written in the Devanagari script. Although the original intentions of the constitution were to phase out English as an official language, provisions were provided so that "Parliament may by law provide for the use ... of ... the English language". The Eighth Schedule of the Indian Constitution lists has 22 languages, which have been referred to as scheduled languages and given recognition, status and official encouragement. In addition, the Government of India has awarded the distinction of classical language to Tamil, Sanskrit, Kannada, Telugu, Malayalam and Odia. The official language of Indonesia is the Indonesian language (Bahasa Indonesia). Bahasa Indonesia is regulated in Chapter XV, 1945 Constitution of Indonesia. On 19 July 2018, the Knesset passed a basic law under the title Israel as the Nation-State of the Jewish People, which defines Hebrew as "the State's language" and Arabic as a language with "a special status in the State" (article 4). The law further says that it should not be interpreted as compromising the status of the Arabic language in practice before the enactment of the basic law, namely, it preserves the status quo and changes the status of Hebrew and Arabic only nominally. Before the enactment of the aforementioned basic law, the status of official language in Israel was determined by the 82nd paragraph of the "Palestine Order in Council" issued on 14 August 1922, for the British Mandate of Palestine, as amended in 1939: All Ordinances, official notices and official forms of the Government and all official notices of local authorities and municipalities in areas to be prescribed by order of the High Commissioner, shall be published in English, Arabic, and Hebrew." This law, like most other laws of the British Mandate, was adopted in the State of Israel, subject to certain amendments published by the provisional legislative branch on 19 May 1948. The amendment states that: In most public schools, the main teaching language is Hebrew, English is taught as a second language, and most students learn a third language, usually Arabic but not necessarily. Other public schools have Arabic as their main teaching language, and they teach Hebrew as a second language and English as a third one. There are also bilingual schools which aim to teach both Hebrew and Arabic equally. Some languages other than Hebrew and Arabic, such as English, Russian, Amharic, Yiddish and Ladino enjoy a somewhat special status but are not official languages. For instance, at least 5% of the broadcasting time of privately owned TV channels must be translated into Russian (a similar privilege is granted to Arabic), warnings must be translated to several languages, and signs are mostly trilingual (Hebrew, Arabic and English), and the government supports Yiddish and Ladino culture (alongside Hebrew culture and Arabic culture). The Official Language Law recognizes Latvian as the sole official language of Latvia, while Latgalian is protected as "a historic variant of Latvian" and Livonian is recognized as "the language of the indigenous (autochthonous) population". Latvia also provides national minority education programmes in Russian, Polish, Hebrew, Ukrainian, Estonian, Lithuanian, and Belarusian. In 2012 there was a constitutional referendum on elevating Russian as a co-official language, but the proposal was rejected by nearly three-quarters of the voters. The official language of Malaysia is the Malay language (Bahasa Melayu), also known as Bahasa Malaysia or just Bahasa for short. Bahasa Melayu is being protected under Article 152 of the Constitution of Malaysia. Dutch is the official language of the Netherlands (a constituent country within the Kingdom of the Netherlands). In the province of Friesland, Frisian is the official second language. While Dutch is therefore the official language of the Caribbean Netherlands (the islands Bonaire, Saba and Sint Eustatius), it is not any of the three islands' main spoken language: Papiamento is the most often spoken language on Bonaire, while English is on both Saba and Sint Eustatius. These languages can be used in official documents (but do not have the same status as Frisian). Low Saxon and Limburgish, languages acknowledged by the European Charter, are spoken in specific regions of the Netherlands. New Zealand has two official languages. The Māori language and New Zealand Sign Language both have restricted de jure official status under the Māori Language Act 1987 and New Zealand Sign Language Act 2006. English is not an official language. In 2018, New Zealand First MP Clayton Mitchell introduced a bill to parliament to statutorily recognise English as an official language. As of May 2020, the bill has not progressed. In 2023 during the 2023 New Zealand general election, The New Zealand First Party released a policy of making English an official language of New Zealand. The official language of Nigeria is English, which was chosen to facilitate the cultural and linguistic unity of the country. British colonial rule ended in 1960. Urdu and English both are official languages in Pakistan. Pakistan has more than 60 other languages. Polish is the official language of Poland. Russian is the official language of the Russian Federation and in all federal subjects, however many minority languages have official status in the areas where they are indigenous. One type of federal subject in Russia, republics, are allowed to adopt additional official languages alongside Russian in their constitutions. Republics are often based around particular native ethnic groups and are often areas where ethnic Russians and native Russian-language speakers are a minority. South Africa has twelve official languages that are mostly indigenous. Due to limited funding, however, the government rarely produces documents in most languages. Accusations of mismanagement and corruption have been leveled against the Pan South African Language Board, established to promote multilingualism, develop the 11 official languages, and protect language rights in the country. The four national languages of Switzerland are German, French, Italian and Romansh. At the federal level German, French and Italian are official languages, the official languages of individual cantons depend on the languages spoken in them. Mandarin is the most common language used in government. After World War II the mainland Chinese-run government made Mandarin the official language, and it was used in the schools and government. Under the National languages development act, political participation can be conducted in any national language, which is defined as a "natural language used by an original people group of Taiwan", which also includes Formosan languages, the Taiwanese variety of Hokkien and Hakka. According to Taiwan's Legislative Yuan, amendments were made to the Hakka Basic Act to make Hakka an official language of Taiwan. According to the constitution of Timor-Leste, Tetum and Portuguese are the official languages of the country, and every official document must be published in both languages; Indonesian and English hold "working language" status in the country. The official language of Ukraine is Ukrainian. The de facto official language of the United Kingdom is English. In Wales, the Welsh language, spoken by approximately 20% of the population, has de jure official status, alongside English. English is the de facto national language of the United States. While there is no official language at the federal level, 32 of the 50 U.S. states and all five inhabited U.S. territories have designated English as one, or the only, official language, while courts have found that residents in the 50 states do not have a right to government services in their preferred language. Public debate in the last few decades has focused on whether Spanish should be recognized by the government, or whether all business should be done in English. California allows people to take their driving test in the following 32 languages: Amharic, Arabic, Armenian, Chinese, Croatian, English, French, German, Greek, Hebrew, Hindi, Hmong, Hungarian, Indonesian, Italian, Japanese, Khmer, Korean, Laotian, Persian, Polish, Portuguese, Punjabi, Romanian, Russian, Samoan, Spanish, Tagalog/Filipino, Thai, Tongan, Turkish, and Vietnamese. New York state provides voter-registration forms in the following five languages: Bengali, Chinese, English, Korean and Spanish. The same languages are also on ballot papers in certain parts of the state (namely, New York City). The pro-English-only website U.S. English sees a multilingual government as one in which its "services actually encourage the growth of linguistic enclaves...[and] contributes to racial and ethnic conflicts". Opponents of an official language policy in the United States argue that it would hamper "the government's ability to reach out, communicate, and warn people in the event of a natural or man-made disaster such as a hurricane, pandemic, or...another terrorist attack". Professor of politics Alan Patten argues that disengagement (officially ignoring the issue) works well in religious issues but that it is not possible with language issues because it must offer public services in some language. Even if it makes a conscious effort not to establish an official language, a de facto official language, or the "national language", will nevertheless emerge. Sometimes an official language definition can be motivated more by national identity than by linguistic concerns. Prior to the breakup in early 1990s, although SFR Yugoslavia had no official language on the federal level, its six constituent republics including two autonomous provinces accounted for four official languages—Serbo-Croatian, Slovene, Macedonian and Albanian. Serbo-Croatian served as the lingua franca for mutual understanding and was also the language of the military, as official in four republics and taught as a second language in the other two. When Croatia declared independence in 1991, it defined its official language as Croatian, while the confederate union of Serbia and Montenegro likewise defined its official language as Serbian in 1992. Bosnia and Herzegovina defined three official languages: Bosnian, Croatian, and Serbian. From the linguistic point of view, the different names refer to national varieties of the same language, which is known under the appellation of Serbo-Croatian. The language used in Montenegro became standardized as the Montenegrin language upon Montenegro's declaration of independence from Serbia and Montenegro in 2006.
[[File:Wedding cake aux fruits-rouges.jpg|thumb|250px|A wedding cake ]] [[File:Piece of chocolate cake on a white plate decorated with chocolate sauce.jpg|thumb|170x170px|A chocolate Cake]] A '''cake''' is a type of (usually) sweet [[dessert]] which is [[baking|baked]]. Originally, it was a [[bread]]-like food, but no longer. Cakes are often made to celebrate special occasions like [[birthday]]s or [[weddings]]. There are many kinds of cakes. It can be baked in an [[oven]]. Some [[savory]] cakes are made on a griddle or a [[frying pan]]. == Construction == Cakes are baked from a batter. [[Batter]] is made by mixing wet ingredients (like [[milk]] and [[Egg (food)|egg]]s) with dry ingredients (like sugar and flour). The batter is baked in an [[oven]]. This way of baking is known as the [[muffin]] method, because [[muffin]]s are made this way. [[Vegan]] versions of the same cake abstain from using animal products such as [[dairy]] or [[Egg (food)|eggs]] and instead substitute them with plant derived products such as nut milk and alginate-lecithin gel.<ref>Salwa, M. "PHYSICAL, CHEMICAL AND SENSORY CHARACTERISTICS OF VEGAN CAKE." Arab Universities Journal of Agricultural Sciences 17.2 (2009): 335-350.</ref> Just like [[bread]], cakes rise in the oven because they contain many small air [[Liquid bubble|bubbles]]. As cakes rise, the air bubbles expand. This is why the cake batter expands in the pan (often to twice its original size). There are two ways of forming the air bubbles, which create different types of cakes. Almost every kind of cake belongs to one of these families. === Creamed Fat === These cakes are made with [[butter]] or another [[fat]], like vegetable shortening. The common way is to mix the fat and sugar, then add eggs, and then add flour. The fat should be soft. It should not be hard or liquid. Mixing [[sugar]] with fat creates many very small air bubbles. Most [[birthday]] cakes are made this way. Cupcakes are also made this way. [[File:My daughter birthday cake.jpg|thumb|A three-pound homestyle birthday by someone's birthday ]] === Egg foam=== These cakes are not made with solid butter or vegetable shortening. Some of these cakes are made with melted butter or vegetable oil. The common way is to mix the eggs and sugar, and then add flour. These cakes are often much taller, lighter and often [[sponge|spongier]] than creamed fat cakes. [[Angelfood cake]] and [[chiffon cakes]] are egg foam cakes. Most grocery stores with bakeries sell angelfood cakes. Angelfood cakes are made by beating egg whites with sugar. This traps a lot of tiny air bubbles. The eggs and sugar are mixed with other ingredients. Then, the cake is baked. == Decorations == A cake can be decorated with [[icing]] (also called "frosting"), chocolate, [[fruit]], and much more. A layer cake is made by stacking cakes with icing or filling between the layers. [[Birthday]] cakes are sometimes decorated with [[candle]]s. Cakes can be served with [[berry|berries]] or other kinds of [[fruit]]. A large, rich cake is often called by the [[French language|French]] word for "cake": "gâteau" (plural: "gâteaux", both singular and plural pronounced "GA-toe").<ref>{{cite web|last=Olver|first=Lynne|title=About cake|url=http://www.foodtimeline.org/foodcakes.html|publisher=foodtimeline.org|accessdate=19 February 2016}}</ref><ref>{{cite book|last=Goldstein|first=Darra|title=The Oxford Companion to Sugar and Sweets|year=2015|publisher=Oxford University Press|isbn=978-0-19-931339-6|pages=92}}</ref> It may have a lot of [[cream]]. Some cakes can have edible paper on the top. ==Gallery== <gallery> File:Bolo Floresta Negra DSCN3536.jpg|Black Forrest cake File:003 2016 10 15 Gebaeck und Kuchen.jpg|Modern style wedding cake File:Chocolate Cake Flourless (1).jpg|A [[flour]]less [[chocolate]] cake File:Sifon pandan.JPG|Pandan cake File:Chocolate angel food cake with various toppings.jpg|Angel food cake File:Matcha tiramisu.jpg|Matcha tiramisu cake File:Muttertagstorte im Mai.jpg|Chocolate cake File:Colorful Cake (Unsplash).jpg|Rainbow coloured cake File:Bolo de Mel.JPG|Honey cake File:Greek revani.JPG|Greek sponge cake File:Homemade cake for mum's day.jpg|Homemade Mother's Day Cake File:Nussecke, whole.jpg|Nusscake </gallery> {{-}} {{Commonscat|Cakes|Cake}} ==References== {{reflist}}[https://www.flavoursguru.com/blogs/a-beginners-guide-to-baking-the-perfect-cake-essential-tips-and-tools/ how to make Beginner’s Baking the Perfect Cake] [[Category:Cakes| ]] [[Category:Basic English 850 words]]
Cake is a flour confection made from flour, sugar, and other ingredients and is usually baked. In their oldest forms, cakes were modifications of bread, but cakes now cover a wide range of preparations that can be simple or elaborate and which share features with desserts such as pastries, meringues, custards, and pies. The most common ingredients include flour, sugar, eggs, fat (such as butter, oil, or margarine), a liquid, and a leavening agent, such as baking soda or baking powder. Common additional ingredients include dried, candied, or fresh fruit, nuts, cocoa, and extracts such as vanilla, with numerous substitutions for the primary ingredients. Cakes can also be filled with fruit preserves, nuts, or dessert sauces (like custard, jelly, cooked fruit, whipped cream, or syrups), iced with buttercream or other icings, and decorated with marzipan, piped borders, or candied fruit. Cake is often served as a celebratory dish on ceremonial occasions, such as weddings, anniversaries, and birthdays. There are countless cake recipes; some are bread-like, some are rich and elaborate, and many are centuries old. Cake making is no longer a complicated procedure; while at one time considerable labor went into cake making (particularly the whisking of egg foams), baking equipment and directions have been simplified so that even the most amateur of cooks may bake a cake. The term "cake" has a long history. The word itself is of Viking origin, from the Old Norse word "kaka". The ancient Greeks called cake πλακοῦς (plakous), which was derived from the word for "flat", πλακόεις (plakoeis). It was baked using flour mixed with eggs, milk, nuts, and honey. They also had a cake called "satura", which was a flat, heavy cake. During the Roman period, the name for cake became "placenta", which was derived from the Greek term. A placenta was baked on a pastry base or inside a pastry case. The Greeks invented beer as a leavener, frying fritters in olive oil, and cheesecakes using goat's milk. In ancient Rome, the basic bread dough was sometimes enriched with butter, eggs, and honey, which produced a sweet and cake-like baked good. The Latin poet Ovid refers to his and his brother's birthday party and cake in his first book of exile, Tristia. Early cakes in England were also essentially bread: the most obvious differences between a "cake" and "bread" were the round, flat shape of the cakes and the cooking method, which turned cakes over once while cooking, while bread was left upright throughout the baking process. Sponge cakes, leavened with beaten eggs, originated during the Renaissance, possibly in Spain. During the Great Depression, there was a surplus of molasses and the need to provide easily made food to millions of economically depressed people in the United States. One company patented a cake-bread mix to deal with this economic situation and thereby established the first line of cake in a box. In so doing, cake, as it is known today, became a mass-produced good rather than a home- or bakery-made specialty. Later, during the post-war boom, other American companies (notably General Mills) developed this idea further, marketing cake mix on the principle of convenience, especially to housewives. When sales dropped heavily in the 1950s, marketers discovered that baking cakes, once a task at which housewives could exercise skill and creativity, had become dispiriting. This was a period in American ideological history when women, retired from the war-time labor force, were confined to the domestic sphere while still exposed to the blossoming consumerism in the US. This inspired psychologist Ernest Dichter to find a solution to the cake mix problem in the frosting. Since making the cake was so simple, housewives and other in-home cake makers could expend their creative energy on cake decorating inspired by, among other things, photographs in magazines of elaborately decorated cakes. Ever since cake in a box has become a staple of supermarkets and is complemented with frosting in a can. Cakes are broadly divided into several categories, based primarily on ingredients and mixing techniques. Although clear examples of the difference between cake and bread are easy to find, the precise classification has always been elusive. Butter cakes are made from creamed butter, sugar, eggs, and flour. They rely on the combination of butter and sugar beaten for an extended time to incorporate air into the batter. A classic pound cake is made with a pound each of butter, sugar, eggs, and flour. Another type of butter cake that takes its name from the proportion of ingredients used is 1-2-3-4 cake: 1 cup butter, 2 cups sugar, 3 cups flour, and 4 eggs. According to Beth Tartan, this cake was one of the most common among the American pioneers who settled North Carolina. Baking powder is in many butter cakes, such as Victoria sponge. The ingredients are sometimes mixed without creaming the butter, using recipes for simple and quick cakes. Sponge cakes (or foam cakes) are made from whipped eggs, sugar, and flour. Traditional sponge cakes are leavened only with eggs. They rely primarily on trapped air in a protein matrix (generally beaten eggs) to provide leavening, sometimes with a bit of baking powder or other chemical leaven added. Egg-leavened sponge cakes are thought to be the oldest cakes made without yeast. Angel food cake is a white cake that uses only the whites of the eggs and is traditionally baked in a tube pan. The French Génoise is a sponge cake that includes clarified butter. Highly decorated sponge cakes with lavish toppings are sometimes called gateau, the French word for cake. Chiffon cakes are sponge cakes with vegetable oil, which adds moistness. Chocolate cakes are butter cakes, sponge cakes, or other cakes flavored with melted chocolate or cocoa powder. German chocolate cake is a variety of chocolate cake. Fudge cakes are chocolate cakes that contain fudge. Coffee cake is generally thought of as a cake to serve with coffee or tea at breakfast or a coffee break. Some types use yeast as a leavening agent, while others use baking soda or baking powder. These cakes often have a crumb topping called streusel or a light glaze drizzle. Baked flourless cakes include baked cheesecakes and flourless chocolate cakes. Layer cakes are cakes made with layers of sponge or butter cake filled with cream, jam, or other filling to hold the layers together. One-egg cakes are made with one egg. They can be made with butter or vegetable shortening. One egg cake was an economical recipe when using two eggs for each cake was too costly. Although clear examples of the difference between cake and bread are easy to find, the precise classification has always been elusive. For example, banana bread may be properly considered either a quick bread or a cake. Yeast cakes are the oldest and are very similar to yeast bread. Such cakes are often very traditional in form and include such pastries as babka and stollen. Cakes may be classified according to the occasion for which they are intended. For example, wedding cakes, birthday cakes, cakes for first communion, Christmas cakes, Halloween cakes, and Passover plava (a type of sponge cake sometimes made with matzo meal) are all identified primarily according to the celebration they are intended to accompany. The cutting of a wedding cake constitutes a social ceremony in some cultures. The Ancient Roman marriage ritual of confarreatio originated in the sharing of a cake. Particular types of cake may be associated with particular festivals, such as stollen or chocolate log (at Christmas), babka and simnel cake (at Easter), or mooncake. There has been a long tradition of decorating an iced cake at Christmas time; other cakes associated with Christmas include chocolate log and mince pies. A Lancashire Courting Cake is a fruit-filled cake baked by a fiancée for her betrothed. The cake has been described as "somewhere between a firm sponge – with a greater proportion of flour to fat and eggs than a Victoria sponge cake – and a shortbread base and was proof of the bride-to-be's baking skills". Traditionally it is a two-layer cake filled and topped with strawberries or raspberries and whipped cream. Cakes are frequently described according to their physical form. Cakes may be small and intended for individual consumption. Larger cakes may be made to be sliced and served as part of a meal or social function. Common shapes include: Special cake flour with a high starch-to-gluten ratio is made from fine-textured, soft, low-protein wheat. It is strongly bleached and compared to all-purpose flour, cake flour tends to result in cakes with a lighter, less dense texture. Therefore, it is frequently specified or preferred in cakes meant to be soft, light or bright white, such as angel food cake. However, if cake flour is called for, a substitute can be made by replacing a small percentage of all-purpose flour with cornstarch or removing two tablespoons from each cup of all-purpose flour. Some recipes explicitly specify or permit all-purpose flour, notably where a firmer or denser cake texture is desired. A cake can fail to bake properly, which is called "falling". In a cake that "falls", parts may sink or flatten, because it was baked at a temperature that is too low or too hot, when it has been underbaked and when placed in an oven that is too hot at the beginning of the baking process. The use of excessive amounts of sugar, flour, fat or leavening can also cause a cake to fall. A cake can also fall when subjected to cool air that enters an oven when the oven door is opened during the cooking process. A finished cake is often enhanced by covering it with icing, or frosting, and toppings such as sprinkles, which are also known as "jimmies" in certain parts of the United States and "hundreds and thousands" in the United Kingdom. The frosting is usually made from powdered (icing) sugar, sometimes a fat of some sort, milk or cream, and often flavorings such as a vanilla extract or cocoa powder. Some decorators use a rolled fondant icing. Commercial bakeries tend to use lard for the fat, and often whip the lard to introduce air bubbles. This makes the icing light and spreadable. Home bakers either use lard, butter, margarine, or some combination thereof. Sprinkles are small firm pieces of sugar and oils that are colored with food coloring. In the late 20th century, new cake decorating products became available to the public. These include several specialized sprinkles and even methods to print pictures and transfer the image onto a cake. Special tools are needed for more complex cake decorating, such as piping bags and various piping tips, syringes and embossing mats. To use a piping bag or syringe, a piping tip is attached to the bag or syringe using a coupler. The bag or syringe is partially filled with icing which is sometimes colored. Using different piping tips and various techniques, a cake decorator can make many different designs. Basic decorating tips include open star, closed star, basketweave, round, drop flower, leaf, multi, petal, and specialty tips. An embossing mat is used to create embossed effects. A cake turntable that cakes are spun upon may be used in cake decoration. Royal icing, marzipan (or a less sweet version, known as almond paste), fondant icing (also known as sugar paste), and buttercream are used as covering icings and to create decorations. Floral sugarcraft or wired sugar flowers are an important part of cake decoration. Cakes for special occasions, such as wedding cakes, are traditionally rich fruit cakes or occasionally Madeira cakes, that are covered with marzipan and iced using royal icing or sugar-paste. They are finished with piped borders (made with royal icing) and adorned with a piped message, wired sugar flowers, hand-formed fondant flowers, marzipan fruit, piped flowers, or crystallized fruits or flowers such as grapes or violets. The shelf life of cakes packages for commercial sale depends on several factors. Cakes are intermediate moisture products prone to mold growth. Commercial cakes are frequently and commonly exposed to different mold varieties before they are packaged for sale, including Aspergillus flavus and various Penicillins, and Aspergillus niger. Preservatives and oxygen absorbents are currently used to control and inhibit mold growth. The CDC has recommended not to eat raw cake batter because it can contain pathogens like Salmonella and E. coli. Cake batter uses raw flour which can contain live bacteria and present a hazard if consumed.
{{for|House the show|House (TV series)}} [[File:Ranch style home in Salinas, California.JPG|thumb|House|300x300px]] A '''house''' is a building that is made for people to live in. It is usually built for a [[family]] (parents and their children).<ref>''Macmillan Dictionary for Students'' Macmillan, Pan Ltd. (1981), page 499. Retrieved 2010-7-23.</ref> It is a "permanent" building that is meant to stay standing. It is not easily packed up and carried away like a [[tent]], or moved like a [[Mobile home|caravan]]. If people live in the same house for more than a short stay, then they call it their "home". Being without a home is called [[homelessness]]. Houses are usually numbered. Some are also named. Houses are usually occupied by a single [[family]] or housemates, like in the cases of [[group home]]s and [[boarding house]]s. Houses come in many different shapes, sizes and architectural styles. They range from one room [[shack]]<nowiki/>s and [[cabin]]<nowiki/>s to the largest of [[palace]]<nowiki/>s, [[mansion]]<nowiki/>s and [[castle]]<nowiki/>s. Houses also are made many different materials from [[Thatching|thatch]] and wood , to brick, stone and modern materials like metal. Typically, a house contains only one to three floor levels/ [[storey]]<nowiki/>s, but four or more stories are common in urban and waterfront areas. [[Basement]]<nowiki/>s are common in certain climatic zones. Houses can be standard shaped or can take unorthodox shapes such as of [[dome]]<nowiki/>s and [[pyramid]]<nowiki/>s. Houses can be detached stand-alone buildings or joined to other houses at the sides to make a "terrace" or "row house" (a connected row of houses). A big building with many levels and apartments is called "a block of flats" (British) or an [[apartment]] building. One of the differences between a house and an apartment is that a house has a front door to the outside world, whereas the main door of an apartment usually opens onto a passage or landing that can be used by other people in the building. Houses have a [[roof]] to keep off the rain and sun, and walls to keep out the wind and cold. They have [[window]] openings to let in light, and a floor. Houses of different places may look different to each other, because of different materials, [[climate]], and styles. If a house is important and historic enough, it can be a [[museum]] showcasing how the house's residents lived. == Types == [[File:HouseII2007.jpg|thumb|right|[[Ranch-style house]] in California]] Most houses have special areas or rooms for people to do the things that they need to live comfortably. A modern house has a place to cook food, a place to eat, places to sleep and a place to wash. These things are usually done in separate rooms, which are called the [[kitchen]], the sitting room, the [[bedroom]]s, the [[bathroom]], the [[toilet]] (or lavatory). Many houses have a separate [[dining room]] for eating meals and a separate [[washing machine|laundry]]. In some houses the toilet is in the bathroom, and in other houses it is separate. Many houses may also have a "study" or computer room and a "family room" where the children can play games and watch television. [[File:Single-family home2.jpg|thumb|A newly built detached house in the US.|279x279px]] === Detached house === In some countries such as the [[United States]], [[Canada]], [[Australia]] and [[New Zealand]] many families live in a "detached house" which is separate from other buildings and surrounded by its own yard, but is close enough to town to have shops, good transport and entertainment nearby. In many other countries, including most of [[Europe]], owning a house like this is something that only the richer families can afford, and is just a dream for most people. [[File:Jakarta slumlife16.JPG|thumb|right|A house in Jakarta]] === Houses in very poor places === In many very poor countries, a lot of people live crowded in houses with only one room. They often have to share a toilet with many other families, and have to do the cooking outside. The house might be made of materials that can be found nearby like mud bricks and grass or from second-hand materials like corrugated iron and cardboard boxes. In many cities there are thousands of small houses crowded together with narrow alleys between them. [[File:EastGarstonCottages(AndrewSmith)Aug2006.jpg|thumb|Cottages in the village of East Garston in England]] === Cottage === A small house is often called a [[cottage]]. In England, where this word comes from, it is used to mean a house that has one main storey, with a second, lower storey of bedrooms which fit under the roof upstairs. Cottages are usually found in [[village]]s or in the [[countryside]]. They are nearly always built from material that can be found nearby. In many places the word cottage is used to mean a small old-fashioned house. In the United States the word cottage is often used to mean a small holiday home. ===Tiny=== A tiny house is one that's usually 500sqft / 48sqm in size and typically built on a [[trailer]] or [[barge]] and thus portable. Tiny houses can also be built on smaller versions of conventional [[foundation]]s. === Semi-detached or duplex === [[File:Numbers 550 to 556 Lordship Lane N22.jpg|thumb|Semi-detached houses in England]] A semi-detached house is a building that has two houses side by side and covered over by just one big roof. Each house has a pathway to one side, leading from the front to the backyard. Semi-detached houses are very common in some cities, and can be single storey or several storeys high. === Terraced houses === A row or terraced house is a house that is part of a row of houses joined at the side walls. Many cities and towns have thousands of row houses because they are a good way to build several houses close together. Many row houses have two rooms downstairs and two rooms upstairs, with a kitchen or wash area at the back. Houses like this were built for poor factory workers and coal miners in many cities and towns. "Terrace houses" are a type of row house where each house looks like the others. [[File:Panorama of the Royal Crescent - panoramio.jpg|left|thumb|Royal Crescent, Bath, England]] Row houses are not always small. Some cities have large beautiful row houses, such as the [[Royal Crescent]], which was built in the [[1770s]] at [[Bath, Somerset|Bath]] in England and is a famous example of [[Georgian architecture]]. === Bungalow === [[File:SydneyBuilding0129.jpg|thumb|A bungalow in Sydney]] In some places, the word "bungalow" is used for any house that is all on one level. The word came from India and for a long time was used for a house that is built all or mostly on one level (though it can have an [[attic]] and or [[basement]]) and has a [[veranda]] where people can sit or work outdoors, but under a shady roof. A bungalow often has a hall down the middle of the house to let the breeze blow through. Bungalows are often seen in countries with hot summers, in India, South East Asia, South Africa, parts of the United States, South America, Australia and New Zealand. In regions with flooding, a bungalow is often built up on wooden "stilts" or a high basement. In the 1800s, bungalows were nearly always built of wood, but from the 1920s it became fashionable to build them of brick as well. [[File:0 Nobressart - Ancienne ferme.JPG|thumb|left|A farmhouse in Belgium]] === Farmhouse === A farmhouse may look like a cottage, a bungalow or a mansion. In many countries a farmhouse can look different to a house in a town. As well as having a place for people to live, it also has a place for animals. Three typical types of farmhouses are found. Many farm houses are long and have two doors. One door leads to the rooms used by the family. The other door opens into a stable for the cows, sheep and chickens. The [[stable]] part often has a loft where hay can be kept to feed the animals in the winter. Another type of farmhouse has two storeys with a big stable and storeroom underneath, with the rooms for the family on the upper floor. Another type of farmhouse has buildings such as the family house, the barn and the stable all joined around a central courtyard. Old farmhouses of these three types can be found in many parts of Europe and Great Britain. === Mansion === [[File:Dalfsen Havezathe Den Berg.jpg|thumb|A mansion in Netherlands]] A mansion is a big grand house, usually with two stories and sometimes more. A mansion often has beautiful architecture, and shows that the person for whom it was designed and built was rich. Mansions often have beautiful gardens. Sometimes a mansion does not belong to a private family, but to a town council, to a big business company, to a church or college and is an [[official residence]] for a person with an important job to live and to entertain guests. A mansion often has rooms which are not found in ordinary houses, such as a drawing room, a ballroom, a [[library]] and a music room. Mansions often need servants to help keep them in order and have [[servants' quarters]] and special rooms where the servants do particular jobs such as cleaning. Well-known mansions are the [[White House]] in [[Washington, D.C.]] where the [[President of the United States]] lives and Mansion House in London where the [[Lord Mayor of London]] lives. [[File:Harewood House, seen from the garden.JPG|thumb|left|Harewood House is really a palace.]] ===Castle=== A [[castle]] is a fortified house constructed in the [[medieval]] period, or a house constructed to look like a medieval fortification. Castles protected [[feudal]] lords from their attacking enemies. In later periods, castles were built as part of romanticist revival movements of the 19th and early 20th centuries. === Palace === A [[palace]] is a house that is very grand for people like Kings and Queens, or other rich people. Many palaces are homes of royal or aristocratic people. Some old palaces been changed by people who lived there over many hundreds of years. One of the best-known palaces in the world is [[Buckingham Palace]], a royal palace in [[London]]. The largest palace in the world is the [[Louvre]] which was built in [[Paris]] for the Kings of [[France]]. It now holds a famous art gallery. {{Further|Palace}} == Building materials == Houses are usually built from types of material that can easily be made or bought near the place where the house is built. Because of this, old houses in different towns and different villages look quite different to each other, even in the same country. In modern times building materials can be transported easily and this means that a builder can choose materials from far away. === Mud and clay === [[File:Bahareque llanos house from Venezuela.JPG|thumb|A farmhouse in Venezuela made of wooden slats and clay daub.]] In many parts of the world, mud or clay are the main building materials. [[Clay]] is a type of soil that sticks together more strongly than most other types of soil. There are three main ways of building walls with mud and clay. * One way to make a house of clay is simply to pile up the mud or wet clay with the hands, and flatten it into a wall shape. Often animal hair, straw and dung is mixed with the mud so that the fibres (stringy bits) help hold it together. Some wooden beams and thin wooden slats are used to hold the roof, which is also made of mud or clay. The roof is flat and strong enough to walk on, but must slope a bit, in case of rain. The walls are often protected from rain with plaster and paint. In many countries the walls are usually white, but in other places all the houses are painted different bright clolours. These types of houses are only found in very dry countries. Houses built in this way give good protection against the very hot sun and hot dry winds of the dessert. In the cold nights they stay quite warm. * A second way to make the walls of a house out of mud of clay is "wattle and daub". "Wattle" is weaving from sticks. First, long sticks are stuck into the ground and then very thin sticks are woven through them, like making a basket. The thin wattle wall is then "daubed" or stuck all over on the inside and the outside with damp clay, mud or plaster, usually with some chopped straw or animal hair to help it stick together. "Wattle and daub" are often found in quite damp places where bendy willow trees or acacias (also called "wattle trees") grow. The walls need good protection from the rain, so houses like this often have roofs that have big "eaves"; the roof sticks out all around the house so that the water does not run down the walls. "Wattle and daub" houses often have big roofs made of straw, reeds, banana leaves, palm leaves, wooden shingles or big sheets of bark. * A third way to make a house of mud or clay is to shape the wet soil into bricks before building the walls. The [[brick]]<nowiki/>s can be shaped by hand, but more often they are shaped in boxes or "moulds" so that all the bricks are just the same size and shape. They are left in the hot sun until they have dried hard. Then walls can be made which are the same thickness all the way up. Clay bricks that are dried in the sun were used in [[Egypt]] and [[Mesopotamia]] for thousands of years. [[File:Buccleuch Cottages Beaulieu New Forest.jpg|thumb|left|Brick and tile houses in Beaulieu, England]] === Fired bricks === It was discovered that if clay was used, rather than mud, very strong [[brick]]s could be made by "firing" (or baking) them in a "[[kiln]]" (a special oven). "Fired" bricks were like [[pottery]]; they did not wash away easily in [[rain]] and would sometimes last for thousands of years. The bricks were joined to each other with a type of [[cement]] called "[[Mortar (masonry)|mortar]]". Some houses built by the Ancient Romans are still standing 2000 years later. Bricks became one of the most common building materials. In most cities, nearly all the houses are built of bricks because they are long lasting, they do not need repairing very often. Brick houses are generally made strong by having all the main walls two bricks thick. Two walls are built side by side which are "bonded" (or joined to each other) by having some bricks set so they make a bridge between the two layers. The pattern of the bonded bricks can be seen from the outside. The two main patterns are called "Flemish Bond" and "British Bond". [[File:Domki budnicze Poznań.jpg|thumb|Tall narrow city houses of Poznań are painted in different colours.]] :Because brick walls are heavy, a brick house needs to have a very firm base or "[[Foundation (engineering)|foundation]]" to sit on. This often means digging deep into the ground to lay a foundation before a brick house can be built. :Houses with brick walls often have tiled roofs, because the tiles can often be made at the brick factory. :Modern bricks and tiles come in a big range of colours, because different materials can be transported from different areas. People who are employing a builder to build a home are nowadays often shown pictures of all the colours of bricks and tiles that are available. This means that although in old towns, every house is made of the same type of bricks, in a modern suburb every house may look quite different. The bricks range from white through shades of red, yellow and brown and are sometimes very patterned, while the roof tiles include blue, green and yellow as well as the traditional red, orange and brown. * In some countries, the clay for bricks does not set very hard, and washes away more easily. If the bricks are not very hard, then it is usual to cover them with plaster or cement, which can then be painted. In other places, brick walls are not very fashionable so the walls are plastered and made more elegant with decoration. Plastering or cementing over brickwork is called "render" or "stucco". In many countries such as Austria, Italy and Spain, the brick houses are usually covered with stucco and are often painted cream, yellow, orange, pink, red or "orchre" (which is a mustard colour). [[File:Maplecroft.jpg|thumb|left|"Maplecroft" is an historic timber house in Texas]] === Timber === Houses made of [[timber]] are found wherever there are, or there has once been big forests. Timber houses are also often found in seaside towns where the sea air makes brick and stone houses feel cold and damp. * One of the oldest ways to make a timber house is to cut tree-trunks into logs. The logs are then split in half so that there is a flat side for the inside of the wall and a round side for the outside. The logs then have big "notches' or holes cut into the ends, so that when they are piled up, the logs lock together at the corners of the building without any nails. These [[log cabin]]s can be seen in Canada, in Switzerland, in Scandinavia, in Eastern Europe and in Japan. * Many houses are made with a wooden "frame" or skeleton. On the outside of the frame, to keep out the weather, rows of flat boards are nailed, which overlap each other so that the rain cannot come in. Walls like this are called "clapboard" or "weatherboard". Clapboard houses are generally painted to preserve the timber. In England and some parts of the United States, they are nearly always painted white. In Australia where "weatherboard" houses are very common, they are usually painted in bright colours. In England and Australia, clapboard houses are usually quite small, but in the United States there are very many grand mansions that are built in this way. * Some houses with a wooden frame have a brick wall on the outside and a wall of plastered board on the inside. This is called "brick veneer". [[File:Weilburg - Tiergarten - Dillhäuser Bauernhaus.jpg|thumb|A half-timbered farmhouse in Germany]] === Half-timbered houses === A well-known type of old-fashioned house is the "half-timbered" house. These are seen in the British Isles, France, and across northern Europe and the Alps. These houses date mainly from about 1200 to about 1800. * Half-timbered houses have a wooden frame built up of lots of thick timber pieces that are generally quite short. The timbers are arranged in a pattern with lots of diagonal pieces to act as braces. The strong wooden frame is then filled in with all sorts of other material. In some houses the "infil" is mainly of brick. In other houses, "wattle and daub" is used between the timber. Other houses have infill of "rubble" (rough stones) in clay or mud. The "infil is generally plastered over and painted white (or in some places a colour) while the timbers are often stained black. If they are not stained, they turn pale silvery grey. Hundreds of these houses exist, including some that are quite grand. In the 1800s and early 1900s, there was a fashion to copy "half-timbering" on modern brick houses. [[File:Blundells' cottage.jpg|thumb|left|A cottage of "undressed" stone with brick chimneys in Australia]] [[File:Benslie cottages.JPG|thumb|left|A cottage of "dressed" stone in Scotland]] === Stone houses === In places where there is lots of stone, many houses are built of it. In many parts of the world, little cottages are built of stone. Many mansions and palaces are also built of stone, as are nearly all castles. * Some of the oldest houses in the world are made of stone, because stone lasts a long time. They are circular houses with walls of flat stone slabs laid carefully on each other. The walls sloped gently inwards to meet in a stone roof. Other prehistoric stone houses had roofs of thatch. * In parts of the world where there is stone that can be broken into flat slabs, then walls can be built up without any "mortar" to join the stones together. But the wind blows through the gaps, so the walls need timber or render on the inside, to keep out the weather. Stone houses of this type often have roofs made of flat stones as well. Stone roofs are very heavy and are laid on very thick beams. * Many stone cottages are built of [[flint]] or "rubble" or "cobble". These are broken stones that are brought from nearby hills, or stones picked up from the farmer's fields, or gathered from a river bed. The walls are built up with stones and mortar. In some places where flint is used, the hard stones are split or "knapped" so that when they are placed on the outside of the wall, they show a shiny flat surface which is more attractive than the dull lumpy outside of a flintstone. * Some types of stone, particularly [[limestone]] and [[sandstone]], can easily be cut into big blocks that can be built up like bricks. It can also be "dressed" or smoothed on its surfaces. In villages that are near a limestone or sandstone [[quarry]] where the stone is cut, many of the cottages are built of neatly cut stones. But in towns that are far away from the quarries, it is only the mansions and palaces that are built of stone, because it is very heavy and expensive to carry across the country. * The good thing about the types of stone that are used for grand houses is that they can easily be carved into decoration. Grand houses often have decorative carvings around the doors and windows. The building stones may also have special textures at the basement or the corners of the building. * Grand stone houses often have roofs of thin stone called slate. They also have roofs of expensive materials like [[copper]] and [[lead]]. === Making modern houses === [[File:Prefabricated house construction.gif|thumb|Building a modern house from pre-made pieces.|421x421px]] Modern houses are often made of "pre-fabricated" parts that are partly built in a factory, and are easy to put together at the site of the building. Many different types of materials for making houses have been developed in the 20th century. * Many houses are now made with steel frames put together with rivets and bolts. * Walls and roofs can be made of boards that combine fibre with cement. These boards are thin, light in weight to transport, easy to put on the frame, and much cheaper than bricks or timber. A similar material can be made into roof tiles. * Metal roofing can be rolled into thin sheets, and made in many different colours, with matching gutters and downpipes. == Notes and references == * Nicholas Pevsner, ''An Outline of European Architecture'', Pelican, ISBN * Trewin Copplestone, Ed., ''World Architecture'', Paul Hamlyn, ISBN * John Summerson, ''Architecture in Britain'', Pelican, {{ISBN|0140560033}} * Trevor Yorke, ''Tracing the History of Villages'', Countryside Books, {{ISBN|1-85306-712-1}} * Richard Reid, ''The Book of Buildings, the Architecture of Europe and North America'', Peerage Books, {{ISBN|0-907408-89-3}} * R. Apperly, R. Irving and P.Reynolds, ''A Pictorial Guide to Identifying Australian Architecture'', Angus and Robertson's, {{ISBN|0-207-16201-8}} * W.R. Dalziel, ''All-Colour Guide to Architecture'', Grosset and Dunlap, {{ISBN|0-448-00863-7}} {{reflist}} [[Category:Basic English 850 words]] [[Category:Houses| ]]
A house is a single-unit residential building. It may range in complexity from a rudimentary hut to a complex structure of wood, masonry, concrete or other material, outfitted with plumbing, electrical, and heating, ventilation, and air conditioning systems. Houses use a range of different roofing systems to keep precipitation such as rain from getting into the dwelling space. Houses may have doors or locks to secure the dwelling space and protect its inhabitants and contents from burglars or other trespassers. Most conventional modern houses in Western cultures will contain one or more bedrooms and bathrooms, a kitchen or cooking area, and a living room. A house may have a separate dining room, or the eating area may be integrated into another room. Some large houses in North America have a recreation room. In traditional agriculture-oriented societies, domestic animals such as chickens or larger livestock (like cattle) may share part of the house with humans. The social unit that lives in a house is known as a household. Most commonly, a household is a family unit of some kind, although households may also be other social groups, such as roommates or, in a rooming house, unconnected individuals. Some houses only have a dwelling space for one family or similar-sized group; larger houses called townhouses or row houses may contain numerous family dwellings in the same structure. A house may be accompanied by outbuildings, such as a garage for vehicles or a shed for gardening equipment and tools. A house may have a backyard or a front yard or both, which serve as additional areas where inhabitants can relax or eat. The English word house derives directly from the Old English word hus, meaning "dwelling, shelter, home, house," which in turn derives from Proto-Germanic husan (reconstructed by etymological analysis) which is of unknown origin. The term house itself gave rise to the letter 'B' through an early Proto-Semitic hieroglyphic symbol depicting a house. The symbol was called "bayt", "bet" or "beth" in various related languages, and became beta, the Greek letter, before it was used by the Romans. Beit in Arabic means house, while in Maltese bejt refers to the roof of the house. Ideally, architects of houses design rooms to meet the needs of the people who will live in the house. Feng shui, originally a Chinese method of moving houses according to such factors as rain and micro-climates, has recently expanded its scope to address the design of interior spaces, with a view to promoting harmonious effects on the people living inside the house, although no actual effect has ever been demonstrated. Feng shui can also mean the "aura" in or around a dwelling, making it comparable to the real estate sales concept of "indoor-outdoor flow". The square footage of a house in the United States reports the area of "living space", excluding the garage and other non-living spaces. The "square metres" figure of a house in Europe reports the area of the walls enclosing the home, and thus includes any attached garage and non-living spaces. The number of floors or levels making up the house can affect the square footage of a home. Humans often build houses for domestic or wild animals, often resembling smaller versions of human domiciles. Familiar animal houses built by humans include birdhouses, hen houses and dog houses, while housed agricultural animals more often live in barns and stables. Many houses have several large rooms with specialized functions and several very small rooms for other various reasons. These may include a living/eating area, a sleeping area, and (if suitable facilities and services exist) separate or combined washing and lavatory areas. Some larger properties may also feature rooms such as a spa room, indoor pool, indoor basketball court, and other 'non-essential' facilities. In traditional agriculture-oriented societies, domestic animals such as chickens or larger livestock often share part of the house with humans. Most conventional modern houses will at least contain a bedroom, bathroom, kitchen or cooking area, and a living room. The names of parts of a house often echo the names of parts of other buildings, but could typically include: Little is known about the earliest origin of the house and its interior; however, it can be traced back to the simplest form of shelters. An exceptionally well-preserved house dating to the fifth millennium BC and with its contents still preserved was for example excavated at Tell Madhur in Iraq. Roman architect Vitruvius' theories have claimed the first form of architecture as a frame of timber branches finished in mud, also known as the primitive hut. Philip Tabor later states the contribution of 17th century Dutch houses as the foundation of houses today. As far as the idea of the home is concerned, the home of the home is the Netherlands. This idea's crystallization might be dated to the first three-quarters of the 17th century, when the Dutch Netherlands amassed the unprecedented and unrivalled accumulation of capital, and emptied their purses into domestic space. In the Middle Ages, the Manor Houses facilitated different activities and events. Furthermore, the houses accommodated numerous people, including family, relatives, employees, servants and their guests. Their lifestyles were largely communal, as areas such as the Great Hall enforced the custom of dining and meetings and the Solar intended for shared sleeping beds. During the 15th and 16th centuries, the Italian Renaissance Palazzo consisted of plentiful rooms of connectivity. Unlike the qualities and uses of the Manor Houses, most rooms of the palazzo contained no purpose, yet were given several doors. These doors adjoined rooms in which Robin Evans describes as a "matrix of discrete but thoroughly interconnected chambers." The layout allowed occupants to freely walk room to room from one door to another, thus breaking the boundaries of privacy. An early example of the segregation of rooms and consequent enhancement of privacy may be found in 1597 at the Beaufort House built in Chelsea, London. It was designed by English architect John Thorpe who wrote on his plans, "A Long Entry through all". The separation of the passageway from the room developed the function of the corridor. This new extension was revolutionary at the time, allowing the integration of one door per room, in which all universally connected to the same corridor. English architect Sir Roger Pratt states "the common way in the middle through the whole length of the house, [avoids] the offices from one molesting the other by continual passing through them." Social hierarchies within the 17th century were highly regarded, as architecture was able to epitomize the servants and the upper class. More privacy is offered to the occupant as Pratt further claims, "the ordinary servants may never publicly appear in passing to and fro for their occasions there." This social divide between rich and poor favored the physical integration of the corridor into housing by the 19th century. Sociologist Witold Rybczynski wrote, "the subdivision of the house into day and night uses, and into formal and informal areas, had begun." Rooms were changed from public to private as single entryways forced notions of entering a room with a specific purpose. Compared to the large scaled houses in England and the Renaissance, the 17th Century Dutch house was smaller, and was only inhabited by up to four to five members. This was because they embraced "self-reliance" in contrast to the dependence on servants, and a design for a lifestyle centered on the family. It was important for the Dutch to separate work from domesticity, as the home became an escape and a place of comfort. This way of living and the home has been noted as highly similar to the contemporary family and their dwellings. By the end of the 17th century, the house layout was transformed to become employment-free, enforcing these ideas for the future. This came in favour for the industrial revolution, gaining large-scale factory production and workers. The house layout of the Dutch and its functions are still relevant today. In the American context, some professions, such as doctors, in the 19th and early 20th century typically operated out of the front room or parlor or had a two-room office on their property, which was detached from the house. By the mid 20th century, the increase in high-tech equipment created a marked shift whereby the contemporary doctor typically worked from an office or hospital. Technology and electronic systems has caused privacy issues and issues with segregating personal life from remote work. Technological advances of surveillance and communications allow insight of personal habits and private lives. As a result, the "private becomes ever more public, [and] the desire for a protective home life increases, fuelled by the very media that undermine it," writes Jonathan Hill. Work has been altered by the increase of communications. The "deluge of information" has expressed the efforts of work conveniently gaining access inside the house. Although commuting is reduced, the desire to separate working and living remains apparent. On the other hand, some architects have designed homes in which eating, working and living are brought together. In many parts of the world, houses are constructed using scavenged materials. In Manila's Payatas neighborhood, slum houses are often made of material sourced from a nearby garbage dump. In Dakar, it is common to see houses made of recycled materials standing atop a mixture of garbage and sand which serves as a foundation. The garbage-sand mixture is also used to protect the house from flooding. In the United States, modern house construction techniques include light-frame construction (in areas with access to supplies of wood) and adobe or sometimes rammed-earth construction (in arid regions with scarce wood-resources). Some areas use brick almost exclusively, and quarried stone has long provided foundations and walls. To some extent, aluminum and steel have displaced some traditional building materials. Increasingly popular alternative construction materials include insulating concrete forms (foam forms filled with concrete), structural insulated panels (foam panels faced with oriented strand board or fiber cement), light-gauge steel, and steel framing. More generally, people often build houses out of the nearest available material, and often tradition or culture govern construction-materials, so whole towns, areas, counties or even states/countries may be built out of one main type of material. For example, a large portion of American houses use wood, while most British and many European houses use stone, brick, or mud. In the early 20th century, some house designers started using prefabrication. Sears, Roebuck & Co. first marketed their Sears Catalog Homes to the general public in 1908. Prefab techniques became popular after World War II. First small inside rooms framing, then later, whole walls were prefabricated and carried to the construction site. The original impetus was to use the labor force inside a shelter during inclement weather. More recently, builders have begun to collaborate with structural engineers who use finite element analysis to design prefabricated steel-framed homes with known resistance to high wind loads and seismic forces. These newer products provide labor savings, more consistent quality, and possibly accelerated construction processes. Lesser-used construction methods have gained (or regained) popularity in recent years. Though not in wide use, these methods frequently appeal to homeowners who may become actively involved in the construction process. They include: In the developed world, energy-conservation has grown in importance in house design. Housing produces a major proportion of carbon emissions (studies have shown that it is 30% of the total in the United Kingdom). Development of a number of low-energy building types and techniques continues. They include the zero-energy house, the passive solar house, the autonomous buildings, the super insulated houses and houses built to the Passivhaus standard. Buildings with historical importance have legal restrictions. New houses in the UK are not covered by the Sale of Goods Act. When purchasing a new house , the buyer has different legal protection than when buying other products. New houses in the UK are covered by a National House Building Council guarantee. With the growth of dense settlement, humans designed ways of identifying houses and parcels of land. Individual houses sometimes acquire proper names, and those names may acquire in their turn considerable emotional connotations. A more systematic and general approach to identifying houses may use various methods of house numbering. Houses may express the circumstances or opinions of their builders or their inhabitants. Thus, a vast and elaborate house may serve as a sign of conspicuous wealth whereas a low-profile house built of recycled materials may indicate support of energy conservation. Houses of particular historical significance (former residences of the famous, for example, or even just very old houses) may gain a protected status in town planning as examples of built heritage or of streetscape. Commemorative plaques may mark such structures. Home ownership provides a common measure of prosperity in economics. Contrast the importance of house-destruction, tent dwelling and house rebuilding in the wake of many natural disasters. Building Functions Types Economics Miscellaneous Institutions Lists
A '''gigabyte''' ('''GB'''), sometimes abbreviated as a ''gig'', is a [[unit of measurement]] in [[computer]]s and similar [[Electronics|electronic devices]]. At first it meant exactly 1 [[1,000,000,000|billion]] [[byte]]s, usually referring to the number of bytes in a [[computer]] [[hard drive]]. Its meaning gradually changed over time, so that today it is often used to mean 1,073,741,824 (2<sup>30</sup>) [[byte]]s, especially when referring to the random access memory ([[RAM]]) of a modern computer. ==Definitions== Nowadays, international standards bodies say a gigabyte should only be used to mean exactly 1 billion [[byte]]s, and that [[gibibyte]] is a better name for 1,073,741,824 bytes. In everyday use, people use ''gigabyte'' to mean 1,073,741,824 bytes when talking about computer memory and 1,000,000,000 bytes when talking about computer disk space. One thousand gigabytes make one [[terabyte]]. ==Applications== This [[measurement]] is often used when measuring the capacity of [[Random access memory|computer memory]], [[hard drive]]s, or other storage [[device]]s. {{tech-stub}} [[Category:Computer science]] [[Category:Units of information]] [[Category:Information technology]]
The gigabyte (/ˈɡɪɡəbaɪt, ˈdʒɪɡəbaɪt/) is a multiple of the unit byte for digital information. The prefix giga means 10 in the International System of Units (SI). Therefore, one gigabyte is one billion bytes. The unit symbol for the gigabyte is GB. This definition is used in all contexts of science (especially data science), engineering, business, and many areas of computing, including storage capacities of hard drives, solid state drives, and tapes, as well as data transmission speeds. The term is also used in some fields of computer science and information technology to denote 1073741824 (1024 or 2) bytes, however, particularly for sizes of RAM. Thus, some usage of gigabyte has been ambiguous. To resolve this difficulty, IEC 80000-13 clarifies that a gigabyte (GB) is 10 bytes and specifies the term gibibyte (GiB) to denote 2 bytes. These differences are still readily seen, for example, when a 400 GB drive's capacity is displayed by Microsoft Windows as 372 GB instead of 372 GiB. Analogously, a memory module that is labeled as having the size "1GB" has one gibibyte (1GiB) of storage capacity. In response to litigation over whether the makers of electronic storage devices must conform to Microsoft Windows' use of a binary definition of "GB" instead of the metric/decimal definition, the United States District Court for the Northern District of California rejected that argument, ruling that "the U.S. Congress has deemed the decimal definition of gigabyte to be the 'preferred' one for the purposes of 'U.S. trade and commerce.'" The term gigabyte has a standard definition of 1000 bytes, as well as a discouraged meaning of 1024 bytes. The latter binary usage originated as compromise technical jargon for byte multiples that needed to be expressed in a power of 2, but lacked a convenient name. As 1024 (2) is approximately 1000 (10), roughly corresponding to SI multiples, it was used for binary multiples as well. In 1998 the International Electrotechnical Commission (IEC) published standards for binary prefixes, requiring that the gigabyte strictly denote 1000 bytes and gibibyte denote 1024 bytes. By the end of 2007, the IEC Standard had been adopted by the IEEE, EU, and NIST, and in 2009 it was incorporated in the International System of Quantities. Nevertheless, the term gigabyte continues to be widely used with the following two different meanings: Based on powers of 10, this definition uses the prefix giga- as defined in the International System of Units (SI). This is the recommended definition by the International Electrotechnical Commission (IEC). This definition is used in networking contexts and most storage media, particularly hard drives, flash-based storage, and DVDs, and is also consistent with the other uses of the SI prefix in computing, such as CPU clock speeds or measures of performance. The file manager of Mac OS X version 10.6 and later versions are a notable example of this usage in software, which report files sizes in decimal units. The binary definition uses powers of the base 2, as does the architectural principle of binary computers. This usage is widely promulgated by some operating systems, such as Microsoft Windows in reference to computer memory (e.g., RAM). This definition is synonymous with the unambiguous unit gibibyte. Since the first disk drive, the IBM 350, disk drive manufacturers expressed hard drive capacities using decimal prefixes. With the advent of gigabyte-range drive capacities, manufacturers labelled many consumer hard drive, solid state drive and USB flash drive capacities in certain size classes expressed in decimal gigabytes, such as "500 GB". The exact capacity of a given drive model is usually slightly larger than the class designation. Practically all manufacturers of hard disk drives and flash-memory disk devices continue to define one gigabyte as 1000000000bytes, which is displayed on the packaging. Some operating systems such as Mac OS X and Ubuntu, and Debian express hard drive capacity or file size using decimal multipliers, while others such as Microsoft Windows report size using binary multipliers. This discrepancy causes confusion, as a disk with an advertised capacity of, for example, 400 GB (meaning 400000000000bytes, equal to 372 GiB) might be reported by the operating system as "372 GB". For RAM, the JEDEC memory standards use IEEE 100 nomenclature which quote the gigabyte as 1073741824bytes (2 bytes). The difference between units based on decimal and binary prefixes increases as a semi-logarithmic (linear-log) function—for example, the decimal kilobyte value is nearly 98% of the kibibyte, a megabyte is under 96% of a mebibyte, and a gigabyte is just over 93% of a gibibyte value. This means that a 300 GB (279 GiB) hard disk might be indicated variously as "300 GB", "279 GB" or "279 GiB", depending on the operating system. As storage sizes increase and larger units are used, these differences become more pronounced. A lawsuit decided in 2019 that arose from alleged breach of contract and other claims over the binary and decimal definitions used for "gigabyte" have ended in favor of the manufacturers, with courts holding that the legal definition of gigabyte or GB is 1 GB = 1,000,000,000 (10) bytes (the decimal definition). Specifically, the courts held that "the U.S. Congress has deemed the decimal definition of gigabyte to be the 'preferred' one for the purposes of 'U.S. trade and commerce' .... The California Legislature has likewise adopted the decimal system for all 'transactions in this state'." Earlier lawsuits had ended in settlement with no court ruling on the question, such as a lawsuit against drive manufacturer Western Digital. Western Digital settled the challenge and added explicit disclaimers to products that the usable capacity may differ from the advertised capacity. Seagate was sued on similar grounds and also settled. Because of their physical design, the capacity of modern computer random access memory devices, such as DIMM modules, is always a multiple of a power of 1024. It is thus convenient to use prefixes denoting powers of 1024, known as binary prefixes, in describing them. For example, a memory capacity of 1073741824bytes (1024 B) is conveniently expressed as 1 GiB rather than as 1.074 GB. The former specification is, however, often quoted as "1 GB" when applied to random access memory. Software allocates memory in varying degrees of granularity as needed to fulfill data structure requirements and binary multiples are usually not required. Other computer capacities and rates, like storage hardware size, data transfer rates, clock speeds, operations per second, etc., do not depend on an inherent base, and are usually presented in decimal units. For example, the manufacturer of a "300 GB" hard drive is claiming a capacity of 300000000000bytes, not 300 × 1024 (which would be 322122547200) bytes. The "gigabyte" symbol is encoded by Unicode at code point U+3387 ㎇ SQUARE GB.
{{wiktionary}} [[File:Indian farmers.jpg|thumbnail|300px|Farmers working on a field in [[India]].]] A '''farmer''' is a person who [[wikt:grow|grows]] [[crops]] and [[wikt:raise|raises]] [[animal]]s to be used as [[food]]. This has been a common [[job]] for people since the beginning of [[civilization]]. == What farmers do == The word ''farmer'' usually refers to a person who has a field, [[orchard]], [[vineyard]], or garden where food is grown. This food is eaten or sold after it is [[harvest]]ed. Farmers may also grow raw materials for industrial purposes. Some examples of raw materials that are used in this way are: * [[cereal]]s for [[alcoholic beverage]]s * hides (animal skin) for [[leather]] * [[maize]]/corn for plastics or fuel * [[wool]] or [[cotton]] for [[yarn]]s and making cloth Some farmers also raise animals to be used for [[meat]], [[milk]], [[Egg (food)|eggs]] or other things. ==Related pages== {{commonscat|Farmers}} *[[Agriculture]] [[Category:Agricultural occupations]] {{stub}} {{Agriculture footer}}
A farmer is a person engaged in agriculture, raising living organisms for food or raw materials. The term usually applies to people who do some combination of raising field crops, orchards, vineyards, poultry, or other livestock. A farmer might own the farm land or might work as a laborer on land owned by others. In most developed economies, a "farmer" is usually a farm owner (landowner), while employees of the farm are known as farm workers (or farmhands). However, in other older definitions a farmer was a person who promotes or improves the growth of plants, land, or crops or raises animals (as livestock or fish) by labor and attention. Over half a billion farmers are smallholders, most of whom are in developing countries and who economically support almost two billion people. Globally, women constitute more than 40% of agricultural employees. Farming dates back as far as the Neolithic, being one of the defining characteristics of that era. By the Bronze Age, the Sumerians had an agriculture specialized labor force by 5000–4000 BCE, and heavily depended on irrigation to grow crops. They relied on three-person teams when harvesting in the spring. The Ancient Egypt farmers farmed and relied and irrigated their water from the Nile. Animal husbandry, the practice of rearing animals specifically for farming purposes, has existed for thousands of years. Dogs were domesticated in East Asia about 15,000 years ago. Goats and sheep were domesticated around 8000 BCE in Asia. Swine or pigs were domesticated by 7000 BCE in the Middle East and China. The earliest evidence of horse domestication dates to around 4000 BCE. In the US of the 1930s, one farmer could produce only enough food to feed three other consumers. A modern farmer produces enough food to feed well over a hundred people. However, some authors consider this estimate to be flawed, as it does not take into account that farming requires energy and many other resources which have to be provided by additional workers, so that the ratio of people fed to farmers is actually smaller than 100 to 1. More distinct terms are commonly used to denote farmers who raise specific domesticated animals. For example, those who raise grazing livestock, such as cattle, sheep, goats and horses, are known as ranchers (U.S.), graziers (Australia & UK) or simply stockmen. Sheep, goat and cattle farmers might also be referred to, respectively, as shepherds, goatherds and cowherds. The term dairy farmer is applied to those engaged primarily in milk production, whether from cattle, goats, sheep, or other milk producing animals. A poultry farmer is one who concentrates on raising chickens, turkeys, ducks or geese, for either meat, egg or feather production, or commonly, all three. A person who raises a variety of vegetables for market may be called a truck farmer or market gardener. Dirt farmer is an American colloquial term for a practical farmer, or one who farms his own land. In developed nations, a farmer (as a profession) is usually defined as someone with an ownership interest in crops or livestock, and who provides land or management in their production. Those who provide only labor are most often called farmhands. Alternatively, growers who manage farmland for an absentee landowner, sharing the harvest (or its profits) are known as sharecroppers or sharefarmers. In the context of agribusiness, a farmer is defined broadly, and thus many individuals not necessarily engaged in full-time farming can nonetheless legally qualify under agricultural policy for various subsidies, incentives, and tax deductions. In the context of developing nations or other pre-industrial cultures, most farmers practice a meager subsistence agriculture—a simple organic-farming system employing crop rotation, seed saving, slash and burn, or other techniques to maximize efficiency while meeting the needs of the household or community. One subsisting in this way may become labelled as a peasant, often associated disparagingly with a "peasant mentality". In developed nations, however, a person using such techniques on small patches of land might be called a gardener and be considered a hobbyist. Alternatively, one might be driven into such practices by poverty or, ironically—against the background of large-scale agribusiness—might become an organic farmer growing for discerning/faddish consumers in the local food market. Farmers are often members of local, regional, or national farmers' unions or agricultural producers' organizations and can exert significant political influence. The Grange movement in the United States was effective in advancing farmers' agendas, especially against railroad and agribusiness interests early in the 20th century. The FNSEA is very politically active in France, especially pertaining to genetically modified food. Agricultural producers, both small and large, are represented globally by the International Federation of Agricultural Producers (IFAP), representing over 600 million farmers through 120 national farmers' unions in 79 countries. There are many organizations that are targeted at teaching young people how to farm and advancing the knowledge and benefits of sustainable agriculture. Farmed products might be sold either to a market, in a farmers' market, or directly from a farm. In a subsistence economy, farm products might to some extent be either consumed by the farmer's family or pooled by the community. There are several occupational hazards for those in agriculture; farming is a particularly dangerous industry. Farmers can encounter and be stung or bitten by dangerous insects and other arthropods, including scorpions, fire ants, bees, wasps and hornets. Farmers also work around heavy machinery which can kill or injure them. Farmers can also establish muscle and joints pains from repeated work. The word 'farmer' originally meant a person collecting taxes from tenants working a field owned by a landlord. The word changed to refer to the person farming the field. Previous names for a farmer were churl and husbandman.
[[Image:Cancer constellation map-bs.svg|thumb|[[Cancer (constellation)|Cancer]] constellation map]] {{commons|category:Constellations|Constellations}} A '''constellation''' is a group of [[star]]s which make up imaginary outline or pattern in the night sky (the [[celestial sphere]]). Usually they are said to represent an animal, [[mythological]] person or creature in a shape. When seen, the group of stars seem to make a pattern. The word constellation comes from Latin: ''con-'', meaning together and ''stella-'' meaning stars. Some examples of constellations are [[Ursa Major|Ursa major]], [[Orion (constellation)|Orion,]] [[Taurus (constellation)|Taurus]], [[Draco (constellation)|Draco]], [[Cancer (constellation)|Cancer]] etc. Constellations were used to group stars. Different places in the world may have different constellations, but today [[astronomy]] has a fixed set of 88 constellations. This set is based on the [[Ancient Greece|Greek]] set and later some southern constellations were added, for example [[Antlia (constellation)|Antlia]] - the [[air]] [[pump]]. Most constellations have names that come from Greek mythology, like Orion or Andromeda. There are 12 constellations in the [[Zodiac]]. The [[Sun]] travels through the Zodiac once each year. There is also a thirteenth constellation [[Ophiuchus (constellation)|Ophiuchus]] - the carrier of a [[Serpens|serpent]], which the Sun goes through. However, most people do not think that it is in the Zodiac. ==History== *No one knows who first saw the constellations. Ancient civilizations, like the Mayans, drew their own [[star map]]s of the skies with their constellations, very few of which we use today. *[[Ptolemy]]'s 48 constellations are still recognized by the [[IAU]] today, and the rest of the constellations were added later {{Constellations/modern}} [[Category:Constellations| ]] {{sci-stub}}
Four views of the constellation Orion: A constellation is an area on the celestial sphere in which a group of visible stars forms a perceived pattern or outline, typically representing an animal, mythological subject, or inanimate object. The origins of the earliest constellations likely go back to prehistory. People used them to relate stories of their beliefs, experiences, creation, or mythology. Different cultures and countries invented their own constellations, some of which lasted into the early 20th century before today's constellations were internationally recognized. The recognition of constellations has changed significantly over time. Many changed in size or shape. Some became popular, only to drop into obscurity. Some were limited to a single culture or nation. Naming constellations also helped astronomers and navigators identify stars more easily. Twelve (or thirteen) ancient constellations belong to the zodiac (straddling the ecliptic, which the Sun, Moon, and planets all traverse). The origins of the zodiac remain historically uncertain; its astrological divisions became prominent c. 400 BC in Babylonian or Chaldean astronomy. Constellations appear in Western culture via Greece and are mentioned in the works of Hesiod, Eudoxus and Aratus. The traditional 48 constellations, consisting of the Zodiac and 36 more (now 38, following the division of Argo Navis into three constellations) are listed by Ptolemy, a Greco-Roman astronomer from Alexandria, Egypt, in his Almagest. The formation of constellations was the subject of extensive mythology, most notably in the Metamorphoses of the Latin poet Ovid. Constellations in the far southern sky were added from the 15th century until the mid-18th century when European explorers began traveling to the Southern Hemisphere. Due to Roman and European transmission, each constellation has a Latin name. In 1922, the International Astronomical Union (IAU) formally accepted the modern list of 88 constellations, and in 1928 adopted official constellation boundaries that together cover the entire celestial sphere. Any given point in a celestial coordinate system lies in one of the modern constellations. Some astronomical naming systems include the constellation where a given celestial object is found to convey its approximate location in the sky. The Flamsteed designation of a star, for example, consists of a number and the genitive form of the constellation's name. Other star patterns or groups called asterisms are not constellations under the formal definition, but are also used by observers to navigate the night sky. Asterisms may be several stars within a constellation, or they may share stars with more than one constellation. Examples of asterisms include the teapot within the constellation Sagittarius, or the big dipper in the constellation of Ursa Major. The word constellation comes from the Late Latin term cōnstellātiō, which can be translated as "set of stars"; it came into use in Middle English during the 14th century. The Ancient Greek word for constellation is ἄστρον (astron). These terms historically referred to any recognisable pattern of stars whose appearance was associated with mythological characters or creatures, earthbound animals, or objects. Over time, among European astronomers, the constellations became clearly defined and widely recognised. Today, there are 88 IAU designated constellations. A constellation or star that never sets below the horizon when viewed from a particular latitude on Earth is termed circumpolar. From the North Pole or South Pole, all constellations south or north of the celestial equator are circumpolar. Depending on the definition, equatorial constellations may include those that lie between declinations 45° north and 45° south, or those that pass through the declination range of the ecliptic or zodiac ranging between 23½° north, the celestial equator, and 23½° south. Stars in constellations can appear near each other in the sky, but they usually lie at a variety of distances away from the Earth. Since each star has its own independent motion, all constellations will change slowly over time. After tens to hundreds of thousands of years, familiar outlines will become unrecognizable. Astronomers can predict the past or future constellation outlines by measuring individual stars' common proper motions or cpm by accurate astrometry and their radial velocities by astronomical spectroscopy. The 88 constellations recognized by the International Astronomical Union as well as those that cultures have recognized throughout history are imagined figures and shapes derived from the patterns of stars in the observable sky. Many officially recognized constellations are based on the imaginations of ancient, Near Eastern and Mediterranean mythologies. H.A. Rey, who wrote popular books on astronomy, pointed out the imaginative nature of the constellations and their mythological and artistic basis, and the practical use of identifying them through definite images, according to the classical names they were given. It has been suggested that the 17,000-year-old cave paintings in Lascaux, southern France, depict star constellations such as Taurus, Orion's Belt, and the Pleiades. However, this view is not generally accepted among scientists. Inscribed stones and clay writing tablets from Mesopotamia (in modern Iraq) dating to 3000 BC provide the earliest generally accepted evidence for humankind's identification of constellations. It seems that the bulk of the Mesopotamian constellations were created within a relatively short interval from around 1300 to 1000 BC. Mesopotamian constellations appeared later in many of the classical Greek constellations. The oldest Babylonian catalogues of stars and constellations date back to the beginning of the Middle Bronze Age, most notably the Three Stars Each texts and the MUL.APIN, an expanded and revised version based on more accurate observation from around 1000 BC. However, the numerous Sumerian names in these catalogues suggest that they built on older, but otherwise unattested, Sumerian traditions of the Early Bronze Age. The classical Zodiac is a revision of Neo-Babylonian constellations from the 6th century BC. The Greeks adopted the Babylonian constellations in the 4th century BC. Twenty Ptolemaic constellations are from the Ancient Near East. Another ten have the same stars but different names. Biblical scholar E. W. Bullinger interpreted some of the creatures mentioned in the books of Ezekiel and Revelation as the middle signs of the four-quarters of the Zodiac, with the Lion as Leo, the Bull as Taurus, the Man representing Aquarius, and the Eagle standing in for Scorpio. The biblical Book of Job also makes reference to a number of constellations, including עיש ‘Ayish "bier", כסיל chesil "fool" and כימה chimah "heap" (Job 9:9, 38:31–32), rendered as "Arcturus, Orion and Pleiades" by the KJV, but ‘Ayish "the bier" actually corresponding to Ursa Major. The term Mazzaroth מַזָּרוֹת, translated as a garland of crowns, is a hapax legomenon in Job 38:32, and it might refer to the zodiacal constellations. There is only limited information on ancient Greek constellations, with some fragmentary evidence being found in the Works and Days of the Greek poet Hesiod, who mentioned the "heavenly bodies". Greek astronomy essentially adopted the older Babylonian system in the Hellenistic era, first introduced to Greece by Eudoxus of Cnidus in the 4th century BC. The original work of Eudoxus is lost, but it survives as a versification by Aratus, dating to the 3rd century BC. The most complete existing works dealing with the mythical origins of the constellations are by the Hellenistic writer termed pseudo-Eratosthenes and an early Roman writer styled pseudo-Hyginus. The basis of Western astronomy as taught during Late Antiquity and until the Early Modern period is the Almagest by Ptolemy, written in the 2nd century. In the Ptolemaic Kingdom, native Egyptian tradition of anthropomorphic figures represented the planets, stars, and various constellations. Some of these were combined with Greek and Babylonian astronomical systems culminating in the Zodiac of Dendera; it remains unclear when this occurred, but most were placed during the Roman period between 2nd to 4th centuries AD. The oldest known depiction of the zodiac showing all the now familiar constellations, along with some original Egyptian constellations, decans, and planets. Ptolemy's Almagest remained the standard definition of constellations in the medieval period both in Europe and in Islamic astronomy. Ancient China had a long tradition of observing celestial phenomena. Nonspecific Chinese star names, later categorized in the twenty-eight mansions, have been found on oracle bones from Anyang, dating back to the middle Shang dynasty. These constellations are some of the most important observations of Chinese sky, attested from the 5th century BC. Parallels to the earliest Babylonian (Sumerian) star catalogues suggest that the ancient Chinese system did not arise independently. Three schools of classical Chinese astronomy in the Han period are attributed to astronomers of the earlier Warring States period. The constellations of the three schools were conflated into a single system by Chen Zhuo, an astronomer of the 3rd century (Three Kingdoms period). Chen Zhuo's work has been lost, but information on his system of constellations survives in Tang period records, notably by Qutan Xida. The oldest extant Chinese star chart dates to that period and was preserved as part of the Dunhuang Manuscripts. Native Chinese astronomy flourished during the Song dynasty, and during the Yuan dynasty became increasingly influenced by medieval Islamic astronomy (see Treatise on Astrology of the Kaiyuan Era). As maps were prepared during this period on more scientific lines, they were considered as more reliable. A well-known map from the Song period is the Suzhou Astronomical Chart, which was prepared with carvings of stars on the planisphere of the Chinese sky on a stone plate; it is done accurately based on observations, and it shows the supernova of the year of 1054 in Taurus. Influenced by European astronomy during the late Ming dynasty, charts depicted more stars but retained the traditional constellations. Newly observed stars were incorporated as supplementary to old constellations in the southern sky, which did not depict the traditional stars recorded by ancient Chinese astronomers. Further improvements were made during the later part of the Ming dynasty by Xu Guangqi and Johann Adam Schall von Bell, the German Jesuit and was recorded in Chongzhen Lishu (Calendrical Treatise of Chongzhen period, 1628). Traditional Chinese star maps incorporated 23 new constellations with 125 stars of the southern hemisphere of the sky based on the knowledge of Western star charts; with this improvement, the Chinese Sky was integrated with the World astronomy. Ancient Greece A lot of well-known constellations also have histories that connect to ancient Greece. Historically, the origins of the constellations of the northern and southern skies are distinctly different. Most northern constellations date to antiquity, with names based mostly on Classical Greek legends. Evidence of these constellations has survived in the form of star charts, whose oldest representation appears on the statue known as the Farnese Atlas, based perhaps on the star catalogue of the Greek astronomer Hipparchus. Southern constellations are more modern inventions, sometimes as substitutes for ancient constellations (e.g. Argo Navis). Some southern constellations had long names that were shortened to more usable forms; e.g. Musca Australis became simply Musca. Some of the early constellations were never universally adopted. Stars were often grouped into constellations differently by different observers, and the arbitrary constellation boundaries often led to confusion as to which constellation a celestial object belonged. Before astronomers delineated precise boundaries (starting in the 19th century), constellations generally appeared as ill-defined regions of the sky. Today they now follow officially accepted designated lines of right ascension and declination based on those defined by Benjamin Gould in epoch 1875.0 in his star catalogue Uranometria Argentina. The 1603 star atlas "Uranometria" of Johann Bayer assigned stars to individual constellations and formalized the division by assigning a series of Greek and Latin letters to the stars within each constellation. These are known today as Bayer designations. Subsequent star atlases led to the development of today's accepted modern constellations. The southern sky, below about −65° declination, was only partially catalogued by ancient Babylonians, Egyptians, Greeks, Chinese, and Persian astronomers of the north. The knowledge that northern and southern star patterns differed goes back to Classical writers, who describe, for example, the African circumnavigation expedition commissioned by Egyptian Pharaoh Necho II in c. 600 BC and those of Hanno the Navigator in c. 500 BC. The history of southern constellations is not straightforward. Different groupings and different names were proposed by various observers, some reflecting national traditions or designed to promote various sponsors. Southern constellations were important from the 14th to 16th centuries, when sailors used the stars for celestial navigation. Italian explorers who recorded new southern constellations include Andrea Corsali, Antonio Pigafetta, and Amerigo Vespucci. Many of the 88 IAU-recognized constellations in this region first appeared on celestial globes developed in the late 16th century by Petrus Plancius, based mainly on observations of the Dutch navigators Pieter Dirkszoon Keyser and Frederick de Houtman. These became widely known through Johann Bayer's star atlas Uranometria of 1603. Fourteen more were created in 1763 by the French astronomer Nicolas Louis de Lacaille, who also split the ancient constellation Argo Navis into three; these new figures appeared in his star catalogue, published in 1756. Several modern proposals have not survived. The French astronomers Pierre Lemonnier and Joseph Lalande, for example, proposed constellations that were once popular but have since been dropped. The northern constellation Quadrans Muralis survived into the 19th century (when its name was attached to the Quadrantid meteor shower), but is now divided between Boötes and Draco. A list of 88 constellations was produced for the International Astronomical Union in 1922. It is roughly based on the traditional Greek constellations listed by Ptolemy in his Almagest in the 2nd century and Aratus' work Phenomena, with early modern modifications and additions (most importantly introducing constellations covering the parts of the southern sky unknown to Ptolemy) by Petrus Plancius (1592, 1597/98 and 1613), Johannes Hevelius (1690) and Nicolas Louis de Lacaille (1763), who introduced fourteen new constellations. Lacaille studied the stars of the southern hemisphere from 1751 until 1752 from the Cape of Good Hope, when he was said to have observed more than 10,000 stars using a refracting telescope with an aperture of 0.5 inches (13 mm). In 1922, Henry Norris Russell produced a list of 88 constellations with three-letter abbreviations for them. However, these constellations did not have clear borders between them. In 1928, the International Astronomical Union (IAU) formally accepted 88 modern constellations, with contiguous boundaries along vertical and horizontal lines of right ascension and declination developed by Eugene Delporte that, together, cover the entire celestial sphere; this list was finally published in 1930. Where possible, these modern constellations usually share the names of their Graeco-Roman predecessors, such as Orion, Leo or Scorpius. The aim of this system is area-mapping, i.e. the division of the celestial sphere into contiguous fields. Out of the 88 modern constellations, 36 lie predominantly in the northern sky, and the other 52 predominantly in the southern. The boundaries developed by Delporte used data that originated back to epoch B1875.0, which was when Benjamin A. Gould first made his proposal to designate boundaries for the celestial sphere, a suggestion on which Delporte based his work. The consequence of this early date is that because of the precession of the equinoxes, the borders on a modern star map, such as epoch J2000, are already somewhat skewed and no longer perfectly vertical or horizontal. This effect will increase over the years and centuries to come. The constellations have no official symbols, though those of the ecliptic may take the signs of the zodiac. Symbols for the other modern constellations, as well as older ones that still occur in modern nomenclature, have occasionally been published. The Great Rift, a series of dark patches in the Milky Way, is more visible and striking in the southern hemisphere than in the northern. It vividly stands out when conditions are otherwise so dark that the Milky Way's central region casts shadows on the ground. Some cultures have discerned shapes in these patches and have given names to these "dark cloud constellations". Members of the Inca civilization identified various dark areas or dark nebulae in the Milky Way as animals and associated their appearance with the seasonal rains. Australian Aboriginal astronomy also describes dark cloud constellations, the most famous being the "emu in the sky" whose head is formed by the Coalsack, a dark nebula, instead of the stars.
A '''flood''' is an overflow of water. Floods are very hard to deal with. ==Overview== Floods are most commonly made due to an overflowing [[river]], a [[dam]] break, [[snowmelt]], or heavy rainfall. Less commonly happening are [[tsunami]]s, [[storm surge]]. The most deadly flooding was in 1931 in China and killed between 2,000,000 and 4,000,000 people.<ref>O'Connor, Jim E. and John E. Costa. 2004. The world's largest floods, past and present: their causes and magnitudes [Circular 1254]. Washington, D.C.: U.S. Department of the Interior, U.S. Geological Survey.</ref> The [[Kerala flood]] in India was another flood that has destroyed people's houses. During a flood, people try to move themselves and their most precious belongings to higher ground quickly. The process of leaving homes in search of a safe place is called a [[Evacuation|flood evacuation]]. Floods have inspired [[myth]]s and [[legend]]s, especially the story of the [[great flood]].{{citation needed|date=November 2024}} ==Pollution of drinking water== [[File:Alicante(30-09-1997).JPG|thumb|right|Autumn [[Mediterranean sea|Mediterranean]] flooding in [[Alicante]], [[Spain]], September 1997]]During a flood there is plenty of water logging and overflow of water; it is mostly [[Water pollution|polluted]] and not safe to [[drink]]. If people drink this dirty water, they may suffer from [[disease]]s such as [[typhoid]] and [[cholera]], [[hepatitis]] and other such diseases. People can get ready to survive a flood by filling many [[container]]s with fresh and clean [[drinking water]] and storing other [[emergency]] supplies like medicine and food. During a flood, people try to go to higher sides because the flood water cannot reach high areas. Also drains overflow and mix with clean water and people who drink it might fall ill.{{citation needed|date=November 2024}} ==Causes== Flooding is usually caused when a [[volume]] of water within a water body, such as a [[lake]], overflows outside it. Sometimes if a [[dam]] breaks, it suddenly releases a large amount of water. The result is that some of the water travels to land, and "floods" the area. Many rivers are in a channel, between [[river]] banks. They flood when the strength of the river causes it to flow beyond the banks. This is more common at bends or [[Meander|meanders]]. Flood damage can be prevented by moving away from places that flood. However, people have long liked to have their homes and businesses alongside water because water is good for [[agriculture]] and [[transport]] and in other ways. Floods are also caused due to improper management of drains.{{citation needed|date=January 2025}} ===Rainfall=== [[Rainfall]] is the most common cause. Snow melt is also a cause of flooding. [[Tsunami]]s and Storm Surge are less common ways that floods happen. Coastal Flooding is another common cause of flooding, and this is caused by [[Storm|low pressure systems or storms]]. There are many ways to control floods, such as improving the drainage and grading in a home, or installing a flood detection system.<ref>{{Cite web|date=2019-10-09|title=Protect Your Home From Flooding {{!}} ALARM-i Video Verification|url=https://alarm-i.ca/protect-your-home-from-flooding/|access-date=2021-03-23|language=en-US|archive-date=2021-01-27|archive-url=https://web.archive.org/web/20210127041323/https://alarm-i.ca/protect-your-home-from-flooding/|url-status=dead}}</ref> ==Related pages== * [[Noah]] * [[Deluge myth]] * [[Flood control]] * [[Natural disaster]] ==References== {{Reflist}} {{commonscat|Floods}} [[Category:Floods| ]] [[Category:Disasters]] [[Category:Natural disasters]]
A flood is an overflow of water (or rarely other fluids) that submerges land that is usually dry. In the sense of "flowing water", the word may also be applied to the inflow of the tide. Floods are an area of study of the discipline hydrology and are of significant concern in agriculture, civil engineering and public health. Human changes to the environment often increase the intensity and frequency of flooding, for example land use changes such as deforestation and removal of wetlands, changes in waterway course or flood controls such as with levees, and larger environmental issues such as climate change and sea level rise. In particular climate change's increased rainfall and extreme weather events increases the severity of other causes for flooding, resulting in more intense floods and increased flood risk. Flooding may occur as an overflow of water from water bodies, such as a river, lake, or ocean, in which the water overtops or breaks levees, resulting in some of that water escaping its usual boundaries, or it may occur due to an accumulation of rainwater on saturated ground in an areal flood. While the size of a lake or other body of water will vary with seasonal changes in precipitation and snow melt, these changes in size are unlikely to be considered significant unless they flood property or drown domestic animals. Floods can also occur in rivers when the flow rate exceeds the capacity of the river channel, particularly at bends or meanders in the waterway. Floods often cause damage to homes and businesses if they are in the natural flood plains of rivers. While riverine flood damage can be eliminated by moving away from rivers and other bodies of water, people have traditionally lived and worked by rivers because the land is usually flat and fertile and because rivers provide easy travel and access to commerce and industry. Flooding can lead to secondary consequences in addition to damage to property, such as long-term displacement of residents and creating increased spread of waterborne diseases and vector-bourne disesases transmitted by mosquitos. Floods can happen on flat or low-lying areas when water is supplied by rainfall or snowmelt more rapidly than it can either infiltrate or run off. The excess accumulates in place, sometimes to hazardous depths. Surface soil can become saturated, which effectively stops infiltration, where the water table is shallow, such as a floodplain, or from intense rain from one or a series of storms. Infiltration also is slow to negligible through frozen ground, rock, concrete, paving, or roofs. Areal flooding begins in flat areas like floodplains and in local depressions not connected to a stream channel, because the velocity of overland flow depends on the surface slope. Endorheic basins may experience areal flooding during periods when precipitation exceeds evaporation. Floods occur in all types of river and stream channels, from the smallest ephemeral streams in humid zones to normally-dry channels in arid climates to the world's largest rivers. When overland flow occurs on tilled fields, it can result in a muddy flood where sediments are picked up by run off and carried as suspended matter or bed load. Localized flooding may be caused or exacerbated by drainage obstructions such as landslides, ice, debris, or beaver dams. Slow-rising floods most commonly occur in large rivers with large catchment areas. The increase in flow may be the result of sustained rainfall, rapid snow melt, monsoons, or tropical cyclones. However, large rivers may have rapid flooding events in areas with dry climates, since they may have large basins but small river channels, and rainfall can be very intense in smaller areas of those basins. Rapid flooding events, including flash floods, more often occur on smaller rivers, rivers with steep valleys, rivers that flow for much of their length over impermeable terrain, or normally-dry channels. The cause may be localized convective precipitation (intense thunderstorms) or sudden release from an upstream impoundment created behind a dam, landslide, or glacier. In one instance, a flash flood killed eight people enjoying the water on a Sunday afternoon at a popular waterfall in a narrow canyon. Without any observed rainfall, the flow rate increased from about 50 to 1,500 cubic feet per second (1.4 to 42 m/s) in just one minute. Two larger floods occurred at the same site within a week, but no one was at the waterfall on those days. The deadly flood resulted from a thunderstorm over part of the drainage basin, where steep, bare rock slopes are common and the thin soil was already saturated. Flash floods are the most common flood type in normally-dry channels in arid zones, known as arroyos in the southwest United States and many other names elsewhere. In that setting, the first flood water to arrive is depleted as it wets the sandy stream bed. The leading edge of the flood thus advances more slowly than later and higher flows. As a result, the rising limb of the hydrograph becomes ever quicker as the flood moves downstream, until the flow rate is so great that the depletion by wetting soil becomes insignificant. Coastal areas may be flooded by storm surges combining with high tides and large wave events at sea, resulting in waves over-topping flood defenses or in severe cases by tsunami or tropical cyclones. A storm surge, from either a tropical cyclone or an extratropical cyclone, falls within this category. A storm surge is "an additional rise of water generated by a storm, over and above the predicted astronomical tides". Due to the effects of climate change (e.g. sea level rise and an increase in extreme weather events) and an increase in the population living in coastal areas, the damage caused by coastal flood events has intensified and more people are being affected. Flooding in estuaries is commonly caused by a combination of storm surges caused by winds and low barometric pressure and large waves meeting high upstream river flows. Urban flooding is the inundation of land or property in a built environment, particularly in more densely populated areas, caused by rainfall overwhelming the capacity of drainage systems, such as storm sewers. Although sometimes triggered by events such as flash flooding or snowmelt, urban flooding is a condition, characterized by its repetitive and systemic impacts on communities, that can happen regardless of whether or not affected communities are located within designated floodplains or near any body of water. Aside from potential overflow of rivers and lakes, snowmelt, stormwater or water released from damaged water mains may accumulate on property and in public rights-of-way, seep through building walls and floors, or backup into buildings through sewer pipes, toilets and sinks. Catastrophic riverine flooding can result from major infrastructure failures, often the collapse of a dam. It can also be caused by drainage channel modification from a landslide, earthquake or volcanic eruption. Examples include outburst floods and lahars. Tsunamis can cause catastrophic coastal flooding, most commonly resulting from undersea earthquakes. Floods are caused by many factors or a combination of any of these generally prolonged heavy rainfall (locally concentrated or throughout a catchment area), highly accelerated snowmelt, severe winds over water, unusual high tides, tsunamis, or failure of dams, levees, retention ponds, or other structures that retained the water. Flooding can be exacerbated by increased amounts of impervious surface or by other natural hazards such as wildfires, which reduce the supply of vegetation that can absorb rainfall. During times of rain, some of the water is retained in ponds or soil, some is absorbed by grass and vegetation, some evaporates, and the rest travels over the land as surface runoff. Floods occur when ponds, lakes, riverbeds, soil, and vegetation cannot absorb all the water. This has been exacerbated by human activities such as draining wetlands that naturally store large amounts of water and building paved surfaces that do not absorb any water. Water then runs off the land in quantities that cannot be carried within stream channels or retained in natural ponds, lakes, and human-made reservoirs. About 30 percent of all precipitation becomes runoff and that amount might be increased by water from melting snow. River flooding is often caused by heavy rain, sometimes increased by melting snow. A flood that rises rapidly, with little or no warning, is called a flash flood. Flash floods usually result from intense rainfall over a relatively small area, or if the area was already saturated from previous precipitation. Periodic floods occur on many rivers, forming a surrounding region known as the flood plain. Even when rainfall is relatively light, the shorelines of lakes and bays can be flooded by severe winds—such as during hurricanes—that blow water into the shore areas. The amount, location, and timing of water reaching a drainage channel from natural precipitation and controlled or uncontrolled reservoir releases determines the flow at downstream locations. Some precipitation evaporates, some slowly percolates through soil, some may be temporarily sequestered as snow or ice, and some may produce rapid runoff from surfaces including rock, pavement, roofs, and saturated or frozen ground. The fraction of incident precipitation promptly reaching a drainage channel has been observed from nil for light rain on dry, level ground to as high as 170 percent for warm rain on accumulated snow. Most precipitation records are based on a measured depth of water received within a fixed time interval. Frequency of a precipitation threshold of interest may be determined from the number of measurements exceeding that threshold value within the total time period for which observations are available. Individual data points are converted to intensity by dividing each measured depth by the period of time between observations. This intensity will be less than the actual peak intensity if the duration of the rainfall event was less than the fixed time interval for which measurements are reported. Convective precipitation events (thunderstorms) tend to produce shorter duration storm events than orographic precipitation. Duration, intensity, and frequency of rainfall events are important to flood prediction. Short duration precipitation is more significant to flooding within small drainage basins. The most important upslope factor in determining flood magnitude is the land area of the watershed upstream of the area of interest. Rainfall intensity is the second most important factor for watersheds of less than approximately 30 square miles or 80 square kilometres. The main channel slope is the second most important factor for larger watersheds. Channel slope and rainfall intensity become the third most important factors for small and large watersheds, respectively. Time of Concentration is the time required for runoff from the most distant point of the upstream drainage area to reach the point of the drainage channel controlling flooding of the area of interest. The time of concentration defines the critical duration of peak rainfall for the area of interest. The critical duration of intense rainfall might be only a few minutes for roof and parking lot drainage structures, while cumulative rainfall over several days would be critical for river basins. Water flowing downhill ultimately encounters downstream conditions slowing movement. The final limitation in coastal flooding lands is often the ocean or some coastal flooding bars which form natural lakes. In flooding low lands, elevation changes such as tidal fluctuations are significant determinants of coastal and estuarine flooding. Less predictable events like tsunamis and storm surges may also cause elevation changes in large bodies of water. Elevation of flowing water is controlled by the geometry of the flow channel and, especially, by depth of channel, speed of flow and amount of sediments in it Flow channel restrictions like bridges and canyons tend to control water elevation above the restriction. The actual control point for any given reach of the drainage may change with changing water elevation, so a closer point may control for lower water levels until a more distant point controls at higher water levels. Effective flood channel geometry may be changed by growth of vegetation, accumulation of ice or debris, or construction of bridges, buildings, or levees within the flood channel. Extreme flood events often result from coincidence such as unusually intense, warm rainfall melting heavy snow pack, producing channel obstructions from floating ice, and releasing small impoundments like beaver dams. Coincident events may cause extensive flooding to be more frequent than anticipated from simplistic statistical prediction models considering only precipitation runoff flowing within unobstructed drainage channels. Debris modification of channel geometry is common when heavy flows move uprooted woody vegetation and flood-damaged structures and vehicles, including boats and railway equipment. Recent field measurements during the 2010–11 Queensland floods showed that any criterion solely based upon the flow velocity, water depth or specific momentum cannot account for the hazards caused by velocity and water depth fluctuations. These considerations ignore further the risks associated with large debris entrained by the flow motion. Some researchers have mentioned the storage effect in urban areas with transportation corridors created by cut and fill. Culverted fills may be converted to impoundments if the culverts become blocked by debris, and flow may be diverted along streets. Several studies have looked into the flow patterns and redistribution in streets during storm events and the implication on flood modelling. The intentional flooding of land that would otherwise remain dry may take place for military, agricultural, or river-management purposes. This is a form of hydraulic engineering. Agricultural flooding may occur in preparing paddy fields for the growing of semi-aquatic rice in many countries. Flooding for river management may occur in the form of diverting flood waters in a river at flood stage upstream from areas that are considered more valuable than the areas that are sacrificed in this way. This may be done ad hoc, as in the 2011 intentional breach of levees by the United States Army Corps of Engineers in Missouri, or permanently, as in the so-called overlaten (literally "let-overs"), an intentionally lowered segment in Dutch riparian levees, like the Beerse Overlaat in the left levee of the Meuse between the villages of Gassel and Linden, North Brabant. Military inundation creates an obstacle in the field that is intended to impede the movement of the enemy. This may be done both for offensive and defensive purposes. Furthermore, in so far as the methods used are a form of hydraulic engineering, it may be useful to differentiate between controlled inundations (as in most historic inundations in the Netherlands under the Dutch Republic and its successor states in that area and exemplified in the two Hollandic Water Lines, the Stelling van Amsterdam, the Frisian Water Line, the IJssel Line, the Peel-Raam Line, and the Grebbe line in that country) and uncontrolled ones (as in the second Siege of Leiden during the first part of the Eighty Years' War, the flooding of the Yser plain during the First World War, and the Inundation of Walcheren, and the Inundation of the Wieringermeer during the Second World War). To count as controlled, a military inundation has to take the interests of the civilian population into account, by allowing them a timely evacuation, by making the inundation reversible, and by making an attempt to minimize the adverse ecological impact of the inundation. That impact may also be adverse in a hydrogeological sense if the inundation lasts a long time. The Itaipu dam caused concern that in times of conflict could be used as a weapon to flood Buenos Aires. Floods can also be a huge destructive power. When water flows, it has the ability to demolish all kinds of buildings and objects, such as bridges, structures, houses, trees, and cars. Economical, social and natural environmental damages are common factors that are impacted by flooding events and the impacts that flooding has on these areas can be catastrophic. There have been numerous flood incidents around the world which have caused devastating damage to infrastructure, the natural environment and human life. Flood risks can be defined as the risk that floods pose to individuals, property and the natural landscape based on specific hazards and vulnerability. The extent of flood risks can impact the types of mitigation strategies required and implemented. Floods can have devastating impacts to human societies. Flooding events worldwide are increasing in frequency and severity, leading to increasing costs to societies. A large amount of the world's population lives in close proximity to major coastlines, while many major cities and agricultural areas are located near floodplains. There is significant risk for increased coastal and fluvial flooding due to changing climatic conditions. The primary effects of flooding include loss of life and damage to buildings and other structures, including bridges, sewerage systems, roadways, and canals. The economic impacts caused by flooding can be severe. Every year flooding causes countries billions of dollars worth of damage that threatens the livelihood of individuals. As a result, there is also significant socio-economic threats to vulnerable populations around the world from flooding. For example, in Bangladesh in 2007, a flood was responsible for the destruction of more than one million houses. And yearly in the United States, floods cause over $7 billion in damage. Flood waters typically inundate farm land, making the land unworkable and preventing crops from being planted or harvested, which can lead to shortages of food both for humans and farm animals. Entire harvests for a country can be lost in extreme flood circumstances. Some tree species may not survive prolonged flooding of their root systems. Urban flooding also has significant economic implications for affected neighborhoods. In the United States, industry experts estimate that wet basements can lower property values by 10–25 percent and are cited among the top reasons for not purchasing a home. According to the U.S. Federal Emergency Management Agency (FEMA), almost 40 percent of small businesses never reopen their doors following a flooding disaster. In the United States, insurance is available against flood damage to both homes and businesses. Economic hardship due to a temporary decline in tourism, rebuilding costs, or food shortages leading to price increases is a common after-effect of severe flooding. The impact on those affected may cause psychological damage to those affected, in particular where deaths, serious injuries and loss of property occur. Fatalities connected directly to floods are usually caused by drowning; the waters in a flood are very deep and have strong currents. Deaths do not just occur from drowning, deaths are connected with dehydration, heat stroke, heart attack and any other illness that needs medical supplies that cannot be delivered. Injuries can lead to an excessive amount of morbidity when a flood occurs. Injuries are not isolated to just those who were directly in the flood, rescue teams and even people delivering supplies can sustain an injury. Injuries can occur anytime during the flood process; before, during and after. During floods accidents occur with falling debris or any of the many fast moving objects in the water. After the flood rescue attempts are where large numbers injuries can occur. Communicable diseases are increased due to many pathogens and bacteria that are being transported by the water.There are many waterborne diseases such as cholera, hepatitis A, hepatitis E and diarrheal diseases, to mention a few. Gastrointestinal disease and diarrheal diseases are very common due to a lack of clean water during a flood. Most of clean water supplies are contaminated when flooding occurs. Hepatitis A and E are common because of the lack of sanitation in the water and in living quarters depending on where the flood is and how prepared the community is for a flood. When floods hit, people lose nearly all their crops, livestock, and food reserves and face starvation. Floods also frequently damage power transmission and sometimes power generation, which then has knock-on effects caused by the loss of power. This includes loss of drinking water treatment and water supply, which may result in loss of drinking water or severe water contamination. It may also cause the loss of sewage disposal facilities. Lack of clean water combined with human sewage in the flood waters raises the risk of waterborne diseases, which can include typhoid, giardia, cryptosporidium, cholera and many other diseases depending upon the location of the flood. Damage to roads and transport infrastructure may make it difficult to mobilize aid to those affected or to provide emergency health treatment. Urban flooding can cause chronically wet houses, leading to the growth of indoor mold and resulting in adverse health effects, particularly respiratory symptoms. Respiratory diseases are a common after the disaster has occurred. This depends on the amount of water damage and mold that grows after an incident. Research suggests that there will be an increase of 30–50% in adverse respiratory health outcomes caused by dampness and mold exposure for those living in coastal and wetland areas. Fungal contamination in homes is associated with increased allergic rhinitis and asthma. Vector borne diseases increase as well due to the increase in still water after the floods have settled. The diseases that are vector borne are malaria, dengue, West Nile, and yellow fever. Floods have a huge impact on victims' psychosocial integrity. People suffer from a wide variety of losses and stress. One of the most treated illness in long-term health problems are depression caused by the flood and all the tragedy that flows with one. Below is a list of the deadliest floods worldwide, showing events with death tolls at or above 100,000 individuals. Floods (in particular more frequent or smaller floods) can also bring many benefits, such as recharging ground water, making soil more fertile and increasing nutrients in some soils. Flood waters provide much needed water resources in arid and semi-arid regions where precipitation can be very unevenly distributed throughout the year and kills pests in the farming land. Freshwater floods particularly play an important role in maintaining ecosystems in river corridors and are a key factor in maintaining floodplain biodiversity. Flooding can spread nutrients to lakes and rivers, which can lead to increased biomass and improved fisheries for a few years. For some fish species, an inundated floodplain may form a highly suitable location for spawning with few predators and enhanced levels of nutrients or food. Fish, such as the weather fish, make use of floods in order to reach new habitats. Bird populations may also profit from the boost in food production caused by flooding. Flooding can bring benefits, such as making the soil more fertile and providing it with more nutrients. For this reason, periodic flooding was essential to the well-being of ancient communities along the Tigris-Euphrates Rivers, the Nile River, the Indus River, the Ganges and the Yellow River among others. The viability of hydropower, a renewable source of energy, is also higher in flood prone regions. In many countries around the world, waterways prone to floods are often carefully managed. Defenses such as detention basins, levees, bunds, reservoirs, and weirs are used to prevent waterways from overflowing their banks. When these defenses fail, emergency measures such as sandbags or portable inflatable tubes are often used to try to stem flooding. Coastal flooding has been addressed in portions of Europe and the Americas with coastal defenses, such as sea walls, beach nourishment, and barrier islands. In the riparian zone near rivers and streams, erosion control measures can be taken to try to slow down or reverse the natural forces that cause many waterways to meander over long periods of time. Flood controls, such as dams, can be built and maintained over time to try to reduce the occurrence and severity of floods as well. In the United States, the U.S. Army Corps of Engineers maintains a network of such flood control dams. In areas prone to urban flooding, one solution is the repair and expansion of human-made sewer systems and stormwater infrastructure. Another strategy is to reduce impervious surfaces in streets, parking lots and buildings through natural drainage channels, porous paving, and wetlands (collectively known as green infrastructure or sustainable urban drainage systems (SUDS)). Areas identified as flood-prone can be converted into parks and playgrounds that can tolerate occasional flooding. Ordinances can be adopted to require developers to retain stormwater on site and require buildings to be elevated, protected by floodwalls and levees, or designed to withstand temporary inundation. Property owners can also invest in solutions themselves, such as re-landscaping their property to take the flow of water away from their building and installing rain barrels, sump pumps, and check valves. In some areas, the presence of certain species (such as beavers) can be beneficial for flood control reasons. Beavers build and maintain beaver dams which will reduce the height of flood waves moving down the river (during periods of heavy rains), and will reduce or eliminate damage to human structures, at the cost of minor flooding near the dams (often on farmland). Besides this, they also boost wildlife populations and filter pollutants (manure, fertilisers, slurry). UK environment minister Rebecca Pow stated that in the future the beavers could be considered a "public good" and landowners would be paid to have them on their land. Flood control (or flood mitigation or flood protection or flood alleviation) methods are used to reduce or prevent the detrimental effects of flood waters. Flood relief methods are used to reduce the effects of flood waters or high water levels. Flooding can be caused by a mix of both natural processes, such as extreme weather upstream, and human changes to waterbodies and runoff. A distinction is made between structural and non-structural flood control measures. Structural methods physically restrain the flood waters, whereas non-structural methods do not. Building hard infrastructure to prevent flooding, such as flood walls, is effective at managing flooding. However, increased best practice within landscape engineering is to rely more on soft infrastructure and natural systems, such as marshes and flood plains, for handling the increase in water. To prevent or manage coastal flooding, coastal management practices have to handle natural processes like tides but also the human cased sea level rise. Flood control and relief is a particularly important part of climate change adaptation and climate resilience. Both sea level rise and changes in the weather (climate change causes more intense and quicker rainfall) mean that flooding of human infrastructure is particularly important the world over. In the United States, the National Weather Service gives out the advice "Turn Around, Don't Drown" for floods; that is, it recommends that people get out of the area of a flood, rather than trying to cross it. At the most basic level, the best defense against floods is to seek higher ground for high-value uses while balancing the foreseeable risks with the benefits of occupying flood hazard zones. Critical community-safety facilities, such as hospitals, emergency-operations centers, and police, fire, and rescue services, should be built in areas least at risk of flooding. Structures, such as bridges, that must unavoidably be in flood hazard areas should be designed to withstand flooding. Areas most at risk for flooding could be put to valuable uses that could be abandoned temporarily as people retreat to safer areas when a flood is imminent. Planning for flood safety involves many aspects of analysis and engineering, including: Each topic presents distinct yet related questions with varying scope and scale in time, space, and the people involved. Attempts to understand and manage the mechanisms at work in floodplains have been made for at least six millennia. In the United States, the Association of State Floodplain Managers works to promote education, policies, and activities that mitigate current and future losses, costs, and human suffering caused by flooding and to protect the natural and beneficial functions of floodplains – all without causing adverse impacts. A portfolio of best practice examples for disaster mitigation in the United States is available from the Federal Emergency Management Agency. Clean-up activities following floods often pose hazards to workers and volunteers involved in the effort. Potential dangers include electrical hazards, carbon monoxide exposure, musculoskeletal hazards, heat or cold stress, motor vehicle-related dangers, fire, drowning, and exposure to hazardous materials. Because flooded disaster sites are unstable, clean-up workers might encounter sharp jagged debris, biological hazards in the flood water, exposed electrical lines, blood or other body fluids, and animal and human remains. In planning for and reacting to flood disasters, managers provide workers with hard hats, goggles, heavy work gloves, life jackets, and watertight boots with steel toes and insoles. A series of annual maximum flow rates in a stream reach can be analyzed statistically to estimate the 100-year flood and floods of other recurrence intervals there. Similar estimates from many sites in a hydrologically similar region can be related to measurable characteristics of each drainage basin to allow indirect estimation of flood recurrence intervals for stream reaches without sufficient data for direct analysis. Physical process models of channel reaches are generally well understood and will calculate the depth and area of inundation for given channel conditions and a specified flow rate, such as for use in floodplain mapping and flood insurance. Conversely, given the observed inundation area of a recent flood and the channel conditions, a model can calculate the flow rate. Applied to various potential channel configurations and flow rates, a reach model can contribute to selecting an optimum design for a modified channel. Various reach models are available as of 2015, either 1D models (flood levels measured in the channel) or 2D models (variable flood depths measured across the extent of a floodplain). HEC-RAS, the Hydraulic Engineering Center model, is among the most popular software, if only because it is available free of charge. Other models such as TUFLOW combine 1D and 2D components to derive flood depths across both river channels and the entire floodplain. Physical process models of complete drainage basins are even more complex. Although many processes are well understood at a point or for a small area, others are poorly understood at all scales, and process interactions under normal or extreme climatic conditions may be unknown. Basin models typically combine land-surface process components (to estimate how much rainfall or snowmelt reaches a channel) with a series of reach models. For example, a basin model can calculate the runoff hydrograph that might result from a 100-year storm, although the recurrence interval of a storm is rarely equal to that of the associated flood. Basin models are commonly used in flood forecasting and warning, as well as in analysis of the effects of land use change and climate change. Anticipating floods before they occur allows for precautions to be taken and people to be warned so that they can be prepared in advance for flooding conditions. For example, farmers can remove animals from low-lying areas and utility services can put in place emergency provisions to re-route services if needed. Emergency services can also make provisions to have enough resources available ahead of time to respond to emergencies as they occur. People can evacuate areas to be flooded. In order to make the most accurate flood forecasts for waterways, it is best to have a long time-series of historical data that relates stream flows to measured past rainfall events. Coupling this historical information with real-time knowledge about volumetric capacity in catchment areas, such as spare capacity in reservoirs, ground-water levels, and the degree of saturation of area aquifers is also needed in order to make the most accurate flood forecasts. Radar estimates of rainfall and general weather forecasting techniques are also important components of good flood forecasting. In areas where good quality data is available, the intensity and height of a flood can be predicted with fairly good accuracy and plenty of lead time. The output of a flood forecast is typically a maximum expected water level and the likely time of its arrival at key locations along a waterway, and it also may allow for the computation of the likely statistical return period of a flood. In many developed countries, urban areas at risk of flooding are protected against a 100-year flood – that is a flood that has a probability of around 63% of occurring in any 100-year period of time. According to the U.S. National Weather Service (NWS) Northeast River Forecast Center (RFC) in Taunton, Massachusetts, a rule of thumb for flood forecasting in urban areas is that it takes at least 1 inch (25 mm) of rainfall in around an hour's time in order to start significant ponding of water on impermeable surfaces. Many NWS RFCs routinely issue Flash Flood Guidance and Headwater Guidance, which indicate the general amount of rainfall that would need to fall in a short period of time in order to cause flash flooding or flooding on larger water basins. In the United States, an integrated approach to real-time hydrologic computer modelling uses observed data from the U.S. Geological Survey (USGS), various cooperative observing networks, various automated weather sensors, the NOAA National Operational Hydrologic Remote Sensing Center (NOHRSC), various hydroelectric companies, etc. combined with quantitative precipitation forecasts (QPF) of expected rainfall and/or snow melt to generate daily or as-needed hydrologic forecasts. The NWS also cooperates with Environment Canada on hydrologic forecasts that affect both the US and Canada, like in the area of the Saint Lawrence Seaway. The Global Flood Monitoring System, "GFMS", a computer tool which maps flood conditions worldwide, is available online. Users anywhere in the world can use GFMS to determine when floods may occur in their area. GFMS uses precipitation data from NASA's Earth observing satellites and the Global Precipitation Measurement satellite, "GPM". Rainfall data from GPM is combined with a land surface model that incorporates vegetation cover, soil type, and terrain to determine how much water is soaking into the ground, and how much water is flowing into streamflow. Users can view statistics for rainfall, streamflow, water depth, and flooding every 3 hours, at each 12-kilometer gridpoint on a global map. Forecasts for these parameters are 5 days into the future. Users can zoom in to see inundation maps (areas estimated to be covered with water) in 1-kilometer resolution. A flood myth or a deluge myth is a myth in which a great flood, usually sent by a deity or deities, destroys civilization, often in an act of divine retribution. Parallels are often drawn between the flood waters of these myths and the primaeval waters which appear in certain creation myths, as the flood waters are described as a measure for the cleansing of humanity, in preparation for rebirth. Most flood myths also contain a culture hero, who "represents the human craving for life". The word "flood" comes from the Old English flōd, a word common to Germanic languages (compare German Flut, Dutch vloed from the same root as is seen in flow, float; also compare with Latin fluctus, flumen), meaning "a flowing of water, tide, an overflowing of land by water, a deluge, Noah's Flood; mass of water, river, sea, wave,". The Old English word flōd comes from the Proto-Germanic floduz (Old Frisian flod, Old Norse floð, Middle Dutch vloet, Dutch vloed, German Flut, and Gothic flodus derives from floduz).
#REDIRECT [[download]]
https://www.google.com/search?client=ms-android-xiaomi-rvo2&sca_esv=589713028&cds=2&cs=0&hl=en-IN&v=12.18.11.23.arm&output=search&q=%E0%A6%B0%E0%A6%98%E0%A7%81%E0%A6%A8%E0%A6%BE%E0%A6%A5%E0%A6%AA%E0%A7%81%E0%A6%B0+%E0%A6%86%E0%A6%A8%E0%A7%8D%E0%A6%A7%E0%A6%BE%E0%A6%B0%E0%A6%95%E0%A7%8B%E0%A6%A0%E0%A6%BE%E2%80%8C+%E0%A6%AC%E0%A6%BE%E0%A6%B2%E0%A6%BE%E0%A6%AC%E0%A6%BE%E0%A6%A1%E0%A6%BC%E0%A6%BF+%E0%A6%B8%E0%A6%BE%E0%A6%B0%E0%A7%8D%E0%A6%AC%E0%A6%9C%E0%A6%A8%E0%A7%80%E0%A6%A8+%E0%A6%B6%E0%A7%8D%E0%A6%B0%E0%A7%80+%E0%A6%B6%E0%A7%8D%E0%A6%B0%E0%A7%80+%E0%A6%B9%E0%A6%B0%E0%A6%BF+%E0%A6%AE%E0%A6%A8%E0%A7%8D%E0%A6%A6%E0%A6%BF%E0%A6%B0&ludocid=11111026167602635361&ibp=gwp;0,7&gsas=1&lsig=AB86z5Uj7oOdhTPp7uvmBgL_vv26&kgs=67aceeeaf6523e41&shndl=-1&shem=lcspc,lsp&source=sh/x/loc/hdr/m1/4#lpg=cid:CgIgAQ%3D%3D refers to transmitting data from one computer system to another through means of a network. Common methods of uploading include: uploading via web browsers, FTP clients], and terminals (SCP/SFTP). Uploading can be used in the context of (potentially many) clients that send files to a central server. While uploading can also be defined in the context of sending files between distributed clients, such as with a peer-to-peer (P2P) file-sharing protocol like BitTorrent, the term file sharing is more often used in this case. Moving files within a computer system, as opposed to over a network, is called file copying. Uploading directly contrasts with downloading, where data is received over a network. In the case of users uploading files over the internet, uploading is often slower than downloading as many internet service providers (ISPs) offer asymmetric connections, which offer more network bandwidth for downloading than uploading. To transfer something (such as data or files), from a computer or other digital device to the memory of another device (such as a larger or remote computer) especially via the internet. Remote file sharing first came into fruition in January 1978, when Ward Christensen and Randy Suess, who were members of the Chicago Area Computer Hobbyists' Exchange (CACHE), created the Computerized Bulletin Board System (CBBS). This used an early file transfer protocol (MODEM, later XMODEM) to send binary files via a hardware modem, accessible by another modem via a telephone number. In the following years, new protocols such as Kermit were released, until the File Transfer Protocol (FTP) was standardized 1985 (RFC 959). FTP is based on TCP/IP and gave rise to many FTP clients, which, in turn, gave users all around the world access to the same standard network protocol to transfer data between devices. The transfer of data saw a significant increase in popularity after the release of the World Wide Web in 1991, which, for the first time, allowed users who were not computer hobbyists to easily share files, directly from their web browser over HTTP. Transfers became more reliable with the launch of HTTP/1.1 in 1997 (RFC 2068), which gave users the option to resume downloads that were interrupted, for instance due to unreliable connections. Before web browsers widely rolled out support, software programs like GetRight could be used to resume downloads. Resuming uploads is not currently supported by HTTP, but can be added with the Tus open protocol for resumable file uploads, which layers resumability of uploads on top of existing HTTP connections. Transmitting a local file to a remote system following the client–server model, e.g., a web browser transferring a video to a website, is called client-to-server uploading. Transferring data from one remote system to another remote system under the control of a local system is called remote uploading or site-to-site transferring. This is used when a local computer has a slow connection to the remote systems, but these systems have a fast connection between them. Without remote uploading functionality, the data would have to first be downloaded to the local system and then uploaded to the remote server, both times over a slower connection. Remote uploading is used by some online file hosting services. Another example can be found in FTP clients, which often support the File eXchange Protocol (FXP) in order to instruct two FTP servers with high-speed connections to exchange files. A web-based example is the Uppy file uploader that can transfer files from a user's cloud storage such as Dropbox, directly to a website without first going to the user's device. Peer-to-peer (P2P) is a decentralized communications model in which each party has the same capabilities, and either party can initiate a communication session. Unlike the client–server model, in which the client makes a service request and the server fulfils the request (by sending or accepting a file transfer), the P2P network model allows each node to function as both client and server. BitTorrent is an example of this, as is the InterPlanetary File System (IPFS). Peer-to-peer allows users to both receive (download) and host (upload) content. Files are transferred directly between the users' computers. The same file transfer constitutes an upload for one party, and a download for the other party. The rising popularity of file sharing during the 1990s culminated in the emergence of Napster, a music-sharing platform specialized in MP3 files that used peer-to-peer (P2P) file-sharing technology to allow users exchange files freely. The P2P nature meant there was no central gatekeeper for the content, which eventually led to the widespread availability of copyrighted material through Napster. The Recording Industry Association of America (RIAA) took notice of Napster's ability to distribute copyrighted music among its user base, and, on December 6, 1999, filed a motion for a preliminary injunction in order to stop the exchange of copyrighted songs on the service. After a failed appeal by Napster, the injunction was granted on March 5, 2001. On September 24, 2001, Napster, which had already shut down its entire network two months earlier, agreed to pay a $26 million dollar settlement. After Napster had ceased operations, many other P2P file-sharing services also shut down, such as Limewire, Kazaa and Popcorn Time. Besides software programs, there were many BitTorrent websites that allowed files to be indexed and searched. These files could then be downloaded via a BitTorrent client. While the BitTorrent protocol itself is legal and agnostic of the type of content shared, many of the services that did not enforce a strict policy to take down copyrighted material would eventually also run into legal difficulties.
{{calendar}} {{day}} == Events == === Up to 1900 === * [[686]] &ndash; [[Pope Conon|Conon]] becomes [[Pope]]. * [[1096]] &ndash; [[People's Crusade]]: The [[Turkey|Turkish]] [[army]] heavily defeats the [[People's Army of the West]]. * [[1097]] &ndash; [[First Crusade]]: The [[Siege of Antioch]] begins. * [[1209]] &ndash; [[Otto IV, Holy Roman Emperor]] is crowned by [[Pope Innocent III]]. * [[1392]] &ndash; [[Emperor Kameyama]] of [[Japan]] abdicates the throne in favour of his arch-rival [[Emperor Go-Komatsu]]. * [[1512]] &ndash; [[Martin Luther]] joins the theological faculty at the [[University of Wittenberg]]. * [[1520]] &ndash; [[Ferdinand Magellan]] enters what is now known as the [[Magellan Strait]]. * [[1600]] &ndash; [[Tokugawa Ieyasu]] defeats the leaders of rival [[Japan]]ese clans in the [[Battle of Sekigahara]], which marks the beginning of the [[Tokugawa shogunate]], who in effect rule Japan until the mid-Nineteenth century. * [[1638]] &ndash; In [[England]], the [[church (building)|church]] at Widecombe-on-the-Moor is struck by [[lightning]] during an afternoon service, leading to new scientific research on lightning strikes. * [[1797]] &ndash; In [[Boston Harbor]], the 44-gun [[United States Navy]] frigate [[USS Constitution|USS ''Constitution'']] is launched. * [[1805]] &ndash; [[Napoleonic Wars]]: [[Battle of Trafalgar]] – a [[United Kingdom|British]] fleet led by Admiral [[Lord Nelson]] defeats a combined [[France|French]] and [[Spain|Spanish]] fleet off the coast of Spain under [[Pierre-Charles Villeneuve|Admiral Villeneuve]]. It signalled the virtual end of French maritime power and left Britain navally unchallenged until the twentieth century. * [[1805]] &ndash; [[Napoleonic Wars]]: Austrian [[Karl Freiherr Mack von Leiberich|General Mack]] surrendurs his army to the [[Grand Armee]] of [[Napoleon]] at [[Battle of Ulm|Ulm]], reaping [[Napoleon]] over 30,000 prisoners and inflicting 10,000 casualties on the losers. Ulm was considered to be one of Napoleon's finest hours. * [[1816]] &ndash; The [[Penang]] Free School is founded in [[George Town, Penang|George Town]], [[Malaysia]]. It is the oldest [[English language]] school in [[Southeast Asia]]. * [[1824]] &ndash; [[Joseph Aspdin]] [[patent]]s [[Portland cement]]. * [[1854]] &ndash; [[Florence Nightingale]] and a staff of 38 [[nurse]]s were sent to the [[Crimean War]]. * [[1861]] &ndash; [[American Civil War]]: [[Battle of Ball's Bluff]] – [[United States|Union]] forces under Colonel [[Edward Baker]] are defeated by [[Confederate States of America|Confederate]] troops in the second major battle of the war. Baker, a close friend of [[Abraham Lincoln]], is killed in the fighting. * [[1867]] &ndash; [[Manifest Destiny]]: [[Medicine Lodge Treaty]] – Near [[Medicine Lodge Creek, Kansas]] a landmark treaty is signed by southern [[Great Plains]] Indian leaders. The treaty requires [[Native American]] Plains tribes to relocate a reservation in western [[Oklahoma]]. * [[1879]] &ndash; Using a filament of [[carbon]]ized thread, [[Thomas Edison]] tests the first practical electric [[light bulb]] (it lasted 13 1/2&nbsp;hours before burning out). * [[1885]] &ndash; In an [[assassination]] attempt, Danish Prime Minister [[Jacob Estrup]] is shot. * [[1888]] &ndash; The [[Switzerland|Swiss]] [[Social Democracy|Social Democratic Party]] is founded. * [[1895]] &ndash; The [[Republic of Taiwan]] collapses as [[Japan]]ese forces invade. === 1901 &ndash; 2000 === * [[1902]] &ndash; In the [[United States]], a five-month [[Strike action|strike]] by [[United Mine Workers]] ends. * [[1907]] &ndash; A magnitude 8.1 [[earthquake]] hits Central [[Asia]], killing 12,000. * [[1910]] &ndash; HMS ''Niobe'' arrives at Halifax Harbour, [[Nova Scotia]] to become the first ship of the [[Royal Canadian Navy]]. * [[1921]] &ndash; US President [[Warren G. Harding]] delivers the first speech by a sitting president against [[lynching]] in the deep south. * [[1930]] &ndash; A mining disaster at Alsdorf, near [[Aachen]], [[Germany]], kills 271 people. * [[1941]] &ndash; [[World War II]]: [[Germany|Germans]] rampage in [[Kingdom of Yugoslavia|Yugoslavia]], killing thousands of civilians. * [[1943]] &ndash; The [[Provisional Government of Free India]] is declared by [[Subhas Chandra Bose]]. * [[1944]] &ndash; The first [[kamikaze]] attack: [[HMAS Australia]] was hit by a [[Japan]]ese plane carrying a 200&nbsp;kg (441&nbsp;pound) [[bomb]] off [[Leyte (island)|Leyte Island]], as the [[Battle of Leyte Gulf]] began. * [[1944]] &ndash; [[World War II]]: [[Aachen]] becomes the first major German city to fall to the [[Allies of World War II|Allies]]. * [[1945]] &ndash; [[Women's suffrage]]: Women are allowed to vote in [[France]] for the first time. * [[1945]] &ndash; [[Argentina|Argentine]] [[military officer]] and [[politician]] [[Juan Perón]] married [[Actor|actress]] [[Eva Perón|Evita]]. * [[1947]] &ndash; 21 die as a fire destroys an [[Psychiatric hospital|asylum]] in [[Hoff, Germany|Hoff]], [[Germany]]. * [[1948]] &ndash; A Lockheed Constellation [[airplane]] crashes at [[Prestwick]], [[Scotland]], killing 39 people. * [[1957]] &ndash; The movie [[Jailhouse Rock (movie)|Jailhouse Rock]], starring [[Elvis Presley]], opens. * [[1959]] &ndash; In [[New York City]], the [[Solomon R. Guggenheim Museum]] opens to the public. It was designed by [[Frank Lloyd Wright]]. * [[1959]] &ndash; US President [[Dwight D. Eisenhower]] signs an [[executive order]] transferring [[Wernher von Braun]] and other [[Germany|German]] scientists from the [[United States Army]] to [[NASA]]. * [[1962]] &ndash; [[Norway|Norwegian]] postal ship ''Sanct Svithun'' sinks, killing 41. * [[1965]] &ndash; [[Comet]] Ikeya-Seki approaches [[Perihelion]], passing 200,000 [[kilometer]]s (279,617 [[mile]]s) within [[Earth]]. * [[1966]] &ndash; [[Aberfan disaster]]: A coal tip falls on the village of [[Aberfan]] in [[Wales]], killing 144 people, mostly schoolchildren * [[1967]] &ndash; [[Vietnam War]]: More than 100,000 [[war protest]]ers gather in [[Washington, DC]]. A peaceful rally at the [[Lincoln Memorial]] is followed by a march to [[The Pentagon]] and clashes with soldiers and [[United States Marshal]]s protecting the facility (event lasts until [[October 23]]; 683 people will be arrested). Similar demonstrations occurred simultaneously in [[Japan]] and [[Western Europe]]. * [[1969]] &ndash; [[Willy Brandt]] is elected [[Chancellor of Germany|Chancellor]] of [[West Germany]]. * [[1971]] &ndash; 22 people are killed in a [[gas]] explosion at a shopping centre in Clarkston, [[East Renfrewshire]], near [[Glasgow]], [[Scotland]]. * [[1973]] &ndash; [[John Paul Getty III]]'s ear is cut off by his kidnappers and sent to a newspaper in [[Rome]]; it does not arrive until [[November 8]]. * [[1977]] &ndash; The [[European Patent Institute]] is founded. * [[1980]] &ndash; [[1980 World Series]]: In 6 games, the [[Philadelphia Phillies]] win their first [[World Series]]. * [[1981]] &ndash; [[Andreas Papandreou]] becomes [[Prime Minister]] of [[Greece]]. * [[1986]] &ndash; In [[Lebanon]], pro–[[Iran]]ian kidnappers claim to have abducted [[United States|American]] writer [[Edward Tracy]] (he will be released in August [[1991]]). *[[1986]] &ndash; [[African Charter on Human and Peoples' Rights]] enters into force. * [[1987]] &ndash; Former [[Miss America]] [[Bess Myerson]] is arrested on charges of [[bribery]], [[conspiracy]], and mail [[fraud]], all involving an alimony-fixing scandal. She is later found not guilty. * [[1994]] &ndash; [[North Korea nuclear weapons program]]: [[North Korea]] and the [[United States]] sign an agreement that requires North Korea to stop its [[nuclear weapon]]s program and agree to inspections. * [[1994]] &ndash; In [[Seoul]], [[South Korea]], 32 people are killed when the Seongsu [[Bridge]] collapses. * [[1997]] &ndash; Hotel owners from the [[Detroit]] area meet to discuss [[Jack Kevorkian]]'s practice of leaving corpses in hotel rooms. * [[1997]] &ndash; The government of [[Singapore]] announces in a widely publicized "[[toilet]] alert" that the drive for toilet cleanliness is a great success; five toilets were selected by citizens as toilet role models. === From 2001 === * [[2004]] &ndash; The [[Boston Red Sox]] win the American League pennant, defeating the [[New York Yankees]] 10–3 in Game 7 of the American League Championship Series, capping off a remarkable comeback from three games to none down to win. * [[2005]] &ndash; [[Hurricane Wilma]] strikes the [[Yucatán Peninsula|Yucatán]], [[Mexico]]. * [[2007]] &ndash; [[Kimi Raikkonen]] wins the [[Formula One]] World Championship. * [[2007]] &ndash; [[Donald Tusk]] is elected [[Prime Minister]] of [[Poland]]. * [[2015]] &ndash; This is the exact date that the main characters travel to in [[Back to the Future Part II]]. * [[2016]] &ndash; A [[train]] crash in Eséka, [[Cameroon]], kills at least 55 people. * [[2017]] &ndash; The [[government]] of [[Spain]] begins moves to take direct control of the region of [[Catalonia]] after the region's disputed referendum on its independence on [[October 1]]. * [[2019]] &ndash; The 2019 [[Canada|Canadian]] federal [[election]] is held, with [[Justin Trudeau]]'s [[Liberal Party of Canada]] retaining power, but with a reduced number of seats and in second place in the popular vote behind [[Andrew Scheer]]'s [[Conservative Party of Canada]]. == Births == === Up to 1900 === * [[1328]] &ndash; [[Hongwu Emperor]] of [[China]] (d. [[1398]]) * [[1449]] &ndash; [[George, Duke of Clarence]] (d. [[1478]]) * [[1527]] &ndash; [[Louis I, Cardinal of Guise]] (d. [[1578]]) * [[1581]] &ndash; [[Domenico Zampieri]], Italian painter (d. [[1641]]) * [[1650]] &ndash; [[Jean Bart]], French [[Navy|naval]] commander (d. [[1702]]) * [[1660]] &ndash; [[Georg Ernst Stahl]], German scientist (d. [[1734]]) * [[1671]] &ndash; King [[Frederick IV of Demmark]] (d. [[1730]]) * [[1672]] &ndash; [[Ludovico Antonio Muratori]], Italian writer (d. [[1750]]) * [[1675]] &ndash; [[Emperor Higashiyama of Japan]] (d. [[1710]]) * [[1687]] &ndash; [[Nicolaus I Bernoulli]], Swiss [[mathematician]] (d. [[1759]]) * [[1728]] &ndash; [[José Moñino, 1st Count of Floridablanca]], Spanish politician (d. [[1808]]) * [[1762]] &ndash; [[Herman Willem Daendels]], Dutch [[statesman]] (d. [[1818]]) * [[1772]] &ndash; [[Samuel Taylor Coleridge]], British [[poet]] (d. [[1834]]) * [[1775]] &ndash; [[Giuseppe Baini]], Italian composer (d. [[1844]]) * [[1790]] &ndash; [[Alphonse de Lamartine]], French writer (d. [[1869]]) * [[1811]] &ndash; [[Filippo Colini]], Italian [[opera]] singer (d. [[1863]]) * [[1823]] &ndash; [[Emilio Arrieta]], Spanish [[composer]] (d. [[1894]]) * [[1833]] &ndash; [[Alfred Nobel]], Swedish inventor (d. [[1896]]) * [[1837]] &ndash; [[James A. Beaver]], 20th [[Governor]] of [[Pennsylvania]] (d. [[1914]]) * [[1839]] &ndash; [[Georg von Siemens]], German banker and politician (d. [[1901]]) * [[1845]] &ndash; [[Will Carleton]], American poet (d. [[1912]]) * [[1846]] &ndash; [[Edmondo De Amicis]], Italian novelist, journalist, poet and short-story writer (d. [[1908]]) * [[1847]] &ndash; [[Giuseppe Giacosa]], Italian writer (d. [[1906]]) * [[1874]] &ndash; [[Henri Guisan]], Swiss general (d. [[1960]]) * [[1884]] &ndash; [[Claire Waldoff]], German singer and entertainer (d. [[1957]]) * [[1886]] &ndash; [[Eugene Burton Ely]], American [[Aviation|aviator]] (d. [[1911]]) * [[1886]] &ndash; [[Karl Polanyi]], [[Austria-Hungary|Austro-Hungarian]] [[economist]] (d. [[1964]]) * [[1887]] &ndash; [[James L. McConaughy]], American politician, [[Governor]] of [[Connecticut]] (d. [[1948]]) * [[1887]] &ndash; [[Krishna Singh]], Indian politician, 1st [[Chief Minister]] of [[Bihar]] (d. [[1961]]) * [[1892]] &ndash; [[Otto Nerz]], German [[footballer]] and coach (d. [[1949]]) * [[1894]] &ndash; [[Edogawa Rampo]], Japanese writer and critic (d. [[1965]]) * [[1895]] &ndash; [[Edna Purviance]], American actress (d. [[1958]]) * [[1895]] &ndash; [[Paavo Johansson]], Finnish [[javelin]] thrower (d. [[1983]]) * [[1896]] &ndash; [[Evgeny Schwarz]], Russian movie writer (d. [[1958]]) === 1901 &ndash; 1950 === * [[1901]] &ndash; [[Margarete Buber-Neumann]], German politician (d. [[1989]]) * [[1907]] &ndash; [[Nikos Engonopoulos]], Greek painter and poet (d. [[1985]]) * [[1911]] &ndash; [[Mary Blair]], American artist and illustrator (d. [[1978]]) * [[1912]] &ndash; [[Georg Solti]], Hungarian [[Conductor (music)|conductor]] (d. [[1997]]) * [[1914]] &ndash; [[Kazimierz Swiatek]], Roman Catholic [[cardinal (Catholicism)|cardinal]] of [[Belarus]] (d. [[2011]]) * [[1914]] &ndash; [[Martin Gardner]], American writer (d. [[2010]]) * [[1917]] &ndash; [[Dizzy Gillespie]], American musician (d. [[1993]]) * [[1918]] &ndash; [[Hulett C. Smith]], [[Governor]] of [[West Virginia]] (d. [[2012]]) * [[1921]] &ndash; [[Malcolm Arnold]], British [[composer]] (d. [[2006]]) * [[1921]] &ndash; [[Ingrid van Houten-Groeneveld]], Dutch [[astronomer]] (d. [[2015]]) * [[1921]] &ndash; [[Sena Jurinac]], [[Bosnia and Herzegovina|Bosnia]]n [[opera]]tic [[soprano]] (d. [[2011]]) * [[1922]] &ndash; [[Liliane Bettencourt]], French [[L'Oreal]] heiress (d. [[2017]]) * [[1924]] &ndash; [[Joyce Randolph]], American actress * [[1924]] &ndash; [[Julie Wilson]], American actress (d. [[2015]]) * [[1925]] &ndash; [[Celia Cruz]], Cuban [[Salsa]] singer (d. [[2003]]) * [[1925]] &ndash; [[Virginia Zeani]], Romanian [[soprano]] * [[1926]] &ndash; [[Leonard Rossiter]], English comedian and actor (d. [[1984]]) * [[1926]] &ndash; [[William Love Waller]], [[Governor]] of [[Mississippi]] (d. [[2011]]) * [[1927]] &ndash; [[Fritz Wintersteller]], Austrian [[mountain]]eer (d. [[2018]]) * [[1927]] &ndash; [[Howard Zieff]], American director (d. [[2009]]) * [[1928]] &ndash; [[Vern Mikkelsen]], American [[basketball]] player (d. [[2013]]) * [[1928]] &ndash; [[Whitey Ford]], American [[baseball]] player * [[1929]] &ndash; [[Pierre Bellemare]], French [[writer]], [[radio]] and [[television]] host (d. [[2018]]) * [[1930]] &ndash; [[Pascale Roberts]], French actress (d. [[2019]]) * [[1930]] &ndash; [[Ivan Silayev]], [[Soviet]] [[Prime Minister]] * [[1931]] &ndash; [[Shammi Kapoor]], Indian actor (d. [[2011]]) * [[1932]] &ndash; [[Alex Ifeanyichukwu Ekwueme]], [[Vice president]] of [[Nigeria]] (d. [[2017]]) * [[1932]] &ndash; [[Pál Csernai]], Hungarian [[footballer]] (d. [[2013]]) * [[1933]] &ndash; [[Georgia Brown]], American actress (d. [[1992]]) * [[1933]] &ndash; [[Joan Hinde]], English [[trumpet]]er and entertainer (d. [[2015]]) * [[1937]] &ndash; [[Hans-Ulrich Schmincke]], German [[Volcanology|volcanologist]] * [[1937]] &ndash; [[Said Afandi al-Chirkawi]], Russian spiritual leader (d. [[2012]]) * [[1937]] &ndash; [[Buranovskiye Babushki|Valentina Pyatchenko]], Russian singer ([[Buranovskiye Babushki]]) * [[1940]] &ndash; [[Geoffrey Boycott]], English [[cricket]]er * [[1940]] &ndash; [[Manfred Mann]], South African-English musician * [[1940]] &ndash; [[Marita Petersen]], [[Prime Minister]] of the [[Faroe Islands]] (d. [[2001]]) * [[1941]] &ndash; [[Steve Cropper]], American musician and songwriter * [[1942]] &ndash; [[Christopher A. Sims]], American [[economist]] * [[1942]] &ndash; [[Paul Churchland]], Canadian [[philosopher]] * [[1942]] &ndash; [[Judith Sheindlin]], American judge ([[Judge Judy]]) * [[1942]] &ndash; [[Lou Lamoriello]], American [[ice hockey]] player, coach and manager * [[1943]] &ndash; [[Tariq Ali]], Pakistani writer and historian * [[1944]] &ndash; [[Jean-Pierre Sauvage]], French [[chemist]], [[2016]] joint [[Nobel Prize in Chemistry]] laureate * [[1944]] &ndash; [[Mandy Rice-Davies]], British media personality (d. [[2014]]) * [[1945]] &ndash; [[Nikita Mikhalkov]], Soviet-Russian movie director * [[1946]] &ndash; [[Jane Heal]], British [[philosopher]] * [[1946]] &ndash; [[Lux Interior]], American singer (d. [[2009]]) * [[1948]] &ndash; [[Tom Everett]], American actor * [[1949]] &ndash; [[Benjamin Netanyahu]], [[Prime Minister]] of [[Israel]] * [[1949]] &ndash; [[LeTanya Richardson]], American actress * [[1949]] &ndash; [[Mike Keenan]], Canadian [[ice hockey]] coach * [[1950]] &ndash; [[Ronald McNair]], American [[astronaut]] (d. [[1986]]) === 1951 &ndash; 1975 === * [[1952]] &ndash; [[Trevor Chappell]], Australian [[cricket]]er * [[1952]] &ndash; [[Brent Mydland]], American [[Keyboard instrument|keyboardist]] (d. [[1990]]) * [[1952]] &ndash; [[Patti Davis]], American actress and novelist, daughter of [[Ronald Reagan]] * [[1953]] &ndash; [[Peter Mandelson]], English politician * [[1953]] &ndash; [[Eric Faulkner]], Scottish [[singer-songwriter]] and [[guitar]]ist * [[1954]] &ndash; [[Brian Tobin]], Canadian politician * [[1955]] &ndash; [[Catherine Hardwicke]], American [[movie director]], [[producer]], [[design]]er and [[screenwriter]] * [[1956]] &ndash; [[Carrie Fisher]], American actress and writer (d. [[2016]]) * [[1957]] &ndash; [[Steve Lukather]], American [[guitarist]] and [[singer]] ([[Toto (band)|Toto]]) * [[1957]] &ndash; [[Wolfgang Ketterle]], German [[physicist]], [[Nobel Prize]] winner * [[1958]] &ndash; [[Andre Geim]], Russian-born [[physicist]], [[Nobel Prize]] winner * [[1959]] &ndash; [[Ken Watanabe]], Japanese actor * [[1960]] &ndash; [[Scott Stearney]], American Vice Admiral (d. [[2018]]) * [[1962]] &ndash; [[David Campese]], Australian [[rugby union|rugby]] player * [[1964]] &ndash; [[Jon Carin]], American musician * [[1965]] &ndash; [[Ion Andoni Goikoetxea]], Basque-Spanish [[footballer]] * [[1967]] &ndash; [[Paul Ince]], English footballer * [[1969]] &ndash; [[Salman bin Hamad bin Isa Al Khalifa]], crown prince of [[Bahrain]] * [[1969]] &ndash; [[Chris Law (politician)|Chris Law]], Scottish politician * [[1971]] &ndash; [[Jade Jagger]], socialite and jewellery designer * [[1971]] &ndash; [[Damien Martyn]], Australian cricketer * [[1971]] &ndash; [[Paul Telfer]], Scottish [[footballer]] * [[1971]] &ndash; [[Thomas Ulsrud]], Norwegian [[Curling|curler]] * [[1972]] &ndash; [[Felicity Anderson]], Australian actress * [[1974]] &ndash; [[Costel Busuioc]], Romanian [[tenor]] === From 1976 === * [[1976]] &ndash; [[Andrew Scott (actor)|Andrew Scott]], Irish actor * [[1979]] &ndash; [[Fernanda Rodrigues]], Brazilian actress and television presenter * [[1979]] &ndash; [[Karl Harris]], British [[motorcycle]] racer (d. [[2014]]) * [[1980]] &ndash; [[Kim Kardashian]], American socialite, model and reality television personality * [[1981]] &ndash; [[Nemanja Vidic]], [[Serbia]]n [[footballer]] * [[1981]] &ndash; [[Martin Castrogiovanni]], Argentine-born Italian [[Rugby football|rugby]] player * [[1983]] &ndash; [[Hrvoje Custic]], Croatian footballer (d. [[2008]]) * [[1983]] &ndash; [[Amber Rose]], American model * [[1983]] &ndash; [[Charlotte Sullivan]], Canadian actress * [[1984]] &ndash; [[Kieran Richardson]], English footballer * [[1986]] &ndash; [[Natalee Holloway]], American teenager (missing since [[2005]]) * [[1986]] &ndash; [[Tamerlan Tsarnaev]], [[Chechnya|Chechen]]–American terrorist (d. [[2013]]) * [[1988]] &ndash; [[James White (basketball)|James White]], American [[basketball]] player * [[1989]] &ndash; [[Sam Vokes]], Welsh [[footballer]] * [[1990]] &ndash; [[Maxime Vachier-Lagrave]], French [[chess]] player * [[1990]] &ndash; [[Ricky Rubio]], Spanish [[basketball]] player * [[1992]] &ndash; [[Bernard Tomic]], Australian [[tennis]] player * [[1993]] &ndash; [[Kane Brown]], American singer * [[1995]] &ndash; [[Cameron Burgess]], Scottish-Australian [[footballer]] * [[1995]] &ndash; [[Shannon Magrane]], American [[singer]] == Deaths == === Up to 1900 === * [[310]] &ndash; [[Pope Eusebius]] * [[1125]] &ndash; [[Cosmas of Prague]], [[Bohemia]]n writer * [[1204]] &ndash; [[Robert de Beaufort, 4th Earl of Leicester]] * [[1221]] &ndash; [[Alix, Duchess of Brittany]] (b. [[1201]]) * [[1266]] &ndash; [[Birger Jarl]], Swedish statesman (b. [[1210]]) * [[1422]] &ndash; King [[Charles VI of France]] (b. [[1368]]) * [[1500]] &ndash; [[Emperor Go-Tsuchimikado]] of [[Japan]] (b. [[1442]]) * [[1558]] &ndash; [[Julius Caesar Scaliger]], [[humanist]] scholar (b. [[1484]]) * [[1600]] &ndash; [[Toda Katsushige]], Japanese [[warlord]] (b. [[1557]]) * [[1623]] &ndash; [[William Wade]], English [[statesman]] and [[diplomat]] (b. [[1546]]) * [[1687]] &ndash; Sir [[Edmund Waller]], English [[poet]] (b. [[1606]]) * [[1708]] &ndash; [[Guru Gobind Singh]], 10th [[Sikhism|Sikh]] [[Guru]] (b. [[1666]]) * [[1765]] &ndash; [[Giovani Paolo Panini]], Italian painter and architect (b. [[1691]]) * [[1775]] &ndash; [[Peyton Randolph]], American politician, President of the [[Continental Congress]] (b. [[1721]]) * [[1805]] &ndash; Killed in the [[Battle of Trafalgar]]: ** [[Horatio Nelson]], [[United Kingdom|British]] admiral (b. [[1758]]) ** [[George Duff]], [[Royal Navy]] captain (b. [[1764]]) ** [[John Cooke]], [[Royal Navy]] captain (b. [[1763]]) * [[1818]] &ndash; [[Michael Howe (bushranger)|Michael Howe]], [[bushranger]] in [[Tasmania]], [[Australia]] * [[1821]] &ndash; [[Dorothea Ackermann]], German actress (b. [[1752]]) * [[1872]] &ndash; [[Jacques Babinet]], French [[physicist]] (b. [[1794]]) * [[1873]] &ndash; [[Johann Sebastian Welhaven]], Norwegian poet (b. [[1807]]) * [[1896]] &ndash; [[James Henry Greathead]], [[United Kingdom|British]] [[engineer]] (b. [[1844]]) === 1901 &ndash; 2000 === * [[1903]] &ndash; [[Jinmaku Kyugoro]], Japanese [[sumo]] wrestler (b. [[1829]]) * [[1904]] &ndash; [[Isabelle Eberhardt]], Swiss explorer and writer (b. [[1877]]) * [[1907]] &ndash; [[Jules Chevalier]], French [[priest]] (b. [[1824]]) * [[1916]] &ndash; [[Count Karl von Stürgkh]], Austrian politician (b. [[1859]]) * [[1929]] &ndash; [[Vasil Radoslavov]], Bulgarian politician (b. [[1854]]) * [[1931]] &ndash; [[Arthur Schnitzler]], [[Austria]]n writer (b. [[1862]]) * [[1931]] &ndash; [[Barbecue Bob]], American [[blues]] musician (b. [[1902]]) * [[1940]] &ndash; [[William G. Conley]], [[Governor]] of [[West Virginia]] (b. [[1866]]) * [[1944]] &ndash; [[Alois Kayser]], [[Germany|German]] missionary, working in [[Nauru]] * [[1944]] &ndash; [[Hilma af Klint]], Swedish artist and mystic (b. [[1862]]) * [[1952]] &ndash; [[Hans Merensky]], South African geologist and philanthropist (b. [[1871]]) * [[1967]] &ndash; [[Ejnar Hertzsprung]], Danish [[astronomer]] (b. [[1873]]) * [[1969]] &ndash; [[Waclaw Sierpinski]], [[Poland|Polish]] mathematician (b. [[1882]]) * [[1969]] &ndash; [[Jack Kerouac]], [[United States|American]] beat novelist (b. [[1922]]) * [[1975]] &ndash; [[Charles Reidpath]], American athlete (b. [[1887]]) * [[1978]] &ndash; [[Anastas Mikoyan]], Soviet politician (b. [[1895]]) * [[1980]] &ndash; [[Hans Asperger]], [[Austrian]] psychologist who discovered [[Asperger's Syndrome]] (b. [[1906]]). * [[1980]] &ndash; [[Vulko Cherenkov]], Bulgarian politician (b. [[1900]]) * [[1984]] &ndash; [[François Truffaut]], [[France|French]] movie director (b. [[1932]]) * [[1985]] &ndash; [[Dan White]], American politician (b. [[1946]]) * [[1986]] &ndash; [[Lionel Keith Murphy|Lionel Murphy]], Australian Labor Party [[politician]] and [[High Court of Australia|High Court]] judge (b. [[1922]]) * [[1990]] &ndash; [[Walther Sommerlath]], German businessman, father of [[Queen Silvia of Sweden]] (b. [[1901]]) * [[1993]] &ndash; [[Melchior Ndadaye]], [[President]] of [[Burundi]] (b. [[1953]]) * [[1995]] &ndash; [[Shannon Hoon]], lead singer of pop band [[Blind Melon]] (b. [[1967]]) * [[1995]] &ndash; [[Maxene Andrews]], American singer (b. [[1916]]) * [[1995]] &ndash; [[Jesús Blasco]], [[Spain|Spanish]] [[comic book]] writer (b. [[1919]]) * [[1996]] &ndash; [[Georgios Zoitakis]], Greek army general and regent (b. [[1910]]) * [[1998]] &ndash; [[Francis W. Sargent]], [[Governor of Massachusetts]] (b. [[1915]]) * [[1999]] &ndash; [[Lars Bo]], Danish artist and writer (b. [[1924]]) === From 2001 === * [[2003]] &ndash; [[Fred Berry]], American actor (b. [[1951]]) * [[2003]] &ndash; [[Luis A. Ferré]], former governor of [[Puerto Rico]] (b. [[1904]]) * [[2003]] &ndash; [[Louise Day Hicks]], US politician (b. [[1916]]) * [[2003]] &ndash; [[Elliott Smith]], American musician (b. [[1969]]) * [[2006]] &ndash; [[Sandy West]], American musician (b. [[1959]]) * [[2010]] &ndash; [[Loki Schmidt]], German [[environmentalism|environmentalist]] and wife of [[Helmut Schmidt]] (b. [[1919]]) * [[2011]] &ndash; [[Edmundo Ros]], [[Trinidad and Tobago|Trinidadian]] musician (b. [[1910]]) * [[2012]] &ndash; [[George McGovern]], American politician (b. [[1922]]) * [[2012]] &ndash; [[Yash Chopra]], Indian [[movie director]], [[screenwriter]] and [[producer]] (b. [[1932]]) * [[2014]] &ndash; [[Gough Whitlam]], 21st [[Prime Minister of Australia]] (b. [[1916]]) * [[2014]] &ndash; [[Benjamin C. Bradlee]], American [[journalist]] (b. [[1921]]) * [[2014]] &ndash; [[Mohammad-Reza Mahdavi Kani]], Iranian politician and cleric (b. [[1931]]) * [[2014]] &ndash; [[Seth Gaaikema]], Dutch comedian and writer (b. [[1939]]) * [[2014]] &ndash; [[Lilli Carati]], Italian actress (b. [[1956]]) * [[2015]] &ndash; [[Marty Ingels]], American actor (b. [[1936]]) * [[2016]] &ndash; [[Manfred Krug]], German actor (b. [[1937]]) * [[2017]] &ndash; [[Martin Eric Ain]], American-Swiss musician (b. [[1967]]) * [[2017]] &ndash; [[Emilio D'Amore]], Italian politician (b. [[1915]]) * [[2017]] &ndash; [[Rosemary Leach]], English actress (b. [[1935]]) * [[2017]] &ndash; [[Lech Ordon]], Polish actor (b. [[1928]]) * [[2017]] &ndash; [[Denise P. Barlow]], British geneticist (b. [[1950]]) * [[2017]] &ndash; [[Judith McGrath]], Australian actress (b. [[1947]]) * [[2017]] &ndash; [[Gilbert Stork]], American organic chemist (b. [[1921]]) * [[2018]] &ndash; [[Earl Bakken]], American inventor and museum founder (b. [[1924]]) * [[2018]] &ndash; [[Ilie Balaci]], Romanian footballer and manager (b. [[1956]]) * [[2018]] &ndash; [[Harry L. Ettlinger]], American engineer (b. [[1926]]) * [[2018]] &ndash; [[Robert Faurisson]], British-French journalist, academic and [[Holocaust]] denier (b. [[1929]]) * [[2018]] &ndash; [[Joachim Ronneberg]], Norwegian military officer and broadcaster (b. [[1919]]) * [[2018]] &ndash; [[Jun-ichi Nishizawa]], Japanese electrical engineer (b. [[1926]]) * [[2018]] &ndash; [[Charles Wang]], Chinese-born American software engineer, philanthropist and sports team owner (b. [[1944]]) * [[2019]] &ndash; [[Gilberto Aceves Navarro]], Mexican painter and sculptor (b. [[1931]]) * [[2019]] &ndash; [[Bengt Feldreich]], Swedish television journalist (b. [[1925]]) * [[2019]] &ndash; [[Taras Kutovy]], Ukrainian politician (b. [[1976]]) * [[2019]] &ndash; [[Lho Shin-yong]], [[Prime Minister of South Korea]] (b. [[1930]]) * [[2019]] &ndash; [[Aila Meriluoto]], Finnish poet (b. [[1924]]) == Observances == * International Day of the Nacho ([[Mexico]] and the [[United States]]) * National [[Nurse]]s' Day ([[Thailand]]) * Overseas Chinese Day ([[Republic of China]]) * [[Apple]] Day ([[UK]]) [[Category:Days of the year]]
October 21 is the 294th day of the year (295th in leap years) in the Gregorian calendar; 71 days remain until the end of the year.
[[File:Stonehenge.jpg|thumb|300px|Stonehenge]] [[File:Stonehenge Summer Solstice eve 02.jpg|thumb|right|250px|Summer [[Solstice]]: visitors are not usually allowed this close!]] [[File:Stonehenge plan.jpg|thumb|right|250px|Plan of Stonehenge in 2004. Trilithon lintels omitted for clarity. Holes that no longer, or never, contained stones are shown as open circles. Stones visible today are shown coloured]]{{Other uses}} '''Stonehenge''' is a [[Prehistory|prehistoric]] [[World Heritage Site]] of [[megalith]]s. It is eight miles (13 kilometres) north of [[Salisbury]] in [[Wiltshire]], [[England]].<ref>UNESCO, [http://whc.unesco.org/en/list/373 "Stonehenge, Avebury and Associated Sites"]; retrieved 2012-4-19.</ref> the site was built between 3100 BC and 1550 BC. It was used until the [[Bronze Age]]. The monument is made of a [[henge]],<ref>Remains of a former ring bank and ditch</ref> with [[standing stone]]s in [[circle]]s. It is likely the most important prehistoric [[monument]] in Britain. The site has attracted visitors from very early times. == Building Stonehenge == Stonehenge was built in three stages. Most of the construction took place between 2640 and 2480 BC.<ref>{{cite news|author=Marc Kaufman|date=January 31, 2007|title=An ancient settlement is unearthed near Stonehenge|newspaper=The Washington Post|url=https://www.washingtonpost.com/wp-dyn/content/article/2007/01/30/AR2007013000661.html|accessdate=2008-05-23}}</ref> The first stage started around 3100 BC. During this stage, people dug a circular ditch and a ring of 56 pits, known as Aubrey Holes. The second stage started around 2100 BC. During this stage, the Stonehenge builders brought huge [https://simple.m.wiktionary.org/wiki/pillar pillars] of rocks from Southwestern [[Wales]] and erected them into [[concentric]] circles around the centre of the site. This double circle was never completed, and it was dismantled during the third period of construction. The final stage probably ended before 1500 BC. During this period, the monument was remodeled. Its builders erected a circle of 30 upright stones, weighing up to 50 [https://simple.m.wiktionary.org/wiki/ton tons] each, capped by ring of stone [[mwod:lintel|lintels]]. These enclosed a [[horseshoe]]-shaped formation of five pairs of upright stones, each pair capped with a stone lintel. Using [[Sequence analysis|DNA analysis]], scientists have discovered that the Stonehenge builders originally came from what is now [[Turkey]].<ref>{{Cite news|last=Rincon|first=Paul|date=2019-04-16|title=DNA reveals origin of Stonehenge builders|language=en-GB|url=https://www.bbc.com/news/science-environment-47938188|accessdate=2019-04-20}}</ref><ref>{{Cite web|url=https://www.itv.com/news/2019-04-16/stonehenge-builders-came-from-as-far-as-modern-day-turkey-dna-suggests/|title=Stonehenge builders came from as far as modern-day Turkey, DNA suggests|website=ITV News|date=16 April 2019|language=en|accessdate=2019-04-20}}</ref> == History == === Bluestones === The first [[stone circle]] was a set of 'bluestones'. These stones are made of [[dolerite]], an [[igneous rock]]. The Stonehenge builders dug holes that held up to 80 standing stones (shown blue on the plan). Only 43 of these can be traced today. There are several [[Theory|theories]] about how these bluestones arrived at Stonehenge. The ''long-distance human transport theory'' says the Stonehenge builders brought the bluestones from the [[Preseli Hills]] in modern-day [[Pembrokeshire]], [[Wales]] - 160 miles (260 km) away from Stonehenge.<ref name="Antiquity">{{cite journal |last1=Parker Pearson |first1=Michael |display-authors=etal |date=December 2015 |title=Craig Rhos-y-felin: a Welsh bluestone megalith quarry for Stonehenge |url=http://journals.cambridge.org/action/displayFulltext?type=6&fid=10057092&jid=AQY&volumeId=89&issueId=348&aid=10057091&bodyId=&membershipNumber=&societyETOCSession=&fulltextType=RA&fileId=S0003598X15001775 |journal=Antiquity |publisher= Antiquity Publications Ltd |volume=89 |issue=348 |pages=1331–1352 |doi=10.15184/aqy.2015.177 |s2cid=162776591 |accessdate=15 December 2015|issn = 0003-598X}}</ref> In 2011, a megalithic bluestone [[quarry]] was discovered at [[Craig Rhos-y-felin]], near [[Crymych]] in Pembrokeshire. This supported the long-distance human transport theory.<ref name="Antiquity" /> Another theory is that the [[Irish Sea Glacier]] brought the stones close to Stonehenge.<ref>John, Brian 2007: ''[http://www.brianjohn.f2s.com/enigma1.html The Stonehenge Bluestones—glacial transport back in favour] {{Webarchive|url=https://web.archive.org/web/20100901221137/http://www.brianjohn.f2s.com/enigma1.html |date=2010-09-01 }}''</ref> However, there is no evidence of glacial deposition within southern central England. For that reason, this theory has less support than the long-distance human transport theory. === Sarsen stones === Later, around 2400 BC, the Stonehenge builders brought thirty huge grey [[sarsen]] stones ([[sandstone]] blocks) to the site. They 'dressed' (worked on) the stones and gave them [[mortice and tenon]] joints. They erected these stones in a circle 33 metres (108 ft) in [[diameter]], with a ring of 30 lintel stones resting on top. They fitted the lintels together using another [[woodworking]] method: the [[tongue and groove joint]]. These stones serve as reminders of the lengths ancient civilizations went to, transporting them vast distances, in their quest to commune with the cosmos.<ref>{{Cite web|last=Nare|first=Alexa|title=Stonehenge Tours|url=https://www.visitstonehengelondon.com/|access-date=2023-09-21|website=|language=en}}</ref> They arranged the remaining bluestones in an inner circle. Each standing stone was about 4.1 metres (13 ft) high by 2.1 metres (6 ft 11 in) wide, and weighed about 25 tons. The stones may have come from a quarry about 25 miles (40 km) north of Stonehenge on the [[Marlborough|Marlborough Downs]]. It is also possible that they were collected from a "litter" of sarsens on the [[chalk downs]], which are closer. The modern Stonehenge consists entirely of original stones (some of which have been replaced in upright position). === Neighbouring sites === There are also several [[passage tomb]]s and many [[Tumulus|tumuli]] nearby. Stonehenge has a number of satellite structures which are part of the 'ritual landscape': *[[Bluehenge]]/Bluestonehenge: a new discovery, one mile to the southeast. *[[Durrington Walls]]: a [[Neolithic]] settlement two miles northeast of Stonehenge. *[[Normanton Down Barrows]]: a Neolithic and Bronze Age [[barrow]] [[cemetery]]. *[[Stonehenge Avenue]]: leads two miles from Stonehenge to [[Bluehenge]] on the River [[River Avon|Avon]]. *[[Stonehenge Cursus]]: the largest monument in the area, not easily visible on the ground. *[[Woodhenge]]: found in 1925 by an aerial survey. It had a henge and a wooden circle. == Function == No one knows who built Stonehenge or why they built it. During the [[summer|summer solstice]], the [[sunrise]] lines up with some of the stones in a particular way. This suggests that the arrangement of stones may work as a [[calendar]]. In [[Egypt]] and [[South America]], similar ancient [[building]]s can be found. They also show the time of the [[solstice]]. Some scientists believe that early people were able to foretell [[Eclipse|eclipses]] of the [[sun]] and the [[moon]] based on their positions in relation to the stone monument. The site may have served as an [[observatory]] where early [[Ritual|rituals]] or [[Religion|religious]] [[Ceremony|ceremonies]] took place on specific days on the year. Stonehenge itself is owned by the [[Monarchy|Crown]] and managed by [[English Heritage]]. The surrounding land is owned by the [[National Trust]].<ref>{{cite web | title =How did Stonehenge come into the care of English Heritage?| work =FAQs on Stonehenge| publisher =[[English Heritage]] | url =http://www.english-heritage.org.uk/server/show/conWebDoc.4189| accessdate =2007-12-17 }}</ref><ref>{{cite web| title =Ancient ceremonial landscape of great archaeological and wildlife interest | work =Stonehenge Landscape| publisher =National Trust | url =http://www.nationaltrust.org.uk/stonehengelandscape| accessdate =2007-12-17 }}</ref> The World Heritage Site includes [[Avebury]] and Stonhenge together, though they are quite distinct. == Related pages == * [[List of World Heritage Sites of the United Kingdom]] * [[Stonehenge, Avebury and Associated Sites]] == References == {{Reflist}} ==Further reading== {{Commons category|Stonehenge}} {{Wikivoyage}} *Malone, Caroline. 2005. ''Neolithic Britain and Ireland''. Tempus, Stroud, Gloucestershire. {{coord|51|10|43.8|N|1|49|34.3|W|type:landmark_region:GB-WIL|display=title}} {{WHSite}} [[Category:Religious buildings]] [[Category:Archaeological sites in Wiltshire]] [[Category:World Heritage Sites in the United Kingdom]] [[Category:Stone circles]] [[Category:Henges]] [[Category:Neolithic]] [[Category:Buildings and structures in Wiltshire]] [[Category:4th millennium BC buildings and structures]]
Stonehenge is a prehistoric monument on Salisbury Plain in Wiltshire, England, two miles (3 km) west of Amesbury. It consists of an outer ring of vertical sarsen standing stones, each around 13 feet (4.0 m) high, seven feet (2.1 m) wide, and weighing around 25 tons, topped by connecting horizontal lintel stones. Inside is a ring of smaller bluestones. Inside these are free-standing trilithons, two bulkier vertical sarsens joined by one lintel. The whole monument, now ruinous, is aligned towards the sunrise on the summer solstice and sunset on the winter solstice. The stones are set within earthworks in the middle of the densest complex of Neolithic and Bronze Age monuments in England, including several hundred tumuli (burial mounds). Archaeologists believe that Stonehenge was constructed in several phases from around 3100 BC to 1600 BC, with the circle of large sarsen stones placed between 2600 BC and 2400 BC. The surrounding circular earth bank and ditch, which constitute the earliest phase of the monument, have been dated to about 3100 BC. Radiocarbon dating suggests that the bluestones were given their current positions between 2400 and 2200 BC, although they may have been at the site as early as 3000 BC. One of the most famous landmarks in the United Kingdom, Stonehenge is regarded as a British cultural icon. It has been a legally protected scheduled monument since 1882, when legislation to protect historic monuments was first successfully introduced in Britain. The site and its surroundings were added to UNESCO's list of World Heritage Sites in 1986. Stonehenge is owned by the Crown and managed by English Heritage; the surrounding land is owned by the National Trust. Stonehenge could have been a burial ground from its earliest beginnings. Deposits containing human bone date from as early as 3000 BC, when the ditch and bank were first dug, and continued for at least another 500 years. The Oxford English Dictionary cites Ælfric's tenth-century glossary, in which henge-cliff is given the meaning 'precipice', or stone; thus, the stanenges or Stanheng "not far from Salisbury" recorded by eleventh-century writers are "stones supported in the air". In 1740, William Stukeley notes: "Pendulous rocks are now called henges in Yorkshire ... I doubt not, Stonehenge in Saxon signifies the hanging stones." Christopher Chippindale's Stonehenge Complete gives the derivation of the name Stonehenge as coming from the Old English words stān 'stone', and either hencg 'hinge' (because the stone lintels hinge on the upright stones) or hen(c)en 'to hang' or 'gallows' or 'instrument of torture' (though elsewhere in his book, Chippindale cites the 'suspended stones' etymology). The "henge" portion has given its name to a class of monuments known as henges. Archaeologists define henges as earthworks consisting of a circular banked enclosure with an internal ditch. As often happens in archaeological terminology, this is a holdover from antiquarian use. Despite being contemporary with true Neolithic henges and stone circles, Stonehenge is in many ways atypical – for example, at more than 24 feet (7.3 m) tall, its extant trilithons' lintels, held in place with mortise and tenon joints, make it unique. Mike Parker Pearson, leader of the Stonehenge Riverside Project based around Durrington Walls, noted that Stonehenge appears to have been associated with burial from the earliest period of its existence: Stonehenge was a place of burial from its beginning to its zenith in the mid third millennium B.C. The cremation burial dating to Stonehenge's sarsen stones phase is likely just one of many from this later period of the monument's use and demonstrates that it was still very much a domain of the dead. Stonehenge evolved in several construction phases spanning at least 1500 years. There is evidence of large-scale construction on and around the monument that perhaps extends the landscape's time frame to 6500 years. Dating and understanding the various phases of activity are complicated by disturbance of the natural chalk by periglacial effects and animal burrowing, poor quality early excavation records, and a lack of accurate, scientifically verified dates. The modern phasing most generally agreed to by archaeologists is detailed below. Features mentioned in the text are numbered and shown on the plan, right. Archaeologists have found four, or possibly five, large Mesolithic postholes (one may have been a natural tree throw), which date to around 8000 BC, beneath the nearby old tourist car-park in use until 2013. These held pine posts around two feet six inches (0.75 m) in diameter, which were erected and eventually rotted in situ. Three of the posts (and possibly four) were in an east–west alignment which may have had ritual significance. Another Mesolithic astronomical site in Britain is the Warren Field site in Aberdeenshire, which is considered the world's oldest lunisolar calendar, corrected yearly by observing the midwinter solstice. Similar but later sites have been found in Scandinavia. A settlement that may have been contemporaneous with the posts has been found at Blick Mead, a reliable year-round spring one mile (1.6 km) from Stonehenge. Salisbury Plain was then still wooded, but 4,000 years later, during the earlier Neolithic, people built a causewayed enclosure at Robin Hood's Ball, and long barrow tombs in the surrounding landscape. In approximately 3500 BC, a Stonehenge Cursus was built 2,300 feet (700 m) north of the site as the first farmers began to clear the trees and develop the area. Other previously overlooked stone or wooden structures and burial mounds may date as far back as 4000 BC. Charcoal from the 'Blick Mead' camp 1.5 miles (2.4 km) from Stonehenge (near the Vespasian's Camp site) has been dated to 4000 BC. The University of Buckingham's Humanities Research Institute believes that the community who built Stonehenge lived here over a period of several millennia, making it potentially "one of the pivotal places in the history of the Stonehenge landscape." The first monument consisted of a circular bank and ditch enclosure made of Late Cretaceous (Santonian Age) Seaford chalk, measuring about 360 feet (110 m) in diameter, with a large entrance to the north east and a smaller one to the south. It stood in open grassland on a slightly sloping spot. The builders placed the bones of deer and oxen in the bottom of the ditch, as well as some worked flint tools. The bones were considerably older than the antler picks used to dig the ditch, and the people who buried them had looked after them for some time prior to burial. The ditch was continuous but had been dug in sections, like the ditches of the earlier causewayed enclosures in the area. The chalk dug from the ditch was piled up to form the bank. This first stage is dated to around 3100 BC, after which the ditch began to silt up naturally. Within the outer edge of the enclosed area is a circle of 56 pits, each about 3.3 feet (1 m) in diameter, known as the Aubrey holes after John Aubrey, the seventeenth-century antiquarian who was thought to have first identified them. These pits and the bank and ditch together are known as the Palisade or Gate Ditch. The pits may have contained standing timbers creating a timber circle, although there is no excavated evidence of them. A recent excavation has suggested that the Aubrey Holes may have originally been used to erect a bluestone circle. If this were the case, it would advance the earliest known stone structure at the monument by some 500 years. In 2013 a team of archaeologists, led by Mike Parker Pearson, excavated more than 50,000 cremated bone fragments, from 63 individuals, buried at Stonehenge. These remains had originally been buried individually in the Aubrey holes, exhumed during a previous excavation conducted by William Hawley in 1920, been considered unimportant by him, and subsequently re-interred together in one hole, Aubrey Hole 7, in 1935. Physical and chemical analysis of the remains has shown that the cremated were almost equally men and women, and included some children. As there was evidence of the underlying chalk beneath the graves being crushed by substantial weight, the team concluded that the first bluestones brought from Wales were probably used as grave markers. Radiocarbon dating of the remains has put the date of the site 500 years earlier than previously estimated, to around 3000 BC. A 2018 study of the strontium content of the bones found that many of the individuals buried there around the time of construction had probably come from near the source of the bluestone in Wales and had not extensively lived in the area of Stonehenge before death. Between 2017 and 2021, studies by Professor Parker Pearson (UCL) and his team suggested that the bluestones used in Stonehenge had been moved there following dismantling of a stone circle of identical size to the first known Stonehenge circle (110m) at the Welsh site of Waun Mawn in the Preseli Hills. It had contained bluestones, one of which showed evidence of having been reused in Stonehenge. The stone was identified by its unusual pentagonal shape and by luminescence soil dating from the filled-in sockets which showed the circle had been erected around 3400-3200 BC, and dismantled around 300–400 years later, consistent with the dates attributed to the creation of Stonehenge. The cessation of human activity in that area at the same time suggested migration as a reason, but it is believed that other stones may have come from other sources. The second phase of construction occurred approximately between 2900 and 2600 BC. The number of postholes dating to the early third millennium BC suggests that some form of timber structure was built within the enclosure during this period. Further standing timbers were placed at the northeast entrance, and a parallel alignment of posts ran inwards from the southern entrance. The postholes are smaller than the Aubrey Holes, being only around 16 inches (0.4 m) in diameter, and are much less regularly spaced. The bank was purposely reduced in height and the ditch continued to silt up. At least twenty-five of the Aubrey Holes are known to have contained later, intrusive, cremation burials dating to the two centuries after the monument's inception. It seems that whatever the holes' initial function, it changed to become a funerary one during Phase two. Thirty further cremations were placed in the enclosure's ditch and at other points within the monument, mostly in the eastern half. Stonehenge is therefore interpreted as functioning as an enclosed cremation cemetery at this time, the earliest known cremation cemetery in the British Isles. Fragments of unburnt human bone have also been found in the ditch-fill. Dating evidence is provided by the late Neolithic grooved ware pottery that has been found in connection with the features from this phase. Archaeological excavation has indicated that around 2600 BC, the builders abandoned timber in favour of stone and dug two concentric arrays of holes (the Q and R Holes) in the centre of the site. These stone sockets are only partly known (hence on present evidence are sometimes described as forming 'crescents'); however, they could be the remains of a double ring. Again, there is little firm dating evidence for this phase. The holes held up to 80 standing stones (shown blue on the plan), only 43 of which can be traced today. It is generally accepted that the bluestones (some of which are made of dolerite, an igneous rock), were transported by the builders from the Preseli Hills, 150 miles (240 km) away in modern-day Pembrokeshire in Wales. Another theory is that they were brought much nearer to the site as glacial erratics by the Irish Sea Glacier although there is no evidence of glacial deposition within southern central England. A 2019 publication announced that evidence of Megalithic quarrying had been found at quarries in Wales identified as a source of Stonehenge's bluestone, indicating that the bluestone was quarried by human agency and not transported by glacial action. The long-distance human transport theory was bolstered in 2011 by the discovery of a megalithic bluestone quarry at Craig Rhos-y-felin, near Crymych in Pembrokeshire, which is the most likely place for some of the stones to have been obtained. Other standing stones may well have been small sarsens (sandstone), used later as lintels. The stones, which weighed about two tons, could have been moved by lifting and carrying them on rows of poles and rectangular frameworks of poles, as recorded in China, Japan and India. It is not known whether the stones were taken directly from their quarries to Salisbury Plain or were the result of the removal of a venerated stone circle from Preseli to Salisbury Plain to "merge two sacred centres into one, to unify two politically separate regions, or to legitimise the ancestral identity of migrants moving from one region to another". Evidence of a 110-metre (360 ft) stone circle at Waun Mawn near Preseli, which could have contained some or all of the stones in Stonehenge, has been found, including a hole from a rock that matches the unusual cross-section of a Stonehenge bluestone "like a key in a lock". Each monolith measures around 6.6 feet (2 m) in height, between 3.3 and 4.9 ft (1 and 1.5 m) wide and around 2.6 feet (0.8 m) thick. What was to become known as the Altar Stone is almost certainly derived from the Senni Beds, perhaps from 50 miles (80 kilometres) east of the Preseli Hills in the Brecon Beacons. The north-eastern entrance was widened at this time, with the result that it precisely matched the direction of the midsummer sunrise and midwinter sunset of the period. This phase of the monument was abandoned unfinished, however; the small standing stones were apparently removed and the Q and R holes purposefully backfilled. The Heel Stone, a Tertiary sandstone, may also have been erected outside the north-eastern entrance during this period. It cannot be accurately dated and may have been installed at any time during phase 3. At first, it was accompanied by a second stone, which is no longer visible. Two, or possibly three, large portal stones were set up just inside the north-eastern entrance, of which only one, the fallen Slaughter Stone, 16 feet (4.9 m) long, now remains. Other features, loosely dated to phase 3, include the four Station Stones, two of which stood atop mounds. The mounds are known as "barrows" although they do not contain burials. Stonehenge Avenue, a parallel pair of ditches and banks leading two miles (3 km) to the River Avon, was also added. During the next major phase of activity, 30 enormous Oligocene–Miocene sarsen stones (shown grey on the plan) were brought to the site. They came from a quarry around 16 miles (26 km) north of Stonehenge, in West Woods, Wiltshire. The stones were dressed and fashioned with mortise and tenon joints before 30 sarsens were erected as a 108-foot (33 m) diameter circle of standing stones, with a ring of 30 lintel stones resting on top. The lintels were fitted to one another using tongue and groove joints – a woodworking method, again. Each standing stone was around 13 feet (4.1 m) high, 6.9 feet (2.1 m) wide and weighed around 25 tons. Each had clearly been worked with the final visual effect in mind: The orthostats widen slightly towards the top in order that their perspective remains constant when viewed from the ground, while the lintel stones curve slightly to continue the circular appearance of the earlier monument. The inward-facing surfaces of the stones are smoother and more finely worked than the outer surfaces. The average thickness of the stones is 3.6 feet (1.1 m) and the average distance between them is 3.3 feet (1 m). A total of 75 stones would have been needed to complete the circle (60 stones) and the trilithon horseshoe (15 stones). It was thought the ring might have been left incomplete, but an exceptionally dry summer in 2013 revealed patches of parched grass which may correspond to the location of missing sarsens. The lintel stones are each around 10 feet (3.2 m) long, 3.3 feet (1 m) wide and 2.6 feet (0.8 m) thick. The tops of the lintels are 16 feet (4.9 m) above the ground. Within this circle stood five trilithons of dressed sarsen stone arranged in a horseshoe shape 45 feet (13.7 m) across, with its open end facing northeast. These huge stones, ten uprights and five lintels, weigh up to 50 tons each. They were linked using complex jointing. They are arranged symmetrically. The smallest pair of trilithons were around 20 feet (6 m) tall, the next pair a little higher, and the largest, single trilithon in the south-west corner would have been 24 feet (7.3 m) tall. Only one upright from the Great Trilithon still stands, of which 22 feet (6.7 m) is visible and a further 7.9 feet (2.4 m) is below ground. The images of a 'dagger' and 14 'axeheads' have been carved on one of the sarsens, known as stone 53; further carvings of axeheads have been seen on the outer faces of stones 3, 4, and 5. The carvings are difficult to date but are morphologically similar to late Bronze Age weapons. Early 21st century laser scanning of the carvings supports this interpretation. The pair of trilithons in the north east are smallest, measuring around 20 feet (6 m) in height; the largest, which is in the south-west of the horseshoe, is almost 25 feet (7.5 m) tall. This ambitious phase has been radiocarbon dated to between 2600 and 2400 BC, slightly earlier than the Stonehenge Archer, discovered in the outer ditch of the monument in 1978, and the two sets of burials, known as the Amesbury Archer and the Boscombe Bowmen, discovered three miles (5 km) to the west. Analysis of animal teeth found two miles (3 km) away at Durrington Walls, thought by Parker Pearson to be the 'builders camp', suggests that, during some period between 2600 and 2400 BC, as many as 4,000 people gathered at the site for the mid-winter and mid-summer festivals; the evidence showed that the animals had been slaughtered around nine months or 15 months after their spring birth. Strontium isotope analysis of the animal teeth showed that some had been brought from as far afield as the Scottish Highlands for the celebrations. At about the same time, a large timber circle and a second avenue were constructed at Durrington Walls overlooking the River Avon. The timber circle was orientated towards the rising Sun on the midwinter solstice, opposing the solar alignments at Stonehenge. The avenue was aligned with the setting Sun on the summer solstice and led from the river to the timber circle. Evidence of huge fires on the banks of the Avon between the two avenues also suggests that both circles were linked. They were perhaps used as a procession route on the longest and shortest days of the year. Parker Pearson speculates that the wooden circle at Durrington Walls was the centre of a 'land of the living', whilst the stone circle represented a 'land of the dead', with the Avon serving as a journey between the two. Later in the Bronze Age, although the exact details of activities during this period are still unclear, the bluestones appear to have been re-erected. They were placed within the outer sarsen circle and may have been trimmed in some way. Like the sarsens, a few have timber-working style cuts in them suggesting that, during this phase, they may have been linked with lintels and were part of a larger structure. This phase saw further rearrangement of the bluestones. They were arranged in a circle between the two rings of sarsens and in an oval at the centre of the inner ring. Some archaeologists argue that some of these bluestones were from a second group brought from Wales. All the stones formed well-spaced uprights without any of the linking lintels inferred in Stonehenge 3 III. The Altar Stone may have been moved within the oval at this time and re-erected vertically. Although this would seem the most impressive phase of work, Stonehenge 3 IV was rather shabbily built compared to its immediate predecessors, as the newly re-installed bluestones were not well-founded and began to fall over. However, only minor changes were made after this phase. Soon afterwards, the northeastern section of the Phase 3 IV bluestone circle was removed, creating a horseshoe-shaped setting (the Bluestone Horseshoe) which mirrored the shape of the central sarsen Trilithons. This phase is contemporary with the Seahenge site in Norfolk. The Y and Z Holes are the last known construction at Stonehenge, built about 1600 BC, and the last usage of it was probably during the Iron Age. Roman coins and medieval artefacts have all been found in or around the monument but it is unknown if the monument was in continuous use throughout British prehistory and beyond, or exactly how it would have been used. Notable is the massive Iron Age hillfort known as Vespasian's Camp (despite its name, not a Roman site) built alongside the Avenue near the Avon. A decapitated seventh-century Saxon man was excavated from Stonehenge in 1923. The site was known to scholars during the Middle Ages and since then it has been studied and adopted by numerous groups. Stonehenge was produced by a culture that left no written records. Many aspects of Stonehenge, such as how it was built and for what purposes it was used, remain subject to debate. A number of myths surround the stones. The site, specifically the great trilithon, the encompassing horseshoe arrangement of the five central trilithons, the heel stone, and the embanked avenue, are aligned to the sunset of the winter solstice and the opposing sunrise of the summer solstice. A natural landform at the monument's location followed this line, and may have inspired its construction. The excavated remains of culled animal bones suggest that people may have gathered at the site for the winter rather than the summer. Further astronomical associations, and the precise astronomical significance of the site for its people, are a matter of speculation and debate. There is little or no direct evidence revealing the construction techniques used by the Stonehenge builders. Over the years, various authors have suggested that supernatural or anachronistic methods were used, usually asserting that the stones were impossible to move otherwise due to their massive size. However, conventional techniques, using Neolithic technology as basic as shear legs, have been demonstrably effective at moving and placing stones of a similar size. The most common theory of how prehistoric people moved megaliths has them creating a track of logs which the large stones were rolled along. Another megalith transport theory involves the use of a type of sleigh running on a track greased with animal fat. Such an experiment with a sleigh carrying a 40-ton slab of stone was successfully conducted near Stonehenge in 1995. A team of more than 100 workers managed to push and pull the slab along the 18-mile (29 km) journey from the Marlborough Downs. Proposed functions for the site include usage as an astronomical observatory or as a religious site. In the 1960s Gerald Hawkins described in detail how the site was apparently set out to observe the Sun and Moon over a recurring 56-year cycle. More recently two major new theories have been proposed. Geoffrey Wainwright, president of the Society of Antiquaries of London, and Timothy Darvill, of Bournemouth University, have suggested that Stonehenge was a place of healing—the primeval equivalent of Lourdes. They argue that this accounts for the high number of burials in the area and for the evidence of trauma deformity in some of the graves. However, they do concede that the site was probably multifunctional and used for ancestor worship as well. Isotope analysis indicates that some of the buried individuals were from other regions. A teenage boy buried approximately 1550 BC was raised near the Mediterranean Sea; a metal worker from 2300 BC dubbed the "Amesbury Archer" grew up near the Alpine foothills of Germany; and the "Boscombe Bowmen" probably arrived from Wales or Brittany, France. On the other hand, Mike Parker Pearson of Sheffield University has suggested that Stonehenge was part of a ritual landscape and was joined to Durrington Walls by their corresponding avenues and the River Avon. He suggests that the area around Durrington Walls Henge was a place of the living, whilst Stonehenge was a domain of the dead. A journey along the Avon to reach Stonehenge was part of a ritual passage from life to death, to celebrate past ancestors and the recently deceased. Both explanations were first mooted in the twelfth century by Geoffrey of Monmouth, who extolled the curative properties of the stones and was also the first to advance the idea that Stonehenge was constructed as a funerary monument. Whatever religious, mystical or spiritual elements were central to Stonehenge, its design includes a celestial observatory function, which might have allowed prediction of eclipse, solstice, equinox and other celestial events important to a contemporary religion. There are other hypotheses and theories. According to a team of British researchers led by Mike Parker Pearson of the University of Sheffield, Stonehenge may have been built as a symbol of "peace and unity", indicated in part by the fact that at the time of its construction, Britain's Neolithic people were experiencing a period of cultural unification. Stonehenge megaliths include smaller bluestones and larger sarsens (a term for silicified sandstone boulders found in the chalk downs of southern England). The bluestones are composed of dolerite, tuff, rhyolite, or sandstone. The igneous bluestones appear to have originated in the Preseli hills of southwestern Wales about 140 miles (230 km) from the monument. The sandstone Altar Stone may have originated in east Wales. Recent analysis has indicated the sarsens originated from West Woods, about 16 miles (26 km) from the monument. Researchers from the Royal College of Art in London have discovered that the monument's igneous bluestones possess "unusual acoustic properties" – when struck they respond with a "loud clanging noise". Rocks with such acoustic properties are frequent in the Carn Melyn ridge of Presili; the Presili village of Maenclochog (Welsh for bell or ringing stones), used local bluestones as church bells until the 18th century. According to the team, these acoustic properties could explain why certain bluestones were hauled such a long distance, a major technical accomplishment at the time. In certain ancient cultures, rocks that ring out, known as lithophonic rocks, were believed to contain mystic or healing powers, and Stonehenge has a history of association with rituals. The presence of these "ringing rocks" seems to support the hypothesis that Stonehenge was a "place for healing" put forward by Darvill, who consulted with the researchers. Researchers studying DNA extracted from Neolithic human remains across Britain determined that the ancestors of the people who built Stonehenge were early European farmers who came from the Eastern Mediterranean, travelling west from there, as well as Western hunter-gatherers from western Europe. DNA studies indicate that early European farmers had a predominantly Aegean ancestry, although their agricultural techniques seem to have come originally from Anatolia. These Aegean farmers moved to Iberia before heading north, reaching Britain in about 4,000 BC. Neolithic individuals in the British Isles were close to Iberian and Central European Early and Middle Neolithic populations, modelled as having about 75% ancestry from Aegean farmers with the rest coming from Western Hunter-Gatherers in continental Europe. They subsequently replaced most of the hunter-gatherer population in the British Isles without mixing much with them. At that time, Britain was inhabited by groups of hunter-gatherers who were the first inhabitants of the island after the last Ice Age ended about 11,700 years ago. Despite their mostly Aegean ancestry, the paternal (Y-DNA) lineages of Neolithic farmers in Britain were almost exclusively of Western Hunter-Gatherer origin. This was also the case among other megalithic-building populations in northwest Europe, meaning that these populations were descended from a mixture of hunter-gatherer males and farmer females. This mixture appears to have happened primarily on the continent before the Neolithic farmers migrated to Britain. The dominance of Western Hunter-Gatherer male lineages in Britain and northwest Europe is also reflected in a general 'resurgence' of hunter-gatherer ancestry, predominantly from males, across western and central Europe in the Middle Neolithic. The Bell Beaker people arrived later, around 2,500 BC, migrating from mainland Europe. The earliest British beakers, most likely speakers of Indo-European languages whose ancestors migrated from the Pontic–Caspian steppe, were similar to those from the Rhine. There was again a large population replacement in Britain. More than 90% of Britain's Neolithic gene pool was replaced with the arrival of the Bell Beaker people, who had approximately 50% WSH ancestry. The Bell Beakers also left their impact on Stonehenge construction. They are also associated with the Wessex culture. The latter appears to have had wide-ranging trade links with continental Europe, going as far as Mycenaean Greece. The wealth from such trade probably permitted the Wessex people to construct the second and third (megalithic) phases of Stonehenge and also indicates a powerful form of social organisation. The Bell Beakers were also associated with the tin trade, which was Britain's only unique export at the time. Tin was important because it was used to turn copper into bronze, and the Beakers derived much wealth from this. There is evidence to suggest that despite the introduction of farming in the British Isles, the practice of cereal cultivation fell out of favor between 3300 and 1500 BC, with much of the population reverting to a pastoralist subsistence pattern focused on hazelnut gathering and pig and cattle rearing. A majority of the major phases of Stonehenge's construction took place during such a period where evidence of large-scale agriculture is equivocal. Similar associations between non-cereal farming subsistence patterns and monumental construction are also seen at Poverty Point and Sannai Maruyama. The Heel Stone lies northeast of the sarsen circle, beside the end portion of Stonehenge Avenue. It is a rough stone, 16 feet (4.9 m) above ground, leaning inwards towards the stone circle. It has been known by many names in the past, including "Friar's Heel" and "Sun-stone". At the Summer solstice an observer standing within the stone circle, looking northeast through the entrance, would see the Sun rise in the approximate direction of the Heel Stone, and the Sun has often been photographed over it. A folk tale relates the origin of the Friar's Heel reference. The Devil bought the stones from a woman in Ireland, wrapped them up, and brought them to Salisbury plain. One of the stones fell into the Avon, the rest were carried to the plain. The Devil then cried out, "No-one will ever find out how these stones came here!" A friar replied, "That's what you think!", whereupon the Devil threw one of the stones at him and struck him on the heel. The stone stuck in the ground and is still there. Brewer's Dictionary of Phrase and Fable attributes this tale to Geoffrey of Monmouth, but though book eight of Geoffrey's Historia Regum Britanniae does describe how Stonehenge was built, the two stories are entirely different. The name is not unique; there was a monolith with the same name recorded in the nineteenth century by antiquarian Charles Warne at Long Bredy in Dorset. The twelfth-century Historia Regum Britanniae ("History of the Kings of Britain"), by Geoffrey of Monmouth, includes a fanciful story of how Stonehenge was brought from Ireland with the help of the wizard Merlin. Geoffrey's story spread widely, with variations of it appearing in adaptations of his work, such as Wace's Norman French Roman de Brut, Layamon's Middle English Brut, and the Welsh Brut y Brenhinedd. According to the tale, the stones of Stonehenge were healing stones, which giants had brought from Africa to Ireland. They had been raised on Mount Killaraus to form a stone circle, known as the Giant's Ring or Giant's Round. The fifth-century king Aurelius Ambrosius wished to build a great memorial to the British Celtic nobles slain by the Saxons at Salisbury. Merlin advised him to use the Giant's Ring. The king sent Merlin and Uther Pendragon (King Arthur's father) with 15,000 men to bring it from Ireland. They defeated an Irish army led by Gillomanius, but were unable to move the huge stones. With Merlin's help, they transported the stones to Britain and re-erected them as they had stood. Mount Killaraus may refer to the Hill of Uisneach. Although the tale is fiction, archaeologist Mike Parker Pearson suggests it may hold a "grain of truth", as evidence suggests the Stonehenge bluestones were brought from the Waun Mawn stone circle on the Irish Sea coast of Wales. Another legend tells how the invading Saxon king Hengist invited British Celtic warriors to a feast but treacherously ordered his men to massacre the guests, killing 420 of them. Hengist erected Stonehenge on the site to show his remorse for the deed. Stonehenge has changed ownership several times since King Henry VIII acquired Amesbury Abbey and its surrounding lands. In 1540 Henry gave the estate to the Earl of Hertford. It subsequently passed to Lord Carleton and then the Marquess of Queensberry. The Antrobus family of Cheshire bought the estate in 1824. During the First World War an aerodrome (Royal Flying Corps "No. 1 School of Aerial Navigation and Bomb Dropping") was built on the downs just to the west of the circle and, in the dry valley at Stonehenge Bottom, a main road junction was built, along with several cottages and a cafe. Stonehenge was one of several lots put up for auction in 1915 by Sir Cosmo Gordon Antrobus, soon after he had inherited the estate from his brother. The auction by Knight Frank & Rutley estate agents in Salisbury was held on 21 September 1915 and included "Lot 15. Stonehenge with about 30 acres, 2 rods, 37 perches [12.44 ha] of adjoining downland." In 1915, Cecil Chubb bought the site for £6,600 (£562,700 in 2023) and gave it to the nation three years later, with certain conditions attached. Although it has been speculated that he purchased it at the suggestion of – or even as a present for – his wife, in fact he bought it on a whim, as he believed a local man should be the new owner. In the late 1920s a nationwide appeal was launched to save Stonehenge from the encroachment of the modern buildings that had begun to rise around it. By 1928 the land around the monument had been purchased with the appeal donations and given to the National Trust to preserve. The buildings were removed (although the roads were not), and the land returned to agriculture. More recently the land has been part of a grassland reversion scheme, returning the surrounding fields to native chalk grassland. During the twentieth century, Stonehenge began to revive as a place of religious significance, this time by adherents of Neopaganism and New Age beliefs, particularly the Neo-druids. The historian Ronald Hutton would later remark that "it was a great, and potentially uncomfortable, irony that modern Druids had arrived at Stonehenge just as archaeologists were evicting the ancient Druids from it." The first such Neo-druidic group to make use of the megalithic monument was the Ancient Order of Druids, who performed a mass initiation ceremony there in August 1905, in which they admitted 259 new members into their organisation. This assembly was largely ridiculed in the press, who mocked the fact that the Neo-druids were dressed up in costumes consisting of white robes and fake beards. Between 1972 and 1984, Stonehenge was the site of the Stonehenge Free Festival. After the Battle of the Beanfield between police and New Age travellers in 1985, this use of the site was stopped for several years and ritual use of Stonehenge is now heavily restricted. Some Druids have arranged an assembling of monuments styled on Stonehenge in other parts of the world as a form of Druidist worship. The earlier rituals were complemented by the Stonehenge Free Festival, loosely organised by the Polytantric Circle, held between 1972 and 1984, during which time the number of midsummer visitors had risen to around 30,000. However, in 1985 the site was closed to festivalgoers by a High Court injunction. A consequence of the end of the festival in 1985 was the violent confrontation between the police and New Age travellers that became known as the Battle of the Beanfield when police blockaded a convoy of travellers to prevent them from approaching Stonehenge. Beginning in 1985, the year of the Battle, no access was allowed into the stones at Stonehenge for any religious reason. This "exclusion-zone" policy continued for almost fifteen years: until just before the arrival of the twenty-first century, visitors were not allowed to go into the stones at times of religious significance, the winter and summer solstices, and the vernal and autumnal equinoxes. However, following a European Court of Human Rights ruling obtained by campaigners such as Arthur Uther Pendragon, the restrictions were lifted. The ruling recognized that members of any genuine religion have a right to worship in their own church, and Stonehenge is a place of worship to Neo-Druids, Pagans and other "Earth based' or 'old' religions. Meetings were organised by the National Trust and others to discuss the arrangements. In 1998, a party of 100 people was allowed access and these included astronomers, archaeologists, Druids, locals, pagans and travellers. In 2000, an open summer solstice event was held and about seven thousand people attended. In 2001, the numbers increased to about 10,000. When Stonehenge was first opened to the public it was possible to walk among and even climb on the stones, but the stones were roped off in 1977 as a result of serious erosion. Visitors are no longer permitted to touch the stones but are able to walk around the monument from a short distance away. English Heritage does, however, permit access during the summer and winter solstice, and the spring and autumn equinox. Additionally, visitors can make special bookings to access the stones throughout the year. Local residents are still entitled to free admission to Stonehenge because of an agreement concerning the moving of a right of way. The access situation and the proximity of the two roads have drawn widespread criticism, highlighted by a 2006 National Geographic survey. In the survey of conditions at 94 leading World Heritage Sites, 400 conservation and tourism experts ranked Stonehenge 75th in the list of destinations, declaring it to be "in moderate trouble". As motorised traffic increased, the setting of the monument began to be affected by the proximity of the two roads on either side—the A344 to Shrewton on the north side, and the A303 to Winterbourne Stoke to the south. Plans to upgrade the A303 and close the A344 to restore the vista from the stones have been considered since the monument became a World Heritage Site. However, the controversy surrounding expensive re-routing of the roads has led to the scheme being cancelled on multiple occasions. On 6 December 2007, it was announced that extensive plans to build Stonehenge road tunnel under the landscape and create a permanent visitors' centre had been cancelled. On 13 May 2009, the government gave approval for a £25 million scheme to create a smaller visitors' centre and close the A344, although this was dependent on funding and local authority planning consent. On 20 January 2010 Wiltshire Council granted planning permission for a centre 1.5 mi (2.4 kilometres) to the west and English Heritage confirmed that funds to build it would be available, supported by a £10m grant from the Heritage Lottery Fund. On 23 June 2013 the A344 was closed to begin the work of removing the section of road and replacing it with grass. The centre, designed by Denton Corker Marshall, opened to the public on 18 December 2013. Throughout recorded history, Stonehenge and its surrounding monuments have attracted attention from antiquarians and archaeologists. John Aubrey was one of the first to examine the site with a scientific eye in 1666, and in his plan of the monument, he recorded the pits that now bear his name, the Aubrey holes. William Stukeley continued Aubrey's work in the early eighteenth century, but took an interest in the surrounding monuments as well, identifying (somewhat incorrectly) the Cursus and the Avenue. He also began the excavation of many of the barrows in the area, and it was his interpretation of the landscape that associated it with the Druids. Stukeley was so fascinated with Druids that he originally named Disc Barrows as Druids' Barrows. The most accurate early plan of Stonehenge was that made by Bath architect John Wood in 1740. His original annotated survey has recently been computer redrawn and published. Importantly Wood's plan was made before the collapse of the southwest trilithon, which fell in 1797 and was restored in 1958. William Cunnington was the next to tackle the area in the early nineteenth century. He excavated some 24 barrows before digging in and around the stones and discovered charred wood, animal bones, pottery and urns. He also identified the hole in which the Slaughter Stone once stood. Richard Colt Hoare supported Cunnington's work and excavated some 379 barrows on Salisbury Plain including on some 200 in the area around the Stones, some excavated in conjunction with William Coxe. To alert future diggers to their work they were careful to leave initialled metal tokens in each barrow they opened. Cunnington's finds are displayed at the Wiltshire Museum. In 1877 Charles Darwin dabbled in archaeology at the stones, experimenting with the rate at which remains sink into the earth for his book The Formation of Vegetable Mould Through the Action of Worms. Stone 22 fell during a fierce storm on 31 December 1900. William Gowland oversaw the first major restoration of the monument in 1901, which involved the straightening and concrete setting of sarsen stone number 56 which was in danger of falling. In straightening the stone he moved it about half a metre from its original position. Gowland also took the opportunity to further excavate the monument in what was the most scientific dig to date, revealing more about the erection of the stones than the previous 100 years of work had done. During the 1920 restoration William Hawley, who had excavated nearby Old Sarum, excavated the base of six stones and the outer ditch. He also located a bottle of port in the Slaughter Stone socket left by Cunnington, helped to rediscover Aubrey's pits inside the bank and located the concentric circular holes outside the Sarsen Circle called the Y and Z Holes. Richard Atkinson, Stuart Piggott and John F. S. Stone re-excavated much of Hawley's work in the 1940s and 1950s, and discovered the carved axes and daggers on the Sarsen Stones. Atkinson's work was instrumental in furthering the understanding of the three major phases of the monument's construction. In 1958 the stones were restored again, when three of the standing sarsens were re-erected and set in concrete bases. The last restoration was carried out in 1963 after stone 23 of the Sarsen Circle fell over. It was again re-erected, and the opportunity was taken to concrete three more stones. Later archaeologists, including Christopher Chippindale of the Museum of Archaeology and Anthropology, University of Cambridge and Brian Edwards of the University of the West of England, campaigned to give the public more knowledge of the various restorations and in 2004 English Heritage included pictures of the work in progress in its book Stonehenge: A History in Photographs. In 1966 and 1967, in advance of a new car park being built at the site, the area of land immediately northwest of the stones was excavated by Faith and Lance Vatcher. They discovered the Mesolithic postholes dating from between 7000 and 8000 BC, as well as a 10-metre (33 ft) length of a palisade ditch – a V-cut ditch into which timber posts had been inserted that remained there until they rotted away. Subsequent aerial archaeology suggests that this ditch runs from the west to the north of Stonehenge, near the avenue. Excavations were once again carried out in 1978 by Atkinson and John Evans, during which they discovered the remains of the Stonehenge Archer in the outer ditch, and in 1979 rescue archaeology was needed alongside the Heel Stone after a cable-laying ditch was mistakenly dug on the roadside, revealing a new stone hole next to the Heel Stone. In the early 1980s Julian C. Richards led the Stonehenge Environs Project, a detailed study of the surrounding landscape. The project was able to successfully date such features as the Lesser Cursus, Coneybury Henge and several other smaller features. In 1993 the way that Stonehenge was presented to the public was called 'a national disgrace' by the House of Commons Public Accounts Committee. Part of English Heritage's response to this criticism was to commission research to collate and bring together all the archaeological work conducted at the monument up to this date. This two-year research project resulted in the publication in 1995 of the monograph Stonehenge in its landscape, which was the first publication presenting the complex stratigraphy and the finds recovered from the site. It presented a rephasing of the monument. More recent excavations include a series of digs held between 2003 and 2008 known as the Stonehenge Riverside Project, led by Mike Parker Pearson. This project mainly investigated other monuments in the landscape and their relationship to the stones — notably, Durrington Walls, where another "Avenue" leading to the River Avon was discovered. The point where the Stonehenge Avenue meets the river was also excavated and revealed a previously unknown circular area which probably housed four further stones, most likely as a marker for the starting point of the avenue. In April 2008, Tim Darvill of the University of Bournemouth and Geoff Wainwright of the Society of Antiquaries began another dig inside the stone circle to retrieve datable fragments of the original bluestone pillars. They were able to date the erection of some bluestones to 2300 BC, although this may not reflect the earliest erection of stones at Stonehenge. They also discovered organic material from 7000 BC, which, along with the Mesolithic postholes, adds support for the site having been in use at least 4,000 years before Stonehenge was started. In August and September 2008, as part of the Riverside Project, Julian C. Richards and Mike Pitts excavated Aubrey Hole 7, removing the cremated remains from several Aubrey Holes that had been excavated by Hawley in the 1920s, and re-interred in 1935. A licence for the removal of human remains at Stonehenge had been granted by the Ministry of Justice in May 2008, in accordance with the Statement on burial law and archaeology issued in May 2008. One of the conditions of the licence was that the remains should be reinterred within two years and that in the intervening period they should be kept safely, privately and decently. A new landscape investigation was conducted in April 2009. A shallow mound, rising to about 16 in (40 centimetres) was identified between stones 54 (inner circle) and 10 (outer circle), clearly separated from the natural slope. It has not been dated but speculation that it represents careless backfilling following earlier excavations seems disproved by its representation in eighteenth- and nineteenth-century illustrations. There is some evidence that, as an uncommon geological feature, it could have been deliberately incorporated into the monument at the outset. A circular, shallow bank, little more than four inches (10 cm) high, was found between the Y and Z hole circles, with a further bank lying inside the "Z" circle. These are interpreted as the spread of spoil from the original Y and Z holes, or more speculatively as hedge banks from vegetation deliberately planted to screen the activities within. In 2010, the Stonehenge Hidden Landscape Project discovered a "henge-like" monument less than 0.62 mi (1 km) away from the main site. This new hengiform monument was subsequently revealed to be located "at the site of Amesbury 50", a round barrow in the Cursus Barrows group. In November 2011, archaeologists from University of Birmingham announced the discovery of evidence of two huge pits positioned within the Stonehenge Cursus pathway, aligned in celestial position towards midsummer sunrise and sunset when viewed from the Heel Stone. The new discovery was made as part of the Stonehenge Hidden Landscape Project which began in the summer of 2010. The project uses non-invasive geophysical imaging technique to reveal and visually recreate the landscape. According to team leader Vince Gaffney, this discovery may provide a direct link between the rituals and astronomical events to activities within the Cursus at Stonehenge. In December 2011, geologists from University of Leicester and the National Museum of Wales announced the discovery of the source of some of the rhyolite fragments found in the Stonehenge debitage. These fragments do not seem to match any of the standing stones or bluestone stumps. The researchers have identified the source as a 230-foot (70 m) long rock outcrop called Craig Rhos-y-felin (51°59′30″N 4°44′41″W / 51.99167°N 4.74472°W / 51.99167; -4.74472 (Craig Rhos-y-Felin)), near Pont Saeson in north Pembrokeshire, located 140 miles (220 km) from Stonehenge. In 2014, the University of Birmingham announced findings including evidence of adjacent stone and wooden structures and burial mounds near Durrington, overlooked previously, that may date as far back as 4000 BC. An area extending to 4.6 square miles (12 km) was studied to a depth of three metres with ground-penetrating radar equipment. As many as seventeen new monuments, revealed nearby, may be Late Neolithic monuments that resemble Stonehenge. The interpretation suggests a complex of numerous related monuments. Also included in the discovery is that the cursus track is terminated by two 16-foot (5 m) wide, extremely deep pits, whose purpose is still a mystery. An announcement in November 2020 stated that a plan to construct a four-lane tunnel for traffic below the site had been approved. This was intended to eliminate the section of the A303 that runs close to the circle. The plan had received opposition from a group of "archaeologists, environmentalists and modern-day druids" according to National Geographic but was supported by others who wanted to "restore the landscape to its original setting and improve the experience for visitors". Opponents of the plan were concerned that artifacts that are underground in the area would be lost or that excavation in the area could de-stabilize the stones, leading to their sinking, shifting or perhaps falling. In February 2021, archaeologists announced the discovery of "vast troves of Neolithic and Bronze Age artifacts" while conducting excavations for the proposed highway tunnel near Stonehenge. The find included Bronze Age graves, late neolithic pottery and C-shaped enclosure on the intended site of the Stonehenge road tunnel. Remains also contained a shale object in one of the graves, burnt flint in C-shaped enclosure and the final resting place of a baby. In January 2022, archaeologists announced the discovery of thousands of prehistoric pits during an electromagnetic induction field survey around Stonehenge. Based on the shape of the pits and the artifacts found inside, the study's lead author, Philippe De Smedt, assumed that six of the 9 large pits excavated were made by prehistoric humans. One of the oldest was about 10000 years old and contained hunting tools. On 14 July 2023, the Department for Transport announced that, despite the original planning application having been overturned by the High Court in 2021, the Transport Secretary, Mark Harper, had approved plans for a 2 mi (3.2 km) road tunnel. In July 2020, a study led by David Nash of the University of Brighton concluded that the large sarsen stones were "a direct chemical match" to those found at West Woods near Marlborough, Wiltshire, some 15 miles (25 km) north of Stonehenge. A core sample, originally extracted in 1958, had recently been returned. First the fifty-two sarsens were analysed using methods including x-ray fluorescence spectrometry to determine their chemical composition which revealed they were mostly similar. Then the core was destructively analysed and compared with stone samples from various locations in southern Britain. Fifty of the fifty-two megaliths were found to match sarsens in West Woods, thereby identifying the probable origin of the stones. During 2017 and 2018, excavations by professor Parker Pearson's team at Waun Mawn, a large stone circle site in the Preseli Hills, suggested that the site had originally housed a 110-metre (360 ft) diameter stone circle of the same size as Stonehenge's original bluestone circle, also orientated towards the midsummer solstice. The circle at Waun Mawn also contained a hole from one stone which had a distinctive pentagonal shape, very closely matching the one pentagonal stone at Stonehenge (stonehole 91 at Waun Mawn / stone 62 at Stonehenge). Soil dating of the sediments within the revealed stone holes, via optically stimulated luminescence (OSL), suggested the absent stones at Waun Mawn had been erected around 3400–3200 BC, and removed around 300–400 years later, a date consistent with theories that the same stones were moved and used at Stonehenge, before later being reorganised into their present locations and supplemented with local sarsens as was already understood. Human activity at Waun Mawn ceased around the same time which has suggested that some people may have migrated to Stonehenge. It has also been suggested that stones from other sources may have been added to Stonehenge, perhaps from other dismantled circles in the region. Further work in 2021 by Parker Pearson's team concluded that the Waun Mawn circle had never been completed, and of the stones which might once have stood at the site, no more than 13 had been removed in antiquity. Historical context Other monuments in the Stonehenge ritual landscape About Stonehenge and replicas of Stonehenge Fiction Similar sites Sites with similar sunrise or sunset alignments Museums with collections from the World Heritage Site
[[File:Aberdeen buildings grey.JPG|thumb|250px|Most buildings in [[Aberdeen]] are made of granite]] [[File:various granites.jpg|thumb|right|250px|Various granites (cut and polished surfaces) The different colours are caused by different proportions of the minerals]] '''Granite''' is a kind of [[igneous rock]], found on [[Earth]] but nowhere else in the [[Solar System]].<ref>[http://www.livescience.com/3070-exciting-evolution-rocks.html Andrea Thompson 2008. The exciting evolution of rocks. ''LiveScience'']</ref><ref>Rosing, Minik T. 2008. On the evolution of minerals. ''Nature'' '''456''' p457.</ref> It is formed from hot, molten [[magma]]. Its colour can be dark or light grey, brown, or even pink, according to the proportions of its [[minerals]]. It is called a plutonic rock because it forms under ground. The magma is forced between other layers of rock by the pressure under the Earth's surface. The magma cools and turns slowly into solid stone. Granite has many different types of minerals in it. These include [[quartz]], [[feldspar]], [[hornblende]], and sometimes [[mica]]. As the magma cools, these minerals form [[crystal]]s. The crystals can be seen easily if the granite is cut and polished. Granite is a common stone on Earth, and makes up a big part of the [[Earth's crust|crust]] (the Earth's outer layer). It is usually found in the [[continental plate]]s of the Earth's crust. Although it forms under the surface of the Earth, there are many places where it has been forced upwards by [[Plate tectonics|tectonic movement]]. When plates in the Earth's crust move together, they get bent and pushed upwards. When this happens, granite [[mountain]]s can be formed. == Origin == Granite forms from silica-rich ([[felsic]]) [[magma]]s. It is the most abundant [[Basement (geology)|basement]] rock. It lies under the thin [[sedimentary rock]]s which cover the continents == Chemical composition == A worldwide average of the chemical composition of granite, by weight percent, based on 2485 analyses:<ref name="Blatt">Blatt, Harvey and Tracy, Robert J. 1997. ''Petrology''. 2nd ed, New York: Freeman, p66. {{ISBN|0-7167-2438-3}}</ref> {| | [[Silicon dioxide|SiO<sub>2</sub>]] || {{bartable|72.04|% (silica)|5}} |- | [[Aluminium oxide|Al<sub>2</sub>O<sub>3</sub>]] || {{bartable|14.42|% (alumina)|5}} |- | [[Potassium oxide|K<sub>2</sub>O]] || {{bartable|4.12|%|5}} |- | [[Sodium oxide|Na<sub>2</sub>O]] || {{bartable|3.69|%|5}} |- | [[Calcium oxide|CaO]] || {{bartable|1.82|%|5}} |- | [[Iron(II) oxide|FeO]] || {{bartable|1.68|%|5}} |- | [[Iron(III) oxide|Fe<sub>2</sub>O<sub>3</sub>]] || {{bartable|1.22|%|5}} |- | [[Magnesium oxide|MgO]] || {{bartable|0.71|%|5}} |- | [[Titanium dioxide|TiO<sub>2</sub>]] || {{bartable|0.30|%|5}} |- | [[Phosphorus pentoxide|P<sub>2</sub>O<sub>5</sub>]] || {{bartable|0.12|%|5}} |- | [[Manganese(II) oxide|MnO]] || {{bartable|0.05|%|5}} |} == Uses == Granite is dense, and can be cut, carved and shaped. It is resistant to water and pollution, and has a range of different colours.<ref>Learn Science intermediate grades 5 to 6, by Mike Evans and Linda Ellis</ref> Kitchen benches are often made of polished granite. Granite is found in many countries of the world. Some countries have beautifully patterned granite which is quarried (cut in open mines) and sold for building material. == Gallery == <gallery class=center> File:The Cheesewring.jpg|The [[Cheesewring]], a granite [[Tor (rock formation)|tor]] on the southern edge of [[Bodmin Moor]], [[Cornwall]] File:Migmatite gran violet.jpg|Polished granite from [[Brazil]] File:Granite wall of chapel La Hougue Bie, Jersey.jpg|A granite wall in [[Jersey]] File:GraniteElephant.jpg|A life-sized elephant carved from granite, [[India]] </gallery> == References == {{reflist}} {{Rock type}} {{commons|Granite}} [[Category:Igneous rocks]]
Granite (/ˈɡrænɪt/) is a coarse-grained (phaneritic) intrusive igneous rock composed mostly of quartz, alkali feldspar, and plagioclase. It forms from magma with a high content of silica and alkali metal oxides that slowly cools and solidifies underground. It is common in the continental crust of Earth, where it is found in igneous intrusions. These range in size from dikes only a few centimeters across to batholiths exposed over hundreds of square kilometers. Granite is typical of a larger family of granitic rocks, or granitoids, that are composed mostly of coarse-grained quartz and feldspars in varying proportions. These rocks are classified by the relative percentages of quartz, alkali feldspar, and plagioclase (the QAPF classification), with true granite representing granitic rocks rich in quartz and alkali feldspar. Most granitic rocks also contain mica or amphibole minerals, though a few (known as leucogranites) contain almost no dark minerals. Granite is nearly always massive (lacking any internal structures), hard, and tough. These properties have made granite a widespread construction stone throughout human history. The word "granite" comes from the Latin granum, a grain, in reference to the coarse-grained structure of such a completely crystalline rock. Granitic rocks mainly consist of feldspar, quartz, mica, and amphibole minerals, which form an interlocking, somewhat equigranular matrix of feldspar and quartz with scattered darker biotite mica and amphibole (often hornblende) peppering the lighter color minerals. Occasionally some individual crystals (phenocrysts) are larger than the groundmass, in which case the texture is known as porphyritic. A granitic rock with a porphyritic texture is known as a granite porphyry. Granitoid is a general, descriptive field term for lighter-colored, coarse-grained igneous rocks. Petrographic examination is required for identification of specific types of granitoids. Granites can be predominantly white, pink, or gray in color, depending on their mineralogy. The alkali feldspar in granites is typically orthoclase or microcline and is often perthitic. The plagioclase is typically sodium-rich oligoclase. Phenocrysts are usually alkali feldspar. Granitic rocks are classified according to the QAPF diagram for coarse grained plutonic rocks and are named according to the percentage of quartz, alkali feldspar (orthoclase, sanidine, or microcline) and plagioclase feldspar on the A-Q-P half of the diagram. True granite (according to modern petrologic convention) contains between 20% and 60% quartz by volume, with 35% to 90% of the total feldspar consisting of alkali feldspar. Granitic rocks poorer in quartz are classified as syenites or monzonites, while granitic rocks dominated by plagioclase are classified as granodiorites or tonalites. Granitic rocks with over 90% alkali feldspar are classified as alkali feldspar granites. Granitic rock with more than 60% quartz, which is uncommon, is classified simply as quartz-rich granitoid or, if composed almost entirely of quartz, as quartzolite. True granites are further classified by the percentage of their total feldspar that is alkali feldspar. Granites whose feldspar is 65% to 90% alkali feldspar are syenogranites, while the feldspar in monzogranite is 35% to 65% alkali feldspar. A granite containing both muscovite and biotite micas is called a binary or two-mica granite. Two-mica granites are typically high in potassium and low in plagioclase, and are usually S-type granites or A-type granites, as described below. Another aspect of granite classification is the ratios of metals that potentially form feldspars. Most granites have a composition such that almost all their aluminum and alkali metals (sodium and potassium) are combined as feldspar. This is the case when K2O + Na2O + CaO > Al2O3 > K2O + Na2O. Such granites are described as normal or metaluminous. Granites in which there is not enough aluminum to combine with all the alkali oxides as feldspar (Al2O3 < K2O + Na2O) are described as peralkaline, and they contain unusual sodium amphiboles such as riebeckite. Granites in which there is an excess of aluminum beyond what can be taken up in feldspars (Al2O3 > CaO + K2O + Na2O) are described as peraluminous, and they contain aluminum-rich minerals such as muscovite. The average density of granite is between 2.65 and 2.75 g/cm (165 and 172 lb/cu ft), its compressive strength usually lies above 200 MPa (29,000 psi), and its viscosity near STP is 3–6·10 Pa·s. The melting temperature of dry granite at ambient pressure is 1215–1260 °C (2219–2300 °F); it is strongly reduced in the presence of water, down to 650 °C at a few hundred megapascals of pressure. Granite has poor primary permeability overall, but strong secondary permeability through cracks and fractures if they are present. A worldwide average of the chemical composition of granite, by weight percent, based on 2485 analyses: The medium-grained equivalent of granite is microgranite. The extrusive igneous rock equivalent of granite is rhyolite. Granitic rock is widely distributed throughout the continental crust. Much of it was intruded during the Precambrian age; it is the most abundant basement rock that underlies the relatively thin sedimentary veneer of the continents. Outcrops of granite tend to form tors, domes or bornhardts, and rounded massifs. Granites sometimes occur in circular depressions surrounded by a range of hills, formed by the metamorphic aureole or hornfels. Granite often occurs as relatively small, less than 100 km stock masses (stocks) and in batholiths that are often associated with orogenic mountain ranges. Small dikes of granitic composition called aplites are often associated with the margins of granitic intrusions. In some locations, very coarse-grained pegmatite masses occur with granite. Granite forms from silica-rich (felsic) magmas. Felsic magmas are thought to form by addition of heat or water vapor to rock of the lower crust, rather than by decompression of mantle rock, as is the case with basaltic magmas. It has also been suggested that some granites found at convergent boundaries between tectonic plates, where oceanic crust subducts below continental crust, were formed from sediments subducted with the oceanic plate. The melted sediments would have produced magma intermediate in its silica content, which became further enriched in silica as it rose through the overlying crust. Early fractional crystallisation serves to reduce a melt in magnesium and chromium, and enrich the melt in iron, sodium, potassium, aluminum, and silicon. Further fractionation reduces the content of iron, calcium, and titanium. This is reflected in the high content of alkali feldspar and quartz in granite. The presence of granitic rock in island arcs shows that fractional crystallization alone can convert a basaltic magma to a granitic magma, but the quantities produced are small. For example, granitic rock makes up just 4% of the exposures in the South Sandwich Islands. In continental arc settings, granitic rocks are the most common plutonic rocks, and batholiths composed of these rock types extend the entire length of the arc. There are no indication of magma chambers where basaltic magmas differentiate into granites, or of cumulates produced by mafic crystals settling out of the magma. Other processes must produce these great volumes of felsic magma. One such process is injection of basaltic magma into the lower crust, followed by differentiation, which leaves any cumulates in the mantle. Another is heating of the lower crust by underplating basaltic magma, which produces felsic magma directly from crustal rock. The two processes produce different kinds of granites, which may be reflected in the division between S-type (produced by underplating) and I-type (produced by injection and differentiation) granites, discussed below. The composition and origin of any magma that differentiates into granite leave certain petrological evidence as to what the granite's parental rock was. The final texture and composition of a granite are generally distinctive as to its parental rock. For instance, a granite that is derived from partial melting of metasedimentary rocks may have more alkali feldspar, whereas a granite derived from partial melting of metaigneous rocks may be richer in plagioclase. It is on this basis that the modern "alphabet" classification schemes are based. The letter-based Chappell & White classification system was proposed initially to divide granites into I-type (igneous source) granite and S-type (sedimentary sources). Both types are produced by partial melting of crustal rocks, either metaigneous rocks or metasedimentary rocks. I-type granites are characterized by a high content of sodium and calcium, and by a strontium isotope ratio, Sr/Sr, of less than 0.708. Sr is produced by radioactive decay of Rb, and since rubidium is concentrated in the crust relative to the mantle, a low ratio suggests origin in the mantle. The elevated sodium and calcium favor crystallization of hornblende rather than biotite. I-type granites are known for their porphyry copper deposits. I-type granites are orogenic (associated with mountain building) and usually metaluminous. S-type granites are sodium-poor and aluminum-rich. As a result, they contain micas such as biotite and muscovite instead of hornblende. Their strontium isotope ratio is typically greater than 0.708, suggesting a crustal origin. They also commonly contain xenoliths of metamorphosed sedimentary rock, and host tin ores. Their magmas are water-rich, and they readily solidify as the water outgasses from the magma at lower pressure, so they less commonly make it to the surface than magmas of I-type granites, which are thus more common as volcanic rock (rhyolite). They are also orogenic but range from metaluminous to strongly peraluminous. Although both I- and S-type granites are orogenic, I-type granites are more common close to the convergent boundary than S-type. This is attributed to thicker crust further from the boundary, which results in more crustal melting. A-type granites show a peculiar mineralogy and geochemistry, with particularly high silicon and potassium at the expense of calcium and magnesium and a high content of high field strength cations (cations with a small radius and high electrical charge, such as zirconium, niobium, tantalum, and rare earth elements.) They are not orogenic, forming instead over hot spots and continental rifting, and are metaluminous to mildly peralkaline and iron-rich. These granites are produced by partial melting of refractory lithology such as granulites in the lower continental crust at high thermal gradients. This leads to significant extraction of hydrous felsic melts from granulite-facies resitites. A-type granites occur in the Koettlitz Glacier Alkaline Province in the Royal Society Range, Antarctica. The rhyolites of the Yellowstone Caldera are examples of volcanic equivalents of A-type granite. M-type granite was later proposed to cover those granites that were clearly sourced from crystallized mafic magmas, generally sourced from the mantle. Although the fractional crystallisation of basaltic melts can yield small amounts of granites, which are sometimes found in island arcs, such granites must occur together with large amounts of basaltic rocks. H-type granites were suggested for hybrid granites, which were hypothesized to form by mixing between mafic and felsic from different sources, such as M-type and S-type. However, the big difference in rheology between mafic and felsic magmas makes this process problematic in nature. Granitization is an old, and largely discounted, hypothesis that granite is formed in place through extreme metasomatism. The idea behind granitization was that fluids would supposedly bring in elements such as potassium, and remove others, such as calcium, to transform a metamorphic rock into granite. This was supposed to occur across a migrating front. However, experimental work had established by the 1960s that granites were of igneous origin. The mineralogical and chemical features of granite can be explained only by crystal-liquid phase relations, showing that there must have been at least enough melting to mobilize the magma. However, at sufficiently deep crustal levels, the distinction between metamorphism and crustal melting itself becomes vague. Conditions for crystallization of liquid magma are close enough to those of high-grade metamorphism that the rocks often bear a close resemblance. Under these conditions, granitic melts can be produced in place through the partial melting of metamorphic rocks by extracting melt-mobile elements such as potassium and silicon into the melts but leaving others such as calcium and iron in granulite residues. This may be the origin of migmatites. A migmatite consists of dark, refractory rock (the melanosome) that is permeated by sheets and channels of light granitic rock (the leucosome). The leucosome is interpreted as partial melt of a parent rock that has begun to separate from the remaining solid residue (the melanosome). If enough partial melt is produced, it will separate from the source rock, become more highly evolved through fractional crystallization during its ascent toward the surface, and become the magmatic parent of granitic rock. The residue of the source rock becomes a granulite. The partial melting of solid rocks requires high temperatures and the addition of water or other volatiles which lower the solidus temperature (temperature at which partial melting commences) of these rocks. It was long debated whether crustal thickening in orogens (mountain belts along convergent boundaries) was sufficient to produce granite melts by radiogenic heating, but recent work suggests that this is not a viable mechanism. In-situ granitization requires heating by the asthenospheric mantle or by underplating with mantle-derived magmas. Granite magmas have a density of 2.4 Mg/m, much less than the 2.8 Mg/m of high-grade metamorphic rock. This gives them tremendous buoyancy, so that ascent of the magma is inevitable once enough magma has accumulated. However, the question of precisely how such large quantities of magma are able to shove aside country rock to make room for themselves (the room problem) is still a matter of research. Two main mechanisms are thought to be important: Of these two mechanisms, Stokes diapirism has been favoured for many years in the absence of a reasonable alternative. The basic idea is that magma will rise through the crust as a single mass through buoyancy. As it rises, it heats the wall rocks, causing them to behave as a power-law fluid and thus flow around the intrusion allowing it to pass without major heat loss. This is entirely feasible in the warm, ductile lower crust where rocks are easily deformed, but runs into problems in the upper crust which is far colder and more brittle. Rocks there do not deform so easily: for magma to rise as a diapir it would expend far too much energy in heating wall rocks, thus cooling and solidifying before reaching higher levels within the crust. Fracture propagation is the mechanism preferred by many geologists as it largely eliminates the major problems of moving a huge mass of magma through cold brittle crust. Magma rises instead in small channels along self-propagating dykes which form along new or pre-existing fracture or fault systems and networks of active shear zones. As these narrow conduits open, the first magma to enter solidifies and provides a form of insulation for later magma. These mechanisms can operate in tandem. For example, diapirs may continue to rise through the brittle upper crust through stoping, where the granite cracks the roof rocks, removing blocks of the overlying crust which then sink to the bottom of the diapir while the magma rises to take their place. This can occur as piecemeal stopping (stoping of small blocks of chamber roof), as cauldron subsidence (collapse of large blocks of chamber roof), or as roof foundering (complete collapse of the roof of a shallow magma chamber accompanied by a caldera eruption.) There is evidence for cauldron subsidence at the Mt. Ascutney intrusion in eastern Vermont. Evidence for piecemeal stoping is found in intrusions that are rimmed with igneous breccia containing fragments of country rock. Assimilation is another mechanism of ascent, where the granite melts its way up into the crust and removes overlying material in this way. This is limited by the amount of thermal energy available, which must be replenished by crystallization of higher-melting minerals in the magma. Thus, the magma is melting crustal rock at its roof while simultaneously crystallizing at its base. This results in steady contamination with crustal material as the magma rises. This may not be evident in the major and minor element chemistry, since the minerals most likely to crystallize at the base of the chamber are the same ones that would crystallize anyway, but crustal assimilation is detectable in isotope ratios. Heat loss to the country rock means that ascent by assimilation is limited to distance similar to the height of the magma chamber. Physical weathering occurs on a large scale in the form of exfoliation joints, which are the result of granite's expanding and fracturing as pressure is relieved when overlying material is removed by erosion or other processes. Chemical weathering of granite occurs when dilute carbonic acid, and other acids present in rain and soil waters, alter feldspar in a process called hydrolysis. As demonstrated in the following reaction, this causes potassium feldspar to form kaolinite, with potassium ions, bicarbonate, and silica in solution as byproducts. An end product of granite weathering is grus, which is often made up of coarse-grained fragments of disintegrated granite. Climatic variations also influence the weathering rate of granites. For about two thousand years, the relief engravings on Cleopatra's Needle obelisk had survived the arid conditions of its origin before its transfer to London. Within two hundred years, the red granite has drastically deteriorated in the damp and polluted air there. Soil development on granite reflects the rock's high quartz content and dearth of available bases, with the base-poor status predisposing the soil to acidification and podzolization in cool humid climates as the weather-resistant quartz yields much sand. Feldspars also weather slowly in cool climes, allowing sand to dominate the fine-earth fraction. In warm humid regions, the weathering of feldspar as described above is accelerated so as to allow a much higher proportion of clay with the Cecil soil series a prime example of the consequent Ultisol great soil group. Granite is a natural source of radiation, like most natural stones. Potassium-40 is a radioactive isotope of weak emission, and a constituent of alkali feldspar, which in turn is a common component of granitic rocks, more abundant in alkali feldspar granite and syenites. Some granites contain around 10 to 20 parts per million (ppm) of uranium. By contrast, more mafic rocks, such as tonalite, gabbro and diorite, have 1 to 5 ppm uranium, and limestones and sedimentary rocks usually have equally low amounts. Many large granite plutons are sources for palaeochannel-hosted or roll front uranium ore deposits, where the uranium washes into the sediments from the granite uplands and associated, often highly radioactive pegmatites. Cellars and basements built into soils over granite can become a trap for radon gas, which is formed by the decay of uranium. Radon gas poses significant health concerns and is the number two cause of lung cancer in the US behind smoking. Thorium occurs in all granites. Conway granite has been noted for its relatively high thorium concentration of 56±6 ppm. There is some concern that some granite sold as countertops or building material may be hazardous to health. Dan Steck of St. Johns University has stated that approximately 5% of all granite is of concern, with the caveat that only a tiny percentage of the tens of thousands of granite slab types have been tested. Resources from national geological survey organizations are accessible online to assist in assessing the risk factors in granite country and design rules relating, in particular, to preventing accumulation of radon gas in enclosed basements and dwellings. A study of granite countertops was done (initiated and paid for by the Marble Institute of America) in November 2008 by National Health and Engineering Inc. of USA. In this test, all of the 39 full-size granite slabs that were measured for the study showed radiation levels well below the European Union safety standards (section 4.1.1.1 of the National Health and Engineering study) and radon emission levels well below the average outdoor radon concentrations in the US. Granite and related marble industries are considered one of the oldest industries in the world, existing as far back as Ancient Egypt. Major modern exporters of granite include China, India, Italy, Brazil, Canada, Germany, Sweden, Spain and the United States. The Red Pyramid of Egypt (c. 2590 BC), named for the light crimson hue of its exposed limestone surfaces, is the third largest of Egyptian pyramids. Pyramid of Menkaure, likely dating 2510 BC, was constructed of limestone and granite blocks. The Great Pyramid of Giza (c. 2580 BC) contains a huge granite sarcophagus fashioned of "Red Aswan Granite". The mostly ruined Black Pyramid dating from the reign of Amenemhat III once had a polished granite pyramidion or capstone, which is now on display in the main hall of the Egyptian Museum in Cairo (see Dahshur). Other uses in Ancient Egypt include columns, door lintels, sills, jambs, and wall and floor veneer. How the Egyptians worked the solid granite is still a matter of debate. Patrick Hunt has postulated that the Egyptians used emery, which has greater hardness on the Mohs scale. The Seokguram Grotto in Korea is a Buddhist shrine and part of the Bulguksa temple complex. Completed in 774 AD, it is an artificial grotto constructed entirely of granite. The main Buddha of the grotto is a highly regarded piece of Buddhist art, and along with the temple complex to which it belongs, Seokguram was added to the UNESCO World Heritage List in 1995. Rajaraja Chola I of the Chola Dynasty in South India built the world's first temple entirely of granite in the 11th century AD in Tanjore, India. The Brihadeeswarar Temple dedicated to Lord Shiva was built in 1010. The massive Gopuram (ornate, upper section of shrine) is believed to have a mass of around 81 tonnes. It was the tallest temple in south India. Imperial Roman granite was quarried mainly in Egypt, and also in Turkey, and on the islands of Elba and Giglio. Granite became "an integral part of the Roman language of monumental architecture". The quarrying ceased around the third century AD. Beginning in Late Antiquity the granite was reused, which since at least the early 16th century became known as spolia. Through the process of case-hardening, granite becomes harder with age. The technology required to make tempered metal chisels was largely forgotten during the Middle Ages. As a result, Medieval stoneworkers were forced to use saws or emery to shorten ancient columns or hack them into discs. Giorgio Vasari noted in the 16th century that granite in quarries was "far softer and easier to work than after it has lain exposed" while ancient columns, because of their "hardness and solidity have nothing to fear from fire or sword, and time itself, that drives everything to ruin, not only has not destroyed them but has not even altered their colour." In some areas, granite is used for gravestones and memorials. Granite is a hard stone and requires skill to carve by hand. Until the early 18th century, in the Western world, granite could be carved only by hand tools with generally poor results. A key breakthrough was the invention of steam-powered cutting and dressing tools by Alexander MacDonald of Aberdeen, inspired by seeing ancient Egyptian granite carvings. In 1832, the first polished tombstone of Aberdeen granite to be erected in an English cemetery was installed at Kensal Green Cemetery. It caused a sensation in the London monumental trade and for some years all polished granite ordered came from MacDonald's. As a result of the work of sculptor William Leslie, and later Sidney Field, granite memorials became a major status symbol in Victorian Britain. The royal sarcophagus at Frogmore was probably the pinnacle of its work, and at 30 tons one of the largest. It was not until the 1880s that rival machinery and works could compete with the MacDonald works. Modern methods of carving include using computer-controlled rotary bits and sandblasting over a rubber stencil. Leaving the letters, numbers, and emblems exposed and the remainder of the stone covered with rubber, the blaster can create virtually any kind of artwork or epitaph. The stone known as "black granite" is usually gabbro, which has a completely different chemical composition. Granite has been extensively used as a dimension stone and as flooring tiles in public and commercial buildings and monuments. Aberdeen in Scotland, which is constructed principally from local granite, is known as "The Granite City". Because of its abundance in New England, granite was commonly used to build foundations for homes there. The Granite Railway, America's first railroad, was built to haul granite from the quarries in Quincy, Massachusetts, to the Neponset River in the 1820s. Engineers have traditionally used polished granite surface plates to establish a plane of reference, since they are relatively impervious, inflexible, and maintain good dimensional stability. Sandblasted concrete with a heavy aggregate content has an appearance similar to rough granite, and is often used as a substitute when use of real granite is impractical. Granite tables are used extensively as bases or even as the entire structural body of optical instruments, CMMs, and very high precision CNC machines because of granite's rigidity, high dimensional stability, and excellent vibration characteristics. A most unusual use of granite was as the material of the tracks of the Haytor Granite Tramway, Devon, England, in 1820. Granite block is usually processed into slabs, which can be cut and shaped by a cutting center. In military engineering, Finland planted granite boulders along its Mannerheim Line to block invasion by Russian tanks in the Winter War of 1939–40. Granite is used as a pavement material. This is because it is extremely durable, permeable and requires little maintenance. For example, in Sydney, Australia black granite stone is used for the paving and kerbs throughout the Central Business District. Curling stones are traditionally fashioned of Ailsa Craig granite. The first stones were made in the 1750s, the original source being Ailsa Craig in Scotland. Because of the rarity of this granite, the best stones can cost as much as US$1,500. Between 60 and 70 percent of the stones used today are made from Ailsa Craig granite. Although the island is now a wildlife reserve, it is still quarried under license for Ailsa granite by Kays of Scotland for curling stones. Granite is one of the rocks most prized by climbers, for its steepness, soundness, crack systems, and friction. Well-known venues for granite climbing include the Yosemite Valley, the Bugaboos, the Mont Blanc massif (and peaks such as the Aiguille du Dru, the Mourne Mountains, the Adamello-Presanella Alps, the Aiguille du Midi and the Grandes Jorasses), the Bregaglia, Corsica, parts of the Karakoram (especially the Trango Towers), the Fitzroy Massif, Patagonia, Baffin Island, Ogawayama, the Cornish coast, the Cairngorms, Sugarloaf Mountain in Rio de Janeiro, Brazil, and the Stawamus Chief, British Columbia, Canada.
{{Year nav|1970}} {{Year in various calendars|1970}} '''1970''' ([[Roman numerals|MCMLXX]]) was a [[common year starting on Thursday]] in the [[Gregorian calendar]], the 1970th year of the [[Common Era]] (CE) and ''[[Anno Domini]]'' (AD) designations, {{Year article header/year number|1970}} == Events == [[File:1978 Ford Pinto hatchback at 2015 Rockville Show 3of5.jpg|thumb|1978 [[Ford Pinto]]]] * [[January 12]] &ndash; The [[Nigerian Civil War]] ends. * [[February 1]] &ndash; The [[Benavidez rail disaster]] near [[Buenos Aires]], Argentina kills 236. * [[February 14]] &ndash; The Who record the iconic live rock album ''[[Live at Leeds]]''. *[[March 1]] &ndash; [[Rhodesia]] declares itself a [[republic]]. * [[April 29]] – The [[United States]] invades [[Cambodia]]. * [[May 4]] – Kent State Shootings: Four students at [[Kent State University]] in [[Ohio]] are killed by [[National Guardsmen]] while protesting against the United States invading Cambodia. * [[May 31]] – A major [[1970 Ancash earthquake|earthquake]] hits southern [[Peru]], killing tens of thousands of people. * [[June 24]] &ndash; The [[United States Senate]] repeals the [[Gulf of Tonkin Resolution]] of [[1964]]. * [[June 28]] &ndash; U.S. ground troops withdraw from [[Cambodia]]. * [[July 1]] – The disputed Territories of [[Pakistan-administered Kashmir]] split to form [[Azad Kashmir]] and [[Gilgit Baltistan]] both of which are autonomous special areas of Pakistan. * [[September 11]] – The [[Ford Motor Company|Ford]] Pinto introduced. * [[September 21]] – ''[[Monday Night Football]]'' debuts on ABC. * [[October 18]] – [[Anwar el-Sadat]] officially becomes president of [[Egypt]]. * [[October 30]] – [[Monsoon]] halts [[Vietnam War]]. *[[November 3]] – [[Salvador Allende]] becomes president of [[Chile]]. *[[December 1]] – [[Luis Echeverría]] becomes [[president of Mexico]]. * The first [[IMAX]] theatre opens in [[Japan]] == Births == [[File:Kevin Smith VidCon 2012.jpg|thumb|[[Kevin Smith]] in 2012]] * [[January 29]] – [[Heather Graham]], [[United States|American]] [[actor|actress]] * [[January 31]] – [[Minnie Driver]], British actress * [[February 28]] – [[Daniel Handler]], American writer (''[[A Series of Unfortunate Events|Lemony Snickets: A Series of Unfortunate Events]]'') * [[March 5]] – [[John Frusciante]], American [[musician]] ([[Red Hot Chili Peppers]]) * [[March 18]] – [[Queen Latifah]], American singer, actress and model * [[April 30]] – [[Wael Dahdouh]], Palestinian journalist * [[May 18]] – [[Tina Fey]], American [[screenwriter]] and actress * [[May 27]] – [[Joseph Fiennes]], [[England|English]] actor June 1 Alex Lalas Popular football player Alison Hinds Queen of Soca popular female soca artiste Of Barbados * [[June 3]] – [[Susan Abulhawa]], Palestinian-American writer and Human rights activist * [[June 22]] - [[Steven Page]], Canadian singer ([[Barenaked Ladies]]) * [[July 8]] – [[Beck]], American singer * [[July 23]] – [[Charisma Carpenter]], American actress * [[August 2]] – [[Kevin Smith]], American [[movie director]] * [[August 8]] – [[Pascal Duquenne]], [[Belgium|Belgian]] actor * [[August 12]] – [[Kristopher Schau]], [[Norway|Norwegian]] singer * [[August 13]] – [[Alan Shearer]], English football player * [[August 20]] – [[John Carmack]], American [[video]] game creator * [[August 31]] - [[Debbie Gibson]], American singer * [[September 9]] – [[Macy Gray]], American singer and actress * [[October 2]] – [[Kelly Ripa]], American actress and [[talk show]] [[Presenter|host]] * [[October 8]] – [[Matt Damon]], American actor * [[October 9]] – [[Steve Jablonsky]], American [[movie]] and [[television]] [[composer]] * [[October 15]] – [[Pernilla Wiberg]], [[Sweden|Swedish]] [[skiing|skier]] * [[October 25]] - [[Ed Robertson]], Canadian singer and [[guitar]]ist ([[Barenaked Ladies]]) * [[October 28]] – [[Greg Eagles]], African-American actor * [[October 29]] – [[Phillip Cocu]], Dutch [[footballer]] * [[October 29]] – [[Edwin van der Sar]], Dutch [[footballer]] * [[December 12]] – [[Jennifer Connelly]], American actress * [[December 18]] – [[Rob Van Dam]], American professional wrestler * [[December 24]] – [[Sascha Fischer]], [[Germany|German]] [[Rugby football|rugby]] player * [[December 29]] – [[Glen Phillips]], American singer and [[guitarist]] == Deaths == === January === [[File:Max Born.jpg|thumb|110px|[[Max Born]]]] * [[January 5]] – [[Max Born]], German physicist (b. [[1882]]) === February === * [[February 2]] – [[Bertrand Russell]], Welsh philosopher and mathematician (b. [[1872]]) * [[February 27]]- [[Marie Dionne]], one of the [[Dionne Quintuplets]] (b. [[1934]]) === March === * [[March 30]] – [[Heinrich Brüning]], [[Chancellor of Germany]] (b. [[1885]]) === April === * [[April 28]] – [[Ed Begley]], American actor (b. [[1901]]) === May === * [[May 12]] – [[Nelly Sachs]], German writer (b. [[1891]]) === June === [[File:Presiden Sukarno.jpg|thumb|110px|[[Sukarno]]]] * [[June 21]] – [[Sukarno]], first [[President of Indonesia]] (b. [[1901]]) === July === * [[July 24]] - [[Beatrice Roberts]], American actress (b. [[1901]]) * [[July 27]] – [[António de Oliveira Salazar]], 12th [[President of Portugal]] (b. [[1889]]) === August === * [[August 10]] – [[Alexander Gode]], German-American linguist (b. [[1906]]) === September === [[File:Jimi Hendrix 1967.png|thumb|110px|[[Jimi Hendrix]]]] [[File:Stevan Kragujevic, Gamal Abdel Naser u Beogradu, 1962.jpg|thumb|110px|[[Gamal Abdel Nasser]]]] * [[September 3]] – [[Vince Lombardi]], American football coach (b. [[1913]]) * [[September 5]] – [[Jochen Rindt]], German racecar driver (b. [[1942]]) * [[September 18]] – [[Jimi Hendrix]], American guitarist (b. [[1942]]) * [[September 25]] – [[Erich Maria Remarque]], German writer (b. [[1898]]) * [[September 28]] – [[Gamal Abdel Nasser]], [[President of Egypt]] (b. [[1918]]) * [[September 29]] - [[Edward Everett Horton]], American actor (b. [[1886]]) === October === [[File:Janis Joplin seated 1970.JPG|thumb|110px|[[Janis Joplin]]]] * [[October 4]] – [[Janis Joplin]], American rock singer (''Mercedes Benz'') (b. [[1943]]) === November === [[File:Bundesarchiv B 145 Bild-F015892-0010, Charles de Gaulle (cropped 2).jpg|thumb|110px|[[Charles de Gaulle]]]] * [[November 4]] – [[Friedrich Kellner]], German writer (b. [[1885]]) * [[November 9]] – [[Charles de Gaulle]], [[President of France]] (b. [[1890]]) === December === * [[December 30]] – [[Sonny Liston]], American boxer (b. [[1932]]) ==Nobel Prizes== *[[Nobel Prize in Physics]] won by [[Hannes Alfvén]], Swedish [[electrical engineer]], [[Plasma (physics)|plasma]] physicist, and [[Louis Néel]], a French [[physicist]] *[[Nobel Prize in Chemistry]] won by [[Luis Federico Leloir]], Argentine [[physician]] and [[biochemist]] *[[Nobel Prize in Physiology or Medicine]] shared by Sir [[Bernard Katz]], [[Ulf von Euler]], and [[Julius Axelrod]] for their work on [[neurotransmitter]]s *[[Nobel Prize in Literature]] won by [[Aleksandr Solzhenitsyn]], Russian writer *[[Nobel Peace Prize]] won by [[Norman Borlaug]], American [[biologist]] and [[humanitarian]] *[[Nobel Memorial Prize in Economic Sciences]] &ndash; [[Paul Samuelson]] == Movies released == * ''[[Airport (1970 movie)|Airport]]'', starring [[Burt Lancaster]] and [[Dean Martin]] * ''[[The Aristocats]]'', an animated movie by [[Walt Disney Productions]] * ''[[Love Story (1970 movie)|Love Story]]'', winner of [[Golden Globe Award]] for Best Motion Picture – Drama * ''[[MASH (movie)|MASH]]'', winner of [[Golden Globe Award]] for Best Motion Picture – Musical or comedy * ''[[Patton (movie)|Patton]]'', winner of 7 [[Academy Awards]] * ''[[Tora! Tora! Tora!]]'' about the [[attack on Pearl Harbor]] in 1941 * ''[[Woodstock (movie)|Woodstock]]'' == Hit songs == [[File:Carpenters - Nixon - Office.png|thumb|[[The Carpenters]] in 1972]] * "(They Long to Be) Close to You" – [[Carpenters]] * "Maybe I'm Amazed" – [[Paul McCartney]] * "After Midnight" – [[Eric Clapton]] * "Ain't No Mountain High Enough" – [[Diana Ross]] * "American Woman" – [[The Guess Who]] * "Amos Moses" – [[Jerry Reed]] * "Border Song (Holy Moses)" – [[Aretha Franklin]] * "Bridge Over Troubled Water" – [[Simon and Garfunkel]] * "Call Me" – [[Aretha Franklin]] * "Can't Stop Loving You" – [[Tom Jones (singer)|Tom Jones]] * "[[Cecilia (Simon & Garfunkel song)|Cecilia]]" – [[Simon and Garfunkel]] * "Celebrate" – [[Three Dog Night]] * "Close To You " – [[Carpenters]] * "Come And Get It" – [[Badfinger]] * "Come Together /Something " – [[The Beatles]] * "Cupid " – [[Johnny Nash]] * "Does Anybody Really Know What Time it is" – [[Chicago]] * "Down on the Corner" – [[Creedence Clearwater Revival]] * "El Condor Pasa" – [[Simon and Garfunkel]] * "Eli's Coming" – [[Three Dog Night]] * "Everybody's Out of Town" – [[B.J. Thomas]] * "Fire and Rain" – [[James Taylor]] * "For You Blue" – [[The Beatles]] * "Hand Me Down World" – [[The Guess Who]] * "Heaven Help Us All" – [[Stevie Wonder]] * "If You Could Read My Mind" – [[Gordon Lightfoot]] * "I'll Never Fall in Love Again" – [[Dionne Warwick]] * "Immigrant Song" – [[Led Zeppelin]] * "Indiana Wants Me – [[R. Dean Taylor]] * "Instant Karma (We All Shine On)" – [[John Lennon]] * "Isn't It a Pity" – [[George Harrison]] * "It Don't Matter to Me" – [[Bread]] * "It's a New Day (Parts 1 & 2)" – James Brown * "It's Only Make Believe" – [[Glen Campbell]] * "I've Lost You" – [[Elvis Presley]] * "Jam Up and Jelly Tight" – [[Tommy Roe]] * "Joanne" – [[Michael Nesmith]] & The First National Band * "Kentucky Rain" – [[Elvis Presley]] * "Leaving on a Jet Plane" – [[Peter, Paul and Mary]] * "Let It Be" – [[The Beatles]] * "Let Me Go to Him" – [[Dionne Warwick]] * "Lola " – [[The Kinks]] * "Lonely Days " – [[The Bee Gees]] * "Long and Winding Road" – [[The Beatles]] * "Looking Out My Back Door " – [[Creedence Clearwater Revival]] * "Love the One You're With" – [[Stephen Stills]] * "Mama Told Me (Not to Come)" – [[Three Dog Night]] * "Mississippi Queen" – [[Mountain]] * "My Sweet Lord" – [[George Harrison]] * "No Matter What" – [[Badfinger]] * "No Time " – [[The Guess Who]] * "Oh Happy Day" – [[Glen Campbell]] * "Only Love Can Break Your Heart" – [[Neil Young]] * "Out In the Country" – [[Three Dog Night]] * "Psychedelic Shack" – [[The Temptations]] * "Raindrops Keep Fallin' On My Head" – [[B.J. Thomas]] * "Reach Out and Touch (Somebody's Hand)" – [[Diana Ross]] * "Remember Me" – [[Diana Ross]] * "Ride Captain Ride" – [[Blues Image]] * "See Me, Feel Me " – [[The Who]] * "Share the Land" – [[The Guess Who]] * "Signed, Sealed, Delivered, I'm Yours" – [[Stevie Wonder]] * "Snowbird" – [[Anne Murray]] * "Son of a Preacher Man" – [[Aretha Franklin]] * "Stoney End" – [[Barbra Streisand]] * "Summertime Blues" – [[The Who]] * "The Long and Winding Road" – [[The Beatles]] * "The Wonder of You" – [[Elvis Presley]] * "Travelin' Band" – [[Creedence Clearwater Revival]] * "We've Only Just Begun " – [[Carpenters]] * "What Is Truth?" – [[Johnny Cash]] * "Whole Lotta Love " – [[Led Zeppelin]] * "Who'll Stop The Rain" – [[Creedence Clearwater Revival]] * "Your Song" – [[Elton John]] == New books == * ''[[Deliverance]]'' by [[James Dickey]] *''[[Fantastic Mr Fox]]'' by [[Roald Dahl]] *''[[Fire from Heaven]]'' by [[Mary Renault]] *''[[Of a Fire on the Moon]]'' by [[Norman Mailer]] *''[[The Guardians (novel)|The Guardians]]'' by [[John Christopher]] ([[Sam Youd]]) *''[[Jonathan Livingston Seagull]]'' by [[Richard Bach]] *''[[The Paper Chase (novel)|The Paper Chase]]'' by [[John Jay Osborn, Jr.]] {{Commonscat|1970}} [[Category:1970| ]]
1970 (MCMLXX) was a common year starting on Thursday of the Gregorian calendar, the 1970th year of the Common Era (CE) and Anno Domini (AD) designations, the 970th year of the 2nd millennium, the 70th year of the 20th century, and the 1st year of the 1970s decade.
{{About|the city in the state of New York|other places sharing the same name|Syracuse}} {{Use American English|date=May 2023}} {{Use mdy dates|date=May 2023}} {{Infobox settlement <!--See the Table at Infobox Settlement for all fields and descriptions of usage-->| name = Syracuse, New York | settlement_type = City | named_for = [[Syracuse, Sicily]] | nickname = The Salt City, 'Cuse, The Emerald City, The Heart of New York, The Paris of New York | motto = | image_skyline = {{multiple image | total_width = 280 | border = infobox | perrow = 1/2/2/2/1 | caption_align = center | image1 = Syracuse NY (cropped).jpg | alt1 = Downtown Syracuse | caption1 = [[Downtown Syracuse]] | image2 = Round Lake (2) - Fayetteville NY.jpg | alt2 = Green Lakes State Park | caption2 = [[Green Lakes State Park]] | image3 = Carrier Dome - New Roof 2021 (Jimhoward03).jpg | alt3 = JMA Wireless Dome | caption3 = [[JMA Wireless Dome]] | image4 = Clinton square ice rink winter.jpg | alt4 = Clinton Square | caption4 = [[Clinton Square]] | image5 = HannoverSquareSyracuse.jpg | alt5 = Hanover Square | caption5 = [[Hanover Square, Syracuse|Hanover Square]] | image6 = Columbus Circle with humans.jpg | alt6 = Columbus Circle Historic District | caption6 = [[Montgomery Street–Columbus Circle Historic District|Columbus Circle Historic District]] | image7 = White Memorial Building.jpg | alt7 = South Salina Street Downtown Historic District | caption7 = [[South Salina Street Downtown Historic District]] | image8 = Columbus Circle panorama, Syracuse, New York - 20210508.jpg | alt8 = Panorama of Columbus Circle Historic District | caption8 = Panorama of [[Montgomery Street–Columbus Circle Historic District|Columbus Circle Historic District]] }} | image_flag = Flag of Syracuse, New York.svg | image_seal = | image_map = {{maplink|frame=yes|plain=yes|frame-align=center|frame-width=280|frame-height=280|frame-coord=SWITCH:{{coord|43|1|49|N|76|8|40|W}}###{{coord|qid=Q114904}}###{{coord|qid=Q1384}}###{{coord|qid=Q30}}|zoom=SWITCH:11;9;5;3|type=SWITCH:shape-inverse;shape;point;point|marker=city|stroke-width=2|stroke-color=#5f5f5f|id2=SWITCH:Q128069;Q114904;Q1384;Q30|type2=shape|fill2=#ffffff|fill-opacity2=SWITCH:0;0.1;0.1;0.1|stroke-width2=2|stroke-color2=#5f5f5f|stroke-opacity2=SWITCH:0;1;1;1|switch=Syracuse;Onondaga County;New York;the United States}} | subdivision_type = Country | subdivision_name = {{US}} | subdivision_type1 = State | subdivision_name1 = {{flag|New York}} | subdivision_type2 = Region | subdivision_name2 = [[Central New York]] | subdivision_type3 = Metro | subdivision_name3 = [[Syracuse Metropolitan Area|Syracuse Metropolitan Statistical Area]] | subdivision_type4 = [[List of counties in New York|County]] | subdivision_name4 = [[Onondaga County, New York|Onondaga]] | government_type = [[Strong Mayor|Strong mayor-council]] | leader_title1 = [[List of mayors of Syracuse, New York|Mayor]] | leader_name1 = [[Ben Walsh]] ([[Independent politician|I]]) | leader_title2 = [[city council|Common Council]] | leader_name2 = {{Collapsible list |title = Members' List |frame_style = border:none; padding: 0; |title_style = <!-- (optional) --> |list_style = text-align:left;display:none; |1 = President: |2 = • Helen Hudson ([[United States Democratic Party|D]]) |3 = At Large Members: |4 = • Chol Majok ([[United States Democratic Party|D]]) |5 = • Rasheada Caldwell ([[United States Democratic Party|D]]) |6 = • Rita Paniagua ([[United States Democratic Party|D]]) |7 = • Amir Gethers ([[United States Democratic Party|D]]) |8 = • D1: Marty Nave ([[United States Democratic Party|D]]) |9 = • D2: Patrick Hogan ([[United States Democratic Party|D]]) |10 = • D3: Corey Williams ([[United States Democratic Party|D]]) |11 = • D4: Patrona Jones-Rowser ([[United States Democratic Party|D]]) |12 = • D5: Jimmy Monto ([[United States Democratic Party|D]]) }} | established_title = Incorporated (village) | established_date = {{start date and age|1825}} | established_title2 = Incorporated (city) | established_date2 = {{start date and age|1847}} | area_magnitude = | unit_pref = Imperial | area_total_km2 = 66.41 | area_total_sq_mi = 25.64 | area_land_km2 = 64.90 | area_land_sq_mi = 25.06 | area_water_km2 = 1.51 | area_water_sq_mi = 0.58 | area_water_percent = 2.15 | area_urban_km2 = | area_urban_sq_mi = | area_metro_km2 = | area_metro_sq_mi = | population_as_of = [[2020 United States Census|2020]] | population_note = <ref name="QF2020">{{cite web |url=https://www.census.gov/quickfacts/fact/table/syracusecitynewyork/PST045219 |title=QuickFacts: Syracuse city, New York |access-date=November 5, 2021 |website=census.gov |archive-date=November 23, 2021 |archive-url=https://web.archive.org/web/20211123053943/https://www.census.gov/quickfacts/fact/table/syracusecitynewyork/PST045219 |url-status=live }}</ref> | population_total = 148620 | population_density_km2 = 2289.88 | population_density_sq_mi = 5930.80 | population_metro = 662,057 (US: [[List of Metropolitan Statistical Areas|91st]]) | population_urban = 413,660 (US: [[List of United States urban areas|102nd]]) | population_density_urban_km2 = 884.7 | population_density_urban_sq_mi = 2,291.3 | population_demonym = Syracusan | timezone = [[North American Eastern Time Zone|Eastern]] | utc_offset = −5 | timezone_DST = Eastern Daylight Time | utc_offset_DST = −4 | coordinates = {{coord|43|02|49|N|76|08|40|W|region:US-NY|display=inline,title}} | elevation_m = 116–135 | elevation_ft = 380–440 | website = {{URL|https://www.syr.gov/}} | postal_code_type = [[ZIP Code]] | postal_code = 132xx | area_codes = [[Area code 315|315]], [[Area code 680|680]] | blank_name = [[Federal Information Processing Standard|FIPS code]] | blank_info = 36-73000 | blank1_name = [[Geographic Names Information System|GNIS]] feature ID | blank1_info = 0966966 | pop_est_as_of = | pop_est_footnotes = | area_footnotes = <ref name="TigerWebMapServer">{{cite web|title=ArcGIS REST Services Directory|url=https://tigerweb.geo.census.gov/arcgis/rest/services/TIGERweb/Places_CouSub_ConCity_SubMCD/MapServer|publisher=United States Census Bureau|accessdate=September 20, 2022}}</ref> | official_name = | population_blank1 = 738,305 (US: [[List of Combined Statistical Areas|72nd]]) <!-- General information --------------->| population_blank1_title = [[Combined statistical area|CSA]] | population_est = | population_footnotes = }} '''Syracuse''' is a [[city]] in the [[U.S. state]] of [[New York (state)|New York]]. It is also the [[county seat]] of [[Onondaga County, New York|Onondaga County]]. Syracuse has a [[population]] of 148,620<ref name="QF2020"/> and a [[metropolitan area]] population of 662,000.<ref>{{Cite web|title=Syracuse Metro Area - FEDERAL RESERVE BANK of NEW YORK|url=https://www.newyorkfed.org/regional-economy/profiles/syracuse|access-date=2023-08-13|website=www.newyorkfed.org}}</ref> Syracuse is sometimes called the "Salt City" because [[salt]] [[mining]] was once the main [[industry]]. Syracuse is named after [[Syracuse, Sicily]], a city on the [[Italy|Italian]] [[island]] of [[Sicily]]. Syracuse is home to [[Syracuse University]] and many company company [[Office|offices]]. ==History== French missionaries were the first Europeans to be in this area. ==References== {{Reflist}} {{Onondaga County, New York}} {{authority control}} {{US-geo-stub}} [[Category:County seats in New York (state)]] [[Category:Syracuse, New York| ]]
Syracuse (/ˈsɪrəkjuːz, ˈsɛr-, -kjuːs/ SIRR-ə-kewz, SERR-, -kewss) is a city and the county seat of Onondaga County, New York, United States. With a population of 148,620 and a metropolitan area of 662,057, it is the fifth-most populated city and 13th-most populated municipality in the state of New York. Formally established in 1820, Syracuse was named after the classical Greek city Syracuse (Siracusa in Italian), a city on the eastern coast of the Italian island of Sicily, for its similar natural features. It has historically functioned as a major crossroads, first between the Erie Canal and its branch canals, then of the railway network. Today, the city is at the intersection of Interstates 81 and 90, and its airport is the largest in Central New York, a five-county region of over one million inhabitants. Syracuse is the economic and educational hub of Central New York. It hosts a number of convention sites, including a large downtown convention complex, and is home to prominent instituions such as Syracuse University, SUNY Upstate Medical University, SUNY ESF, and Le Moyne College. French missionaries were the first Europeans to come to this area, arriving to work with and convert the Native Americans in the mid-17th century. At the invitation of the Onondaga Nation, one of the five nations of the Iroquois Confederacy, a Jesuit priest by the name of Simon Le Moyne, accompanied by soldiers and coureurs des bois, including Pierre Esprit Radisson, set up a mission, known as Ste. Marie de Gannentaha on the northeast shore of Onondaga Lake. Jesuit missionaries reported salty brine springs around the southern end of what they referred to as "Salt Lake", known today as Onondaga Lake in honor of the historic tribe. French fur traders established trade throughout the New York area among the Iroquois. Dutch and English colonists also were traders, and the English nominally claimed the area, from their upstate base at Albany, New York. During the American Revolutionary War, the highly decentralized Iroquois divided into groups and bands that supported the British, and two tribes that supported the American-born rebels, or patriots. Settlers came into central and western New York from eastern parts of the state and New England after the American Revolutionary War and various treaties with and land sales by Native American tribes. The subsequent designation of this area by the state of New York as the Onondaga Salt Springs Reservation provided the basis for commercial salt production. Such production took place from the late 1700s through the early 1900s. Brine from wells that tapped into halite (common salt) beds in the Salina shale near Tully, New York, 15 miles south of the city, was developed in the 19th century. It is the north-flowing brine from Tully that is the source of salt for the "salty springs" found along the shoreline of Onondaga Lake. The rapid development of this industry in the 18th and 19th centuries led to the nicknaming of this area as "The Salt City". The original settlement of Syracuse was a conglomeration of several small towns and villages and was not recognized with a post office by the United States Government. Establishing the post office was delayed because the settlement did not have a name. Joshua Forman wanted to name the village Corinth. When John Wilkinson applied for a post office in that name in 1820, it was denied because the same name was already in use in Saratoga County, New York. Having read a poetic description of Syracuse, Sicily (Siracusa), Wilkinson saw similarities to the lake and salt springs of this area, which had both "salt and freshwater mingling together". On February 4, 1820, Wilkinson proposed the name "Syracuse" to a group of fellow townsmen; it became the name of the village and the new post office. The first Solvay Process Company plant in the United States was erected on the southwestern shore of Onondaga Lake in 1884. The village was called Solvay to commemorate the inventor Ernest Solvay. In 1861, he developed the ammonia-soda process for the manufacture of soda ash (anhydrous sodium carbonate) from brine wells dug in the southern end of Tully valley (as a source of sodium chloride) and limestone (as a source of calcium carbonate). The process was an improvement over the earlier Leblanc process. The Syracuse Solvay plant was the incubator for a large chemical industry complex owned by Allied Signal in Syracuse. While this industry stimulated development and provided many jobs in Syracuse, it left Onondaga Lake as the most polluted in the nation. The salt industry declined after the Civil War, but a new manufacturing industry arose in its place. Throughout the late 1800s and early 1900s, numerous businesses and stores were established, including the Franklin Automobile Company, which produced the first air-cooled engine in the world; the Century Motor Vehicle Company; the Smith Corona company; and the Craftsman Workshops, the center of Gustav Stickley's handmade furniture empire. On March 24, 1870, Syracuse University was founded. The State of New York granted the new university its own charter, independent of Genesee College, which had unsuccessfully tried to move to Syracuse the year before. The university was founded as coeducational. President Peck stated at the opening ceremonies, "The conditions of admission shall be equal to all persons... there shall be no invidious discrimination here against woman.... brains and heart shall have a fair chance... " Syracuse implemented this policy and attracted a high proportion of women students. In the College of Liberal Arts, the ratio between male and female students during the 19th century was approximately even. The College of Fine Arts was predominantly female, and a low ratio of women enrolled in the College of Medicine and the College of Law. The first New York State Fair was held in Syracuse in 1841. Between 1842 and 1889, the Fair was held among 11 New York cities before finding a permanent home in Syracuse. It has been an annual event since then, except between 1942 and 1947, when the grounds were used as a military base during World War II, and in 2020, due to the outbreak of the COVID-19 pandemic. As part of the racial incidents happening all over the country during the 1919 Red Summer, on July 31, 1919, there was a violent riot between white and black workers of the Syracuse Globe Malleable Iron Works. Syracuse is home to the only "green on top" traffic light. The "green on top" traffic light was installed in 1928 as a result of local youths throwing rocks at the "British red" light that was originally on top. These locals became known as "stonethrowers" and the neighborhood now has the Tipperary Hill Heritage Memorial on the corner of Tompkins Street and Milton Avenue to commemorate this history. World War II stimulated significant industrial expansion in the area: of specialty steel, fasteners, and custom machining. After the war, two of the Big Three automobile manufacturers (General Motors and Chrysler) had major operations in the area. Syracuse was also headquarters for Carrier Corporation, and Crouse-Hinds manufactured traffic signals in the city. General Electric, with its headquarters in Schenectady to the east, had its main television manufacturing plant at Electronics Parkway in Syracuse. The manufacturing industry in Syracuse began to falter in the 1970s, as the industry restructured nationwide. Many small businesses failed during this time, which contributed to the already increasing unemployment rate. Rockwell International moved its factory outside New York state. General Electric moved its television manufacturing operations to Suffolk, Virginia, and later offshore to Asia. The Carrier Corporation moved its headquarters out of Syracuse, relocated its manufacturing operations out of state, and outsourced some of its production to Asian facilities. Although the city population has declined since 1950, the Syracuse metropolitan area population has remained fairly stable, growing by 2.5 percent since 1970. While this growth rate is greater than much of Upstate New York, it is far below the national average during that period. The Syracuse Community Grid project is a 2023 highway teardown project taking place in Downtown Syracuse, with the goal of improving the city. Syracuse is located at 43°2′49″N 76°8′40″W / 43.04694°N 76.14444°W / 43.04694; -76.14444 (43.046899, −76.144423). It is located about 87 miles (140 km) east of Rochester, 150 miles (240 km) east of Buffalo, and 145 miles (230 km) west of the state capital, Albany. It is also the halfway point between New York City and Toronto, about 245 miles (390 km) from each, Toronto to the northwest and NYC to the southeast. According to the United States Census Bureau, the city has a total area of 66 square kilometres (25.6 square miles), of which 65 square kilometres (25.1 square miles) is land and 1.6 square kilometres (0.6 square miles) (2.15%) water. The city developed at the northeast corner of the Finger Lakes region. The city has many neighborhoods that were originally independent villages, which joined the city over the years. Although the central part of Syracuse is flat, many of its neighborhoods are on small hills such as University Hill and Tipperary Hill. Land to the north of Syracuse is generally flat, while land to the south is hilly. About 27 percent of Syracuse's land area is covered by 890,000 trees — a higher percentage than in Albany, Rochester or Buffalo. The Labor Day Storm of 1998 was a derecho that destroyed approximately 30,000 trees. The sugar maple accounts for 14.2 percent of Syracuse's trees, followed by the Northern white cedar (9.8 percent) and the European buckthorn (6.8 percent). The most common street tree is the Norway maple (24.3 percent), followed by the honey locust (9.3 percent). The densest tree cover in Syracuse is in the two Valley neighborhoods, where 46.6 percent of the land is covered by trees. The lowest tree cover percentage is found in the densely developed downtown, which has only 4.6 percent trees. Syracuse's main water source is Skaneateles Lake, one of the country's cleanest lakes, located 15 mi (24 km) southwest of the city. Water from nearby Onondaga Lake is not potable due to industrial dumping that spanned many decades, leaving the lake heavily polluted. Incoming water is left unfiltered, and chlorine is added to prevent bacterial growth. Most of the environmental work to achieve lake cleanup was scheduled to be completed by 2016; however Honeywell, the company tasked with the cleanup, announced the project's completion in late 2017. For periods of drought, there is also a backup line which uses water from Lake Ontario. Onondaga Creek, a waterway that runs through downtown, flows northward through the city. The Onondaga Creekwalk borders this, connecting the Lakefront, Inner Harbor, Franklin Square and Armory Square neighborhoods. The creek continues through the Valley and ultimately to the Onondaga Nation. The creek is navigable, but it can be a challenge. Its channelized nature speeds up its flow, particularly in the spring, when it may be dangerous. After some youngsters drowned in the creek, some residential areas fenced-off the creek in their neighborhoods. The City of Syracuse officially recognizes 26 neighborhoods within its boundaries. Some of these have small additional neighborhoods and districts inside of them. In addition, Syracuse also owns and operates Syracuse Hancock International Airport on the territory of four towns north of the city. Syracuse's neighborhoods reflect the historically ethnic and multicultural population. Traditionally, Irish, Polish and Ukrainian Americans settled on its west side (see Tipperary Hill); Jewish Americans on its east side; German and Italian Americans on the north side; and African-Americans on its south side. In recent years, large numbers of refugees from the Middle East have settled mainly on the north side as well. Syracuse has a hot-summer humid continental climate (Dfa), as mean July temperatures are just above the 71.6 °F (22.0 °C) threshold needed for a hot-summer climate. The city is known for its high snowfall, 115.6 inches (2.94 meters) on average; Syracuse receives the most annual average snow of any metropolitan area in the United States. Syracuse usually wins the Golden Snowball Award among Upstate cities. Its record seasonal (July 1 to June 30 of the following year) snowfall so far is 192.1 in (4.88 m) during the winter of 1992–93, while the snowiest calendar month was January 2004, with 78.1 in (1.98 m) accumulated. The high snowfall is a result of the city receiving both heavy snow from the lake effect of nearby Lake Ontario (of the Great Lakes) and nor'easter snow from storms driven from the Atlantic Ocean. Snow most often falls in small (about 1–3 inches or 2.5–7.6 centimetres), almost daily doses, over a period of several days. Larger snowfalls do occur occasionally, and even more so in the northern suburbs. The Blizzard of 1993 was described as the Storm of the Century. Some 42.9 in (109 cm) fell on the city within 48 hours, with 35.6 in (90 cm) falling within the first 24 hours. Syracuse received more snow than any other city in the country during this storm, which shattered a total of eight local records, including the most snow in a single snowstorm. A second notable snowfall was the Blizzard of 1966, with 42.3 in (107 cm). The Blizzard of '58 occurred in February (16–17th) across Oswego and Onondaga counties. This storm was classified as a blizzard due to the high winds, blowing snow, and cold; 26.1 in (66 cm) of snow was measured at Syracuse and drifts reached 20 ft (6.1 m) in Oswego County. (See Thirtieth Publication of the Oswego County Historical Society, (1969) and The Climate and Snow Climatology of Oswego N.Y., (1971) Syracuse on average receives an annual precipitation of 38.47 inches (977 millimeters), with the months of July through September being the wettest in terms of total precipitation, while precipitation occurs on more days each month during the snow season. The normal monthly mean temperature ranges from 23.6 °F (−4.7 °C) in January to 71.3 °F (21.8 °C) in July. The record high of 102 °F (39 °C) was recorded on July 9, 1936, and the record low of −26 °F (−32 °C) has occurred three times since 1942, the last being February 18, 1979. In the early 21st century, a handful of previous heat records have been broken in the city. For example, July 2020 became the hottest month on record, with a mean temperature of 77.1 °F (25.1 °C), while the summers (June–August) of 2005, 2020, and 2012 were, respectively, the hottest, third-hottest, and fourth-hottest summers on record. Additionally, 2017 and 2018 saw consecutive monthly high temperature records broken in February, of 71 °F (22 °C) on February 24, 2017, and 75 °F (24 °C) on February 21, 2018, in addition to four consecutive days at or above 60 °F (16 °C). The latter was the warmest winter day on record. As of the census of 2010, there were 145,170 people, 57,355 households, and 28,455 families residing in the city. The racial makeup of the city was 56.0% White, 29.5% African American, 1.1% Native American, 5.5% Asian, 0.03% Pacific Islander, 2.7% from other races, and 5.1% from two or more races. Hispanic or Latino residents of any race were 8.3% of the population. The largest ancestries include Italian (29.5%), Irish (18.4%), Polish (15.3%), German (9.6%), English (4.5%), and Slovak (3.6%). Non-Hispanic Whites were 52.8% of the population in 2010, down from 87.2% in 1970. Suburbanization attracted residents outside the city, even as new immigrant and migrant groups increased. There were 57,355 households, out of which 29% had children under the age of 18 living with them, 9.3% were married couples living together, 20.8% had a female householder with no husband present, and 50.4% were non-families. 38.4% of all households were made up of individuals, and 10.4% had someone living alone who was 65 years of age or older. The average household size was 2.31 and the average family size was 3.14. The city's age distribution was as follows: 19% of residents were under the age of 15, 23% from 15 to 24, 25.6% from 25 to 44, 21.7% from 45 to 64, and 10.5% who were 65 years of age or older. The median age was 29.6 years. For every 100 females, there were 91 males. For every 100 females age 18 and over, there were 87.89 males. According to the 2014 estimates from the American Community Survey, the median income for a household in the city was $31,566, and the median income for a family was $38,794. Males had a median income of $39,537 versus $33,983 for females. The per capita income for the city was $19,283. About 28.2% of families and 35.1% of the population were below the poverty line, including 50% of those under age 18 and 16.7% of those age 65 and over. As of 2017, the United States Census Bureau indicated an estimated population of 146,396. Syracuse ranks 50th in the United States for transit ridership and 12th for most pedestrian commuters. Each day, 38,332 people commute into Onondaga County from the four adjoining counties (2006). In 2021, Syracuse experienced a population growth for the first time in over 70 years, growing 0.24% Work Area Profile Report Worker Age Earnings Christianity: Most Christians in Syracuse are Catholic, reflecting the influence of 19th and early 20th-century immigration patterns, when numerous Irish, German, Italian and eastern European Catholics settled in the city. The city has the Roman Catholic Cathedral of the Immaculate Conception. Syracuse is also home to the combined novitiate of the United States Northeast (UNE) and Maryland Provinces of the Society of Jesus (Jesuits). The historic Basilica of the Sacred Heart of Jesus is located near downtown (Roman Catholic, with Mass, offered in English and Polish). Tridentine Mass is offered multiple times a week at Transfiguration Parish in the Eastside neighborhood. Another major historic church is the Episcopal St. Paul's Cathedral. Both cathedrals are located at Columbus Circle. They represent their respective dioceses, the Diocese of Syracuse (Roman Catholic) and the Diocese of Central New York (Episcopal). The Assembly of God, the American Baptist Churches of the US, the Southern Baptist Convention, and the United Church of Christ are other Protestant denominations, and they have their state offices in the Greater Syracuse area. The dozens of churches in Syracuse include Seventh-Day Adventist, Eastern Orthodox, Jehovah's Witness, Christian Science, Reformed Presbyterian, and Metaphysical Christian. Buddhism: Buddhism is represented by the Zen Center of Syracuse on the Seneca Turnpike; as well as a center on Park Street, on the city's Northside. Hinduism: Hindu houses of worship include the Hindu Mandir of Central New York in Liverpool. Islam: The Islamic Society of Central New York Mosque is located on Comstock Avenue and Muhammad's Study Group on West Kennedy Street. Judaism: Several synagogues are located in the Syracuse metropolitan area, including Beth Shalom-Chevra Chas, Temple Adath Yeshurun, Shaarei Torah Orthodox Congregation of Syracuse, and Temple Concord, considered the ninth-oldest Jewish house of worship in the United States. Sikhism: The gurdwara is at the Sikh Foundation of Syracuse, in Liverpool. Unitarian Universalism: Two Unitarian Universalist societies in Syracuse: May Memorial Unitarian Society and First Unitarian Universalist Society of Syracuse. Formerly a manufacturing center, Syracuse's economy has faced challenges over the past decades as industrial jobs have left the area. The number of local and state government jobs also has been declining for several years. Syracuse's top employers now are primarily in higher education, research, health care and services; some high-tech manufacturing remains. University Hill is Syracuse's fastest-growing neighborhood, fueled by expansions by Syracuse University and Upstate Medical University (a division of the State University of New York), as well as dozens of small medical office complexes. Micron Technology plans to spend up to $100 billion building a mega-complex of computer chip plants in Syracuse's northern suburbs, about a 15-minute drive from downtown Syracuse, in what would be the largest single private investment in New York history. Micron Technology announced it will begin construction in 2024. Micron's Clay mega-complex of memory chip fabs would create up to 9,000 direct jobs and additional 40,000 supply-chain and construction jobs over the next 20 years in Syracuse area. In Phase 1, the company will spend $20 billion to build its first plant, which it estimates will require about 5,000 workers for construction and initially employ about 3,000 manufacturing workers. Micron's Syracuse investment was influenced by passage of the CHIPS and Science Act. Through the CHIPs and Science Act, Congress and the Biden administration established a powerful investment platform with the potential to expand the region's productive capacity for decades. Top employers in the Syracuse region and the size of their workforce include the following: Since 1927 the State Tower Building has been the tallest in Syracuse. In addition to the dominant Destiny USA shopping mall in Syracuse's Lakefront neighborhood, many of the city's more traditional neighborhoods continue to have active business districts: Live jazz music is the centerpiece of two annual outdoor festivals in Syracuse, the Syracuse Jazz Festival, Polish Festival as well as the CNY Jazz Arts Foundation's Jazz in the Square Festival. Performers in the last five years have included Chuck Mangione, Joshua Redman, Smokey Robinson, Branford Marsalis, The Bad Plus, Randy Brecker, Stanley Clarke, Jimmy Heath, Terrence Blanchard, Slide Hampton, Bobby Watson, Dr. John, and Aretha Franklin. The Polish Festival hosted Grammy winners Jimmy Sturr and his Orchestra, Polish music legend Stan Borys and Irena Jarocka, Grammy nominee Lenny Goumulka, LynnMarie, Dennis Polisky & The Maestro's Men, Jerry Darlak and the Buffalo Touch & The John Gora Band. Syracuse was home to the 75-member Syracuse Symphony Orchestra (SSO), founded in 1961. The SSO's former music directors include Daniel Hege, Frederik Prausnitz and Kazuyoshi Akiyama. The orchestra performed over 200 concerts annually for an audience of over 250,000. The SSO filed for Chapter 7 Bankruptcy in 2011 and was replaced by the Syracuse Symphoria in 2013. The Clinton String Quartet has been active for over 15 years and is based in the Syracuse area. All four members were also members of the Syracuse Symphony Orchestra. The Syracuse Friends of Chamber Music for more than a half century have presented a series of concerts by various chamber ensembles. The Society for New Music, founded in 1971, is the oldest new music organization in the state outside of New York City, and the only year-round new music group in upstate New York. The Society commissions at least one new work each year from a regional composer who awards the annual Brian Israel Prize to a promising composer under 30 years of age and produces the weekly "Fresh Ink" radio broadcast for WCNY-FM. The Syracuse Opera Company is a professional company that generally performs three operas each season. Founded in 1963 as the Opera Chorus of the Syracuse Symphony Orchestra, it became independent in 1973. In addition to full performances, it offers several free outdoor concerts each year in Armory Square, Thornden Park, and elsewhere. The company has an annual budget of US$1 million and is the only professional opera company in upstate New York. The Syracuse Shakespeare Festival is a charitable, educational, not-for-profit corporation dedicated to performing the works of William Shakespeare. It was founded in 2002 and is best known for its annual free Shakespeare-in-the-Park program at the Thornden Park Amphitheatre that has attracted more than 12,000 people since its inception. Syracuse Stage presents experimental and creative theater; a number of its productions have been world premieres and have moved to Broadway. The venue was designed by its most famous former artistic director Arthur Storch. Its artistic director is Robert Hupp. The Red House Arts Center, which opened in 2004, is a small theater housed in a converted hotel that offers performances by local, national, and international artists, and hosts regular exhibits in its art gallery, and screenings of independent films. Syracuse is also known for a large contemporary music scene, particularly in the heavy metal, hardcore, ska, and punk rock genres. From 1997 to 2003, Syracuse (or its suburbs) was home to Hellfest, a major hardcore music festival. Onondaga County Public Library (OCPL) operates Syracuse's public libraries. Including the Central Library, ten city libraries, and 21 independent libraries in suburban Onondaga County. A library card from any OCPL library will work at any of the other OCPL libraries. City libraries Suburban libraries The Syracuse City School District consists of 34 schools and 4 alternative education programs. In the 2014–2015 school year, the K-12 enrollment was 20,084. 15% of students were classified as English Language Learners, 20% as students with disabilities, and 77% as economically disadvantaged. The drop-out rate was 6%. Syracuse City School District is collaborating with Say Yes to Education with the goal of every public school student graduating high school with the preparation and support to attain, afford, and complete a college or other postsecondary education. They are also one of the "Big 5," which consists of the five New York State School districts with populations over 125,000. "Big 5" school budgets are approved by annually by the Board of Education and city government as opposed to voters in an annual vote. One of Syracuse's major research universities is Syracuse University, located on University Hill. It had an enrollment of 22,484 for the 2017–2018 academic year. Immediately adjacent to Syracuse University are two doctoral-degree granting universities, the SUNY Upstate Medical University and SUNY College of Environmental Science and Forestry. Both institutions have long-standing ties to Syracuse University. Upstate Medical University is also one of Syracuse's major research universities and is one of only about 125 academic medical centers in the country. The medical university directly generates 10,959 jobs, making it Central New York's largest employer. In addition, the Norton College of Medicine at SUNY Upstate is the only medical school in the Central New York region providing state of the art education to over 700+ students. Also serving Syracuse are Le Moyne College on the city's eastern border, and Onondaga Community College, which has its main campus in the adjacent Town of Onondaga and has two smaller campuses, downtown and in Liverpool. A branch of SUNY's Empire State College is in downtown Syracuse, along with a campus of the nationwide Bryant & Stratton College. There are also the Pomeroy College of Nursing at Crouse Hospital and St. Joseph's College of Nursing. Other colleges and universities in the area include Cornell University and Ithaca College in Ithaca, Hamilton College in Clinton, Oswego State College in Oswego, SUNY Cortland in Cortland, Morrisville State College in Morrisville, Colgate University in Hamilton, Cazenovia College in Cazenovia (closed 2023), Wells College in Aurora, and both Utica College and SUNY Institute of Technology in Utica. The City of Syracuse maintains over 170 parks, fields, and recreation areas, totaling over 1,000 acres (4.0 km). Burnet Park includes the first public golf course in the United States (1901) and Rosamond Gifford Zoo. Other major parks include Thornden Park, Schiller Park, Sunnycrest Park, Onondaga Park and Kirk Park. There are 12 public pools, two public ice rinks (Sunnycrest and Meachem), and two public nine-hole golf courses (Burnet and Sunycrest Parks) in the city. Onondaga Park, located in the historic Strathmore neighborhood, features Hiawatha Lake, and a beautiful gazebo, often used for prom photos and wedding shoots. Right outside the city proper, along the east side and north end of Onondaga Lake, is Onondaga Lake Park. The adjacent Onondaga Lake Parkway is closed to vehicular traffic several hours on Sundays during the summer months, so it can be used for walking, running, biking, and rollerblading. During the holiday season, the park hosts Lights on the Lake, a two-mile (3.2 km) drive-through light show. Syracuse University sports are by far the most attended sporting events in the Syracuse area. Basketball games often draw over 30,000 fans, and football games over 40,000. The university has bred dozens of famous professional players since starting an athletics program in the late nineteenth century, including all-time greats Ernie Davis, Jim Brown, Larry Csonka and Dave Bing. Both teams play in the JMA Wireless Dome. In addition to many former professional minor league teams, Syracuse was previously the home of several top-level pro teams, most notably the Syracuse Nationals who played a total of 17 seasons between the NBL and NBA, and won the 1955 NBA Finals before moving to Philadelphia and becoming the Philadelphia 76ers. Syracuse was also the home of two different Major League Baseball teams: the Syracuse Stars of the National League in 1879, which did not finish their first season; and the Syracuse Stars of the American Association in 1890. The city is headed by an elected mayor who is limited to two four-year terms. Syracuse has a Strong mayor-council form of government. On November 7, 2017, Ben Walsh was elected mayor. He began in January 2018 as the first independent mayor of Syracuse in over 100 years. The last independent mayor of Syracuse was Louis Will, who was elected in 1913. The previous mayor was former Common Councilor at Large Stephanie Miner, who was elected on November 3, 2009; she was the first female mayor of Syracuse. Miner was preceded by former Syracuse Common Council President Matthew Driscoll, who first assumed the position in 2001 after the former mayor, Roy Bernardi, resigned upon his appointment by President George W. Bush to a position in the Department of Housing and Urban Development. After serving the remaining term, Driscoll was re-elected that year, and again in 2005. The legislative branch of Syracuse is the Syracuse Common Council. It consists of a president and nine members: The Onondaga County Supreme and County Court is the trial court of general jurisdiction for Syracuse. It is also the administrative court for the Fifth District of the New York State Unified Court System. Judges for these courts are elected at-large. The U.S. District Court for the Northern District of New York also holds court in downtown Syracuse at the James Hanley Federal Building. syracuse.com is the most popular local media site in Central New York, according to ComScore, with an average of 3.98 Million unique users. Advance Media NY is the home of syracuse.com, who also produces the local newspaper, The Post-Standard. The two media units combined reach 422,000 in the Syracuse DMA, according to Nielsen, 2022. Advance Media NY is a digital media and marketing agency, who helps businesses tell their stories in print, digital and visuals. It has a full service internal ad agency. syracuse.com features the latest in local, sports news, breaking news and entertainment, with daily CNY weather updates. Syracuse has one major daily morning newspaper, The Post-Standard. Until 2001, Syracuse also had an evening paper, The Herald-Journal. It focuses on local news throughout Central New York, and has a reporter in the Washington, DC. Before the merger with the evening paper, the Post-Standard was named among the "10 best newspapers in America with a circulation of under 100,000" by Al Neuharth of USA Today (run by a competing organization). Since the merger, circulation has increased to over 120,000. Even outside of its four-county delivery area, the paper is available in many convenience stores and supermarkets from the Canada–US border to the New York–Pennsylvania border. The newspaper partly caters to this audience as well, covering many stories from the Ithaca, Utica, and Watertown areas. Since opening a new printing press in 2002, the paper calls itself "America's Most Colorful Newspaper," as almost every page contains color. The Daily Orange, the newspaper of Syracuse University and SUNY ESF students, is read by over 20,000 people daily, and is widely distributed in the University Hill neighborhood and Armory Square. The Dolphin, the weekly student newspaper of Le Moyne College is also available, read mainly by Le Moyne students. There are other popular free newspapers, including Eagle Newspaper's downtown edition, the City Eagle, and Table Hopping, which focuses on the restaurant and entertainment scene. Additionally, there's a weekly newspaper, CNY Vision, that publishes news and information focusing on the local African American community. There is also a Hispanic-based monthly publication, called the CNY Latino newspaper, published by the CNY Latino Media Consortium (www.cnylatino.com) in BOTH paper format and online at www.cnylatinonewspaper.com, covering not only the city of Syracuse, also all the cities and towns between Rochester & Albany AND Watertown & Binghamton. The Syracuse area is covered in a regional lifestyle publication called "The Good Life, Central New York Magazine," mostly known as "Central New York Magazine" (www.readcnymagazine.com). The magazine is bi-monthly (six issues per year) and offers print + digital and digital only subscriptions; it is also sold at local independent retailers, Wegmans, Tops Friendly Markets, and Barnes & Noble. The magazine covers the greater Syracuse and Central New York area, including Onondaga, Oswego, Madison, Oneida Cortland, Tompkins and Cayuga counties. Central New York Magazine premiered in May 2006 and tells "positive and uniquely CNY stories." Coverage areas include local shops and small businesses, regional travel destinations, food and drink, home decor, attractions and things to do, artisans, changemakers and area trends. According to Nielsen Media Research, Syracuse is the fifth largest television market in New York State and the 87th largest in the United States (as of the 2020–2021 TV season). Six major full-power stations serve the city: WSTM-TV 3 (NBC), WTVH 5 (CBS), WSYR-TV 9 (ABC), WCNY-TV 24/cable 11 (PBS), WSPX-TV 56/cable 4 (Ion), and WSYT 68/cable 8 (Fox). WSTM-TV also operates the area's CW affiliate on its DT2 subchannel and cable channel 6, and WSYT carries the MyNetworkTV affiliation on channel 43 and cable channel 7; both stations were previously separately-licensed stations before having their licenses returned to the FCC. Additionally, networks such as Cornerstone Television channel 11 & 22, Univision, and MTV2 are broadcast by low-power television stations. Syracuse University's student-run TV station is CitrusTV. CitrusTV programming is broadcast on the university campus on the Orange Television Network. Syracuse's cable television provider is Charter Spectrum (Charter Communications acquired Time Warner Cable in 2016), which, as a part of its regular and digital offerings, provides a 24-hour local news channel (Spectrum News Central New York), public access channel, and an additional PBS channel. Several suburbs also have access to Verizon Fios for cable television. Dish Network and DirecTV also provide local satellite television subscribers with local broadcast stations. Syracuse is served by the Central New York Regional Transportation Authority, or Centro. Centro operates bus service in Syracuse and its suburbs, as well as to outlying metropolitan area cities such as Auburn, Fulton, and Oswego. Proposed public transit projects In 2005, local millionaire Tom McDonald proposed an aerial tramway system, called Salt City Aerial Transit (S.C.A.T.), to link the university to the transportation center. The first segment from Syracuse University to downtown was estimated to cost $5 million, which McDonald planned to raise himself. Due to perceived low operating costs, the system was envisioned as running continuously. Syracuse (station stop code SYR) is served by Amtrak's Empire Service, Lake Shore Limited, and Maple Leaf lines. Amtrak's station is part of the William F. Walsh Regional Transportation Center. The Empire Service runs twice daily in each direction between Niagara Falls, NY and New York Penn Station, with major stops in Buffalo, Rochester, Syracuse, Utica, and Albany along the way. The Maple Leaf runs once daily in each direction, and follows the same route as the Empire Service, however instead of terminating in Niagara Falls, it continues on to Toronto. Empire Service and Maple Leaf trains stop at the seasonal New York State Fair – NYF station during the New York State Fair's annual run each August. The NYF Station is located along the southern part of the fairgrounds, near the historic train car display of the Central New York Chapter, of the National Railway Historical Society. The Lake Shore Limited runs once daily in each direction between Chicago and Boston or New York City (via two sections splitting Albany-Rensselaer). It follows the same route as the Empire Service and Maple Leaf between New York City and Buffalo-Depew, where it diverges and continues on through Cleveland and Toledo to Chicago. A regional commuter rail service, OnTrack, was active from 1994 until it was discontinued in 2007 due to low ridership. Its sole route connected the Carousel Center to southern Syracuse, often extending to Jamesville in the summer. Greyhound Lines, Megabus, OurBus and Trailways provide long-distance bus service to destinations including New York City, Boston, Buffalo, Albany, and Toronto. Greyhound, Megabus, and Trailways use the William F. Walsh Regional Transportation Center in the northern area of the city, while OurBus stops near the campus of Syracuse University. Syracuse is served by the Syracuse Hancock International Airport in nearby Salina, near Mattydale. The airport is named after Clarence E. Hancock, a former US Congressman representing Syracuse. The airport is served by 8 major airlines, which provide non-stop flights to important airline hubs and business centers such as Atlanta, Boston, Charlotte, Chicago, Detroit, Denver, Ft. Lauderdale, New York City, Orlando, Philadelphia, Tampa, Washington, DC, as well as connecting service to 147 foreign cities in 87 countries. Cargo carriers FedEx and UPS also serve the airport. New York City can be reached in under an hour flight. The City of Syracuse owns the airport and property, while a public for-benefit corporation runs the airport, the Syracuse Regional Airport Authority. The airport is protected by the 174th Attack Wing's Fire Department, and patrolled by Syracuse Airport Police Department officers. Community Grid In late May, 2023, The Community Grid project was officially approved for construction in the Syracuse. The highly controversial plan consists of removing the I-81 viaduct that runs through the downtown of the city, and replacing it with the Boulevard style Business Loop-81. The heavily congested Interstate-81 will be re-routed around the city onto the already existing Interstate-481. This project is estimated to cost around 2.25 billion dollars, and will be completed over a multi-year process. Construction has already begun in portions of North Syracuse at the I-81 and I-481 interchange. Two US Highways run through the Syracuse area: New York State Route Expressways: New York State Routes Public services such as garbage pickup, street plowing, sewage, and street and traffic maintenance are provided by the Department of Public Works (DPW). The Syracuse water system was one of the few water systems built and operated before federal funding. The water system was constructed mainly to support the industries around Syracuse, New York. Construction of Syracuse's water system began in 1868. The water is brought in on a gravity fed system from Skaneateles Lake, through an unfiltered system, and carried into the city. It is noted for having some of the best drinking water in the nation, due to the quality of the lake. In 2015, the city experienced an average of at least one water main break per day. Between 2005 and 2015, the city suffered 2,000 water main breaks. Mayor Stephanie Miner estimated of the cost to fix the city's water infrastructure at $1 billion over a 10–15-year period. On February 25, 2015, Miner testified before a joint hearing of the state Assembly Ways and Means Committee and state Senate Finance Committee. Miner testified that the 2014 polar vortex contributed to the increase in Syracuse's water main break. On March 3, the 100th water main break in Syracuse in 2015 occurred on James Street. Early in 2015, Miner lobbied the state for funding to fix the city's aging water system. New York Governor Andrew Cuomo declined to help, stating that the city should improve its economy and increase tax revenues, which would enable the city to fund their own water pipe repairs. The Syracuse Police Department (SPD) is the principal law enforcement agency of the city of Syracuse, New York. For 2017–18, the police department budget was $48.5 million (~$57.2 million in 2022). Effective April 22, 2022, longtime Deputy Chief Joe Cecile is Chief of the SPD, following his predecessor Kenton Buckner's retirement. Police headquarters is in the John C. Dillon Public Safety Building at 511 South State Street. The SPD is divided into three patrol zones North (Lakefront, Northside, Eastwood, Tip Hill), South West (Strathmore, Valley, Southside, Near-Westside), and Southeast (University Area, Downtown, Meadowbrook, Eastside). In 2019, a jury awarded Elijah Johnson $35,000 (~$40,061 in 2022) after he was beaten with unreasonable force by three police officers while being arrested. In addition, the city was forced to pay attorneys fees, at a total cost to taxpayers of $213,000. Established in 2011, SPD operates a network of 521 surveillance cameras called the Criminal Observation and Protection System (COPS). Between 2011 and 2014 more than 40 utility pole mounted cameras were installed, mainly in the Southwest and Northeast neighborhoods. The cameras were funded by federal, state, and private grants. In Summer 2014, 10 cameras were approved for installation in Downtown Syracuse, the first area not targeted because of high levels of violent crime. Live monitoring of Clinton Square for suspicious people during events and festivals was planned, although police agreed to a prohibition on the use of cameras to monitor protests. Twenty-five additional cameras were planned to be installed in 2016. In spring 2017, the surveillance system was augmented with the installation of ShotSpotter gunshot detection sensors. Syracuse Mayor Stephanie Miner cited increasing public acceptance of police cameras and lower technology costs as factors in the decision. The Syracuse Fire Department (SFD) has the responsibility of protecting the City of Syracuse from fires and other dangers. The department provides multiple services in addition to fire related calls: multi-county regional HAZ-MAT response,first response to medical and trauma calls, unmanned aerial vehicle (drone) capabilities,and teams experienced in high-angle rope,swift water,and confined space rescue operations. The Chief of Fire is Michael J. Monds. SFD headquarters is in the John C. Dillon Public Safety Building at 511 South State Street. The department has a Class 1 rating from the Insurance Services Office, which is the best rating obtainable. This rating has a direct effect on the fire insurance of properties within the city. The SFD operates out of 11 fire stations, organized into three districts (akin to battalions), located throughout the city. The SFD currently maintains nine engine companies (operating with nine corresponding "mini" units),five truck companies,one heavy rescue company,a manpower-squad company,and several special and support units. The department also provides Aircraft Rescue and Firefighting ARFF coverage and specialized fire,rescue,medical and hazardous materials coverage to the Syracuse Clarance E. Hancock International Airport (station 4). Syracuse's sister cities are:
A '''soap opera''' is a [[television]] or [[radio]] program that [[broadcast|airs]] in [[episode|episodic]] series. This means that each episode continues telling a story, which, in turn, tells more of the last episode's story. A single story on a soap opera can be told for weeks, months, or sometimes even years. In [[United States|America]], each soap opera airs every day, Monday through Friday, in the afternoon. Because they air at this time, they are sometimes called '''daytime serials'''. In the [[United Kingdom|U.K.]] and other countries, the soap operas air in the evening, twice or three times a week. This is a list of soap operas that air in America, in their current order of popularity: #''[[The Young and the Restless]]'' #''[[The Bold and the Beautiful]]'' #''[[General Hospital]]'' #''[[Supernatural (TV series)|Supernatural]]'' #''[[Lost (TV series)|Lost]]'' #''[[Days of Our Lives]]'' #''[[All My Children]]'' #''[[As the World Turns]]'' #''[[One Life to Live]]'' #''[[Guiding Light]]'' #''[[Passions]]'' There are still two soap operas that air in [[Australia]], [[Network Ten]]'s ''[[Neighbours]]'' and the [[Seven Network]]'s ''[[Home and Away]]''. Strangely, ''Neighbours'' is more popular in England than it is in Australia. In the U.K., the most popular soap operas are ''[[Coronation Street]]'' and ''[[EastEnders]]''. ''Coronation Street'' is about people who live in [[Manchester]], in the northern part of England. ''EastEnders'' is about people who live in an imaginary place called Albert Square in the [[East London]]. Another popular soap opera is ''[[Emmerdale]]'', and it is about people who live in a small village in [[Yorkshire]], also in the northern part of England. In the 1970s ''[[Crossroads (British TV series)|Crossroads]]'' became the most watched soap in England. It was shown from 1964 to 1988.<ref>{{cite web |url=http://www.screenonline.org.uk/tv/id/540415/index.html |title=Crossroads (1964-88, 2001-03) |website=screenonline.org.uk |access-date=21 January 2020}}</ref> In [[Latin America]] (including [[Brazil]]), [[Spain]], [[Portugal]] and [[Portuguese language|Portuguese]]-speaking countries in [[Africa]] (like [[Angola]]), soap operas are called ''[[telenovela]]s'', and usually run for a limited time. ==Related pages== *[[Dorama]] *[[Serial (radio and television)|Serial]] *[[Radio drama]] *[[Playwright]] *[[Theatre]] == References == {{reflist}} [[Category:Soap operas| ]] [[Category:Television genres]]
A soap opera, daytime drama, or soap for short, is typically a long-running radio or television serial, frequently characterized by melodrama, ensemble casts, and sentimentality. The term "soap opera" originated from radio dramas originally being sponsored by soap manufacturers. The term was preceded by "horse opera", a derogatory term for low-budget Westerns. BBC Radio's The Archers, first broadcast in 1950, is the world's longest-running radio soap opera. The longest-running current television soap is Coronation Street, which was first broadcast on ITV in 1960, with the record for the longest-running soap opera in history being held by Guiding Light, which began on radio in 1937, transitioned to television in 1952, and ended in 2009. According to Albert Moran, one of the defining features that make a television program a soap opera is "that form of television that works with a continuous open narrative. Each episode ends with a promise that the storyline is to be continued in another episode". In 2012, Los Angeles Times columnist Robert Lloyd wrote of daily dramas: Although melodramatically eventful, soap operas such as this also have a luxury of space that makes them seem more naturalistic; indeed, the economics of the form demand long scenes, and conversations that a 22-episodes-per-season weekly series might dispense with in half a dozen lines of dialogue may be drawn out, as here, for pages. You spend more time even with the minor characters; the apparent villains grow less apparently villainous. Soap opera storylines run concurrently, intersect and lead into further developments. An individual episode of a soap opera will generally switch between several concurrent narrative threads that may at times interconnect and affect one another or may run entirely independent to each other. Episodes may feature some of the show's current storylines, but not always all of them. Especially in daytime serials and those that are broadcast each weekday, there is some rotation of both storyline and actors so any given storyline or actor will appear in some but usually not all of a week's worth of episodes. Soap operas rarely conclude all the current storylines at the same time. When one storyline ends, there are several other story threads at differing stages of development. Soap opera episodes typically end on some sort of cliffhanger, and the season finale (if a soap incorporates a break between seasons) ends in the same way, only to be resolved when the show returns for the start of a new yearly broadcast. Evening soap operas and those that air at a rate of one episode per week are more likely to feature the entire cast in each episode and present all current storylines. Evening soap operas and serials that run for only part of the year tend to bring things to a dramatic end-of-season cliffhanger. In 1976, Time magazine described American daytime television as "TV's richest market", noting the loyalty of the soap opera fan base and the expansion of several half-hour series into hour-long broadcasts in order to maximize ad revenues. The article explained that at that time, many prime time series lost money, while daytime serials earned profits several times more than their production costs. The issue's cover notably featured its first daytime soap stars, Bill Hayes and Susan Seaforth Hayes of Days of Our Lives, a married couple whose onscreen and real-life romance was widely covered by both the soap opera magazines and the mainstream press at large. The first program generally considered to be a "soap opera" or daytime serial by scholars of the genre is Painted Dreams, which premiered on WGN radio Chicago, on October 20, 1930. It was regularly broadcast in a daytime time slot, where most listeners would be housewives; thus, the shows were aimed at – and consumed by – a predominantly female audience. Clara, Lu, 'n Em would become the first network radio serial of the type when it aired on the NBC Blue Network at 10:30 p.m. Eastern Time on January 27, 1931. Although it did not make the move until February 15, 1932, Clara, Lu 'n Em would become the first network serial of the type to move to a weekday daily timeslot, and so also became the first network daytime serial. The main characteristics that define soap operas are "an emphasis on family life, personal relationships, sexual dramas, emotional and moral conflicts; some coverage of topical issues; set in familiar domestic interiors with only occasional excursions into new locations". Fitting in with these characteristics, most soap operas follow the lives of a group of characters who live or work in a particular place, or focus on a large extended family. The storylines follow the day-to-day activities and personal relationships of these characters. "Soap narratives, like those of film melodramas, are marked by what Steve Neale has described as 'chance happenings, coincidences, missed meetings, sudden conversions, last-minute rescues and revelations, deus ex machina endings.'" These elements may be found across the gamut of soap operas, from EastEnders to Dallas. In many soap operas, in particular daytime serials in the US, the characters are frequently attractive, seductive, glamorous and wealthy. Soap operas from the United Kingdom and Australia tend to focus on more everyday characters and situations, and are frequently set in working-class environments. Many of the soaps produced in those two countries explore social realist storylines such as family discord, marriage breakdown or financial problems. Both UK and Australian soap operas feature comedic elements, often affectionate comic stereotypes such as the gossip or the grumpy old man, presented as a comic foil to the emotional turmoil that surrounds them. This diverges from US soap operas where such comedy is rare. UK soap operas frequently make a claim to presenting "reality" or purport to have a "realistic" style. UK soap operas also frequently foreground their geographic location as a key defining feature of the show while depicting and capitalising on the exotic appeal of the stereotypes connected to the location. As examples, EastEnders focuses on the tough and grim life in the East End of London; Coronation Street and its characters exhibit the stereotypical characteristic of "northern straight talking". If we want to blend an actor back into a show, there's always a way. You can generally find a way to twist and manipulate something. You rarely see a dead body, but hey, even if you do, he or she can always come back to play the evil identical twin. Marlena Laird in 1992, during her time as a line producer and director for General Hospital. Romance, secret relationships, extramarital affairs, and genuine hate have been the basis for many soap opera storylines. In US daytime serials, the most popular soap opera characters, and the most popular storylines, often involved a romance of the sort presented in paperback romance novels. Soap opera storylines weave intricate, convoluted and sometimes confusing tales of characters who have affairs, meet mysterious strangers and fall in love, and who commit adultery, all of which keeps audiences hooked on the unfolding story. Crimes such as kidnapping, assault (sometimes sexual), and even murder may go unpunished if the perpetrator is to be retained in the ongoing story. Australian and UK soap operas also feature a significant proportion of romance storylines. In Russia, most popular serials explore the "romantic quality" of criminal and/or oligarch life. In soap opera storylines, previously unknown children, siblings and twins (including the evil variety) of established characters often emerge to upset and reinvigorate the set of relationships examined by the series. Unexpected calamities disrupt weddings, childbirths, and other major life events with unusual frequency. As in comic books – another popular form of linear storytelling pioneered in the US during the 20th century – a character's death is not guaranteed to be permanent. On The Bold and the Beautiful, Taylor Hayes (Hunter Tylo) was shown to flatline and have a funeral. Once Tylo reprised her character in 2005, a retcon explained that Taylor had actually gone into a coma. Stunts and complex physical action are largely absent, especially from daytime serials. Such story events often take place off-screen and are referred to in dialogue instead of being shown. This is because stunts or action scenes are difficult to adequately depict without complex movements, multiple takes, and post-production editing. When episodes were broadcast live, post-production work was impossible. Though all serials have long switched to being taped, extensive post-production work and multiple takes, while possible, are not feasible due to the tight taping schedules and low budgets. The first daytime TV soap opera in the United States was These Are My Children in 1949, though earlier melodramas had aired in the evenings as once-a-week programs. Soap operas quickly became a fixture of American daytime television in the early 1950s, joined by game shows, sitcom reruns and talk shows. In 1988, H. Wesley Kenney, who at the time served as the executive producer of General Hospital, said to The New York Times: I think people like stories that continue so they can relate to these people. They become like a family, and the viewer becomes emotionally involved. There seem to be two attitudes by viewers. One, that the stories are similar to what happened to them in real life, or two, thank goodness that isn't me. Many long-running US soap operas established particular environments for their stories. The Doctors and General Hospital, in the beginning, told stories almost exclusively from inside the confines of a hospital. As the World Turns dealt heavily with Chris Hughes' law practice and the travails of his wife Nancy who, tired of being "the loyal housewife" in the 1970s, became one of the first older women on the American serials to enter the workforce. Guiding Light dealt with Bert Bauer (Charita Bauer) and her alcoholic husband Bill, and their endless marital troubles. When Bert's status shifted to caring mother and town matriarch, her children's marital troubles were showcased. Search for Tomorrow mostly told its story through the eyes of Joanne Gardner (Mary Stuart). Even when stories revolved around other characters, Joanne was frequently a key player in their storylines. Days of Our Lives initially focused on Dr. Tom Horton and his steadfast wife Alice. The show later branched out to focus more on their five children. The Edge of Night featured as its central character Mike Karr, a police detective (later an attorney), and largely dealt with organized crime. The Young and the Restless first focused on two families, the prosperous Brooks family with four daughters, and the working-class Foster family of a single working mother with three children. Its storylines explored realistic problems including cancer, mental illness, poverty, and infidelity. In contrast, Dark Shadows (1966–1971), Port Charles (1997–2003) and Passions (1999–2008) featured supernatural characters and dealt with fantasy and horror storylines. Their characters included vampires, witches, ghosts, goblins, and angels. The American soap opera Guiding Light (originally titled The Guiding Light until 1975) started as a radio drama in January 1937 and subsequently transferred to television in June 1952. With the exception of several years in the late 1940s, during which creator Irna Phillips was involved in a dispute with Procter & Gamble, Guiding Light was heard or seen nearly every weekday from 1937 to 2009, making it the longest story ever told in a broadcast medium. Originally serials were broadcast as 15-minute installments each weekday in daytime slots. In 1956, As the World Turns and The Edge of Night, both produced by Procter & Gamble Productions, debuted as the first half-hour soap operas on the CBS television network. All soap operas broadcast half-hour episodes by the end of the 1960s. With increased popularity in the 1970s, most soap operas had expanded to an hour in length by the end of the decade (Another World even expanded to 90 minutes for a short time from 1979 to 1980). More than half of the serials had expanded to one-hour episodes by 1980. As of 2012, three of the four U.S. serials air one-hour episodes each weekday; only The Bold and the Beautiful airs 30-minute episodes. Soap operas were originally broadcast live from the studio, creating what many at the time regarded as a feeling similar to that of a stage play. As nearly all soap operas were originated at that time from New York City, a number of soap actors were also accomplished stage actors who performed live theater during breaks from their soap roles. In the 1960s and 1970s, new serials such as General Hospital, Days of our Lives, and The Young and the Restless were produced in Los Angeles. Their success made the West Coast a viable alternative to New York-produced soap operas, which were becoming more costly to perform. By the early 1970s, nearly all soap operas had transitioned to being taped. As the World Turns and The Edge of Night were the last to make the switch, in 1975. Port Charles used the practice of running 13-week "story arcs," in which the main events of the arc are played out and wrapped up over the 13 weeks, although some storylines did continue over more than one arc. According to the 2006 Preview issue of Soap Opera Digest, it was briefly discussed that all ABC shows might do telenovela arcs, but this was rejected. Though U.S. daytime soap operas are not generally rerun by their networks, occasionally they are rebroadcast elsewhere; CBS and ABC have made exceptions to this, airing older episodes (either those aired earlier in the current season or those aired years prior) on major holidays when special event programming is not scheduled or because of last-minute deferrals of scheduled episodes to the following day because of breaking news coverage. (Temporary production stoppages caused by the COVID-19 pandemic similarly resulted in CBS and ABC airing older reruns of The Young and the Restless, The Bold and the Beautiful and General Hospital during the Spring and Summer of 2020 in order to ration first-run episodes and, eventually, to fill airtime after the programs ran out of new episodes to broadcast; Days of Our Lives, which produces its episodes roughly eight months ahead of their initial broadcast, did not resort to airing older episodes during this time as it had a larger first-run episode backlog.) Early episodes of Dark Shadows were rerun on PBS member stations in the early 1970s after the show's cancellation, and the entire series (except for a single missing episode) was rerun on the Sci-Fi Channel in the 1990s. After The Edge of Night's 1984 cancellation, reruns of the show's final five years were shown late nights on USA Network from 1985 to 1989. On January 20, 2000, a digital cable and satellite network dedicated to the genre, Soapnet, began re-airing soaps that originally aired on ABC, NBC and CBS. Newer broadcast networks since the late 1980s, such as Fox and cable television networks, have largely eschewed soap operas in their daytime schedules, instead running syndicated programming and reruns. No cable television outlet has produced its own daytime serial, although DirecTV's The 101 Network took over existing serial Passions, continuing production for one season; while TBS and CBN Cable Network respectively aired their own soap operas, The Catlins (a primetime soap that utilized the daily episode format of its daytime counterparts) and Another Life (a soap that combined standard serial drama with religious overtones), during the 1980s. Fox, the fourth "major network", carried a short-lived daytime soap Tribes in 1990. Yet, other than this and a couple of pilot attempts, Fox mainly stayed away from daytime soaps, and has not attempted them since their ascension to major-network status in 1994 (it did later attempt a series of daily prime time soaps, which aired on newly created sister network MyNetworkTV, but the experiment was largely a failure). Due to the masses of episodes produced for a series, release of soap operas to DVD (a popular venue for distribution of current and vintage television series) is considered impractical. With the exception of occasional specials, daytime soap operas are notable by their absence from DVD release schedules (an exception being the supernatural soap opera, Dark Shadows, which did receive an essentially complete release on both VHS and DVD; the single lost episode #1219 is reconstructed by means of an off-the-air audio recording, still images, and recap material from adjacent episodes). Soap opera performers in the United States are typically divided into two main groups: primary characters (sometimes referred to as "contract players" – as their portrayers signed contracts of employment – or leading characters) and secondary characters (sometimes referred to as recurring characters). These two groups of characters make up the vast majority of the people who appear on any given soap. There are also characters who appear only for a short time as dictated by a specific storyline, and even characters who may only get a first name and no fleshed-out character history with little dialogue (these are sometimes referred to as "under-5s" since they receive under five lines of dialogue in each episode). Due to the longevity of these shows, it is not uncommon for a single character to be played by multiple actors. The key character of Mike Karr on The Edge of Night was played by three actors. Conversely, several actors have remained playing the same character for many years, or decades even. Helen Wagner played Hughes family matriarch Nancy Hughes on American soap As the World Turns from its April 2, 1956, debut through her death in May 2010. She is listed in the Guinness Book of World Records as the actor with the longest uninterrupted performance in a single role. A number of performers played roles for 20 years or longer, occasionally on more than one show. Rachel Ames played Audrey Hardy on both General Hospital and Port Charles from 1964 until 2007, and returned in 2009. Susan Lucci played Erica Kane in All My Children from the show's debut in January 1970 until it ended its network television run on ABC on September 23, 2011. Erika Slezak played Victoria Lord #3 on One Life to Live from 1971 until the show ended its network television run on ABC on January 13, 2012, and resumed the role in its short-lived online revival on April 29, 2013. Other actors have played several characters on different shows. Millette Alexander, Bernard Barrow, Doris Belack, David Canary, Judith Chapman, Keith Charles, Jordan Charney, Joan Copeland, Nicolas Coster, Jacqueline Courtney, Augusta Dabney, Louis Edmonds, Don Hastings, Larry Haines, Vincent Irizarry, Lenore Kasdorf, Teri Keane, Lois Kibbee, John Loprieno, Lori March, Maeve McGuire, Robert Mili, James Mitchell, Lee Patterson, Christopher Pennock, Antony Ponzini, William Prince, Rosemary Prinz, Louise Shaffer, Mary Stuart, Richard Thomas, Diana van der Vlis, Mary K. Wells, Lesley Woods and Michael Zaslow, among many others, have all played multiple soap roles. For several decades, most daytime soap operas concentrated on family and marital discord, legal drama and romance. The action rarely left interior settings, and many shows were set in fictional, medium-sized Midwestern towns. Social issue storylines were typically verboten when soaps were starting, due to heavy network-imposed censorship at that time, but writer and producer Agnes Nixon introduced these storylines slowly but surely, first in 1962 when the matriarch of The Guiding Light, Bert Bauer, developed uterine cancer (as the actress, Charita Bauer, had been diagnosed with the same illness in real life). The storyline encouraged many women to get pap smears and the CBS mailroom in New York City received a then-record amount of fan mail wishing Bauer (both Bert and Charita) well. Nixon would go on to tell many socially relevant storylines on her soaps One Life to Live and All My Children in the late 1960s and into the 1970s. Exterior shots were slowly incorporated into the series The Edge of Night and Dark Shadows. Unlike many earlier serials that were set in fictional towns, The Best of Everything and Ryan's Hope were set in a real-world location, New York City. The first exotic location shoot was made by All My Children, to St. Croix in 1978. Many other soap operas planned lavish storylines after the success of the All My Children shoot. Soap operas Another World and Guiding Light both went to St. Croix in 1980, the former show culminating a long-running storyline between popular characters Mac, Rachel and Janice, and the latter to serve as an exotic setting for Alan Spaulding and Rita Bauer's torrid affair. Search for Tomorrow taped for two weeks in Hong Kong in 1981. Later that year, some of the cast and crew ventured to Jamaica to tape a love consummation storyline between the characters of Garth and Kathy. During the 1980s, perhaps as a reaction to the evening drama series that were gaining high ratings, daytime serials began to incorporate action and adventure storylines, more big-business intrigue, and an increased emphasis on youthful romance. One of the most popular couples was Luke Spencer and Laura Webber on General Hospital. Luke and Laura helped to attract both male and female fans. Even actress Elizabeth Taylor was a fan and at her own request was given a guest role in Luke and Laura's wedding episode. Luke and Laura's popularity led to other soap producers striving to reproduce this success by attempting to create supercouples of their own. With increasingly bizarre action storylines coming into vogue, Luke and Laura saved the world from being frozen, brought a mobster down by finding his black book in a left-handed boy statue, and helped a princess find her Aztec treasure in Mexico. Other soap operas attempted similar adventure storylines, often featuring footage shot on location – frequently in exotic locales. During the 1990s, the mob, action and adventure stories fell out of favor with producers, due to generally declining ratings for daytime soap operas at the time. With the resultant budget cuts, soap operas were no longer able to go on expensive location shoots overseas as they were able to do in the 1980s. During that decade, soap operas increasingly focused on younger characters and social issues, such as Erica Kane's drug addiction on All My Children, the re-emergence of Viki Lord's dissociative identity disorder on One Life to Live, and Stuart Chandler dealing with his wife Cindy dying of AIDS on All My Children. Other social issues included cancer, rape, abortion and racism. Several shows during the 1990s and 2000s incorporated supernatural and science fiction elements into their storylines in an attempt to boost their ratings. One of the main characters on the earlier soap opera Dark Shadows was Barnabas Collins, a vampire, and One Life to Live featured an angel named Virgil. Both shows featured characters who traveled to and from the past. In 1995, Days of our Lives featured a storyline in which fan favorite character Marlena Evans was possessed by the devil, and in 1998, Guiding Light featured a cloning storyline involving legacy character Reva Shayne. Modern U.S. daytime soap operas largely stay true to the original soap opera format. The duration and format of storylines and the visual grammar employed by U.S. daytime serials set them apart from soap operas in other countries and from evening soap operas. Stylistically, UK and Australian soap operas, which are usually produced for early evening timeslots, fall somewhere in between U.S. daytime and evening soap operas. Similar to U.S. daytime soap operas, UK and Australian serials are shot on videotape, and the cast and storylines are rotated across the week's episodes so that each cast member will appear in some but not all episodes. UK and Australian soap operas move through storylines at a faster rate than daytime serials, making them closer to U.S. evening soap operas in this regard. American daytime soap operas feature stylistic elements that set them apart from other shows: Soap opera ratings have significantly fallen in the U.S. since the 2000s. No new major daytime soap opera has been created since Passions in 1999, while many have been cancelled. As of September 2022, only four daytime soap operas – General Hospital, Days of Our Lives, The Young and the Restless and The Bold and the Beautiful – are still in production with three airing on two broadcast networks and one on streaming, down from a total of 12 soaps broadcast during the 1990–91 season and a high of 19 in the 1969–70 season. This marks the first time since 1953 that there have been only four soap operas airing on broadcast television. The Young and the Restless, the highest-rated soap opera from 1988 to the present, had fewer than 5 million daily viewers as of February 2012, a number exceeded by several non-scripted programs such as Judge Judy. Circulations of soap opera magazines have decreased and some have even ceased publication. Soapnet, which largely aired soap opera reruns, began to be phased out in 2012 and fully ceased operations the following year. The Daytime Emmy Awards, which honor soap operas and other daytime shows, moved from prime time network television to smaller cable channels in 2012, then failed to get any TV broadcast at all in 2014, 2016, and 2017. Several of the U.S.'s most established soaps ended between 2009 and 2012. The longest-running drama in television and radio history, Guiding Light, barely reached 2.1 million daily viewers in 2009 and ended on September 18 of that year, after a 72-year run (including radio). As the World Turns aired its final episode on September 17, 2010, after a 54-year run. As the World Turns was the last of 20 soap operas produced by Procter & Gamble, the soap and consumer goods company from which the genre got its name. As the World Turns and Guiding Light were also among the last of the soaps that originated from New York City. All My Children, another New York–based soap, moved its production out to Los Angeles in an effort to reduce costs and raise sagging ratings; however, both it and One Life to Live, each with a 40-year-plus run, were cancelled in 2011. All My Children aired its network finale in September 2011, with One Life to Live following suit in January 2012. Both All My Children and One Life to Live were briefly revived online in 2013, before being cancelled again that same year. In 2019, production of Days of Our Lives was put on "indefinite hiatus" and all of the cast's contracts were terminated, raising concerns within soap publications that cancellation would ensue, though the show was later renewed through September 2021. In 2022, NBC announced that Days of Our Lives would be moved to its streaming service, Peacock, making NBC the first of the big three networks not to air any daytime soap operas. As women increasingly worked outside of the home, daytime television viewing declined. New generations of potential viewers were not raised watching soap operas with their mothers, leaving the shows' long and complex storylines foreign to younger audiences. As viewers age, ratings continue to drop among young adult women, the demographic group for which soap opera advertisers pay the most. Those who might watch in workplace breakrooms are not counted, as Nielsen does not track television viewing outside the home. The rise of cable and the Internet has also provided new sources of entertainment during the day. The genre's decline has additionally been attributed to reality television displacing soap operas as TV's dominant form of melodrama. An early term for the reality TV genre was docu-soap. A precursor to reality TV, the televised 1994–95 O. J. Simpson murder case, both preempted and competed with an entire season of soaps, transforming viewing habits and leaving soap operas with 10 percent fewer viewers after the trial ended. Daytime programming alternatives such as talk shows, game shows, and court shows cost up to 50% less to produce than scripted dramas, making those formats more profitable and attractive to networks, even if they receive the same or slightly lower ratings than soap operas. A network may even prefer to return a time slot to its local stations to keeping a soap opera with disappointing ratings on the air, as was the case with Sunset Beach and Port Charles. Compounding the financial pressure on scripted programming in the 2007–2010 period was a decline in advertising during the Great Recession, which led shows to reduce their budgets and cast sizes. In addition to these external factors, a litany of production decisions has been cited by soap opera fans as contributing to the genre's decline, such as clichéd plots, a lack of diversity that narrowed audience appeal, and the elimination of core families. Serials produced for prime time slots have also found success. The first prime time soap opera was Faraway Hill (1946), which aired on October 2, 1946, on the now-defunct DuMont Television Network. Faraway Hill ran for 12 episodes and was primarily broadcast live, interspersed with short pre-recorded film clips and still photos to remind the audience of the previous week's episode. The first long-running prime time soap opera was Peyton Place (1964–1969) on ABC. It was based in part on the eponymous 1957 film (which, in turn, was based on the 1956 novel). The popularity of Peyton Place prompted the CBS network to spin off popular As the World Turns character Lisa Miller into her own evening soap opera, Our Private World (originally titled "The Woman Lisa" in its planning stages). Our Private World was broadcast from May to September 1965. The character of Lisa (and her portrayer Eileen Fulton) returned to As The World Turns after the series ended. The structure of Peyton Place, with its episodic plots and long-running story arcs, set the mold for the prime time serials of the 1980s, when the format reached its pinnacle. The successful prime time serials of the 1980s included Dallas, its spin-off Knots Landing, Dynasty, and Falcon Crest. These shows frequently dealt with wealthy families, and their personal and big-business travails. Common characteristics were sumptuous sets and costumes, complex storylines examining business schemes and intrigue, and spectacular disaster cliffhanger situations. Each of these series featured a wealthy, domineering, promiscuous, and passionate antagonist as a key character in the storyline – respectively, J.R. Ewing (Larry Hagman), Abby Cunningham (Donna Mills), Alexis Colby (Joan Collins), and Angela Channing (Jane Wyman). These villainous schemers became immensely popular figures that audiences "loved to hate". Unlike daytime serials, which are shot on video in a studio using the multi-camera setup, these evening series were shot on film using a single camera setup, and featured substantial location-shot footage, often in picturesque locales. Dallas, its spin-off Knots Landing, and Falcon Crest all initially featured episodes with self-contained stories and specific guest stars who appeared in just that episode. Each story was completely resolved by the end of the episode, and there were no end-of-episode cliffhangers. After the first couple of seasons, all three shows changed their story format to that of a pure soap opera, with interwoven ongoing narratives that ran over several episodes. Dynasty featured this format throughout its run. The soap opera's distinctive open plot structure and complex continuity was increasingly incorporated into American prime time television programs of the period. The first significant drama series to do this was Hill Street Blues. This series, produced by Steven Bochco, featured many elements borrowed from soap operas, such as an ensemble cast, multi-episode storylines, and extensive character development over the course of the series. It and the later Cagney & Lacey overlaid the police series formula with ongoing narratives exploring the personal lives and interpersonal relationships of the regular characters. The success of these series prompted other drama series, such as St. Elsewhere and situation comedy series, to incorporate serialized stories and story structure to varying degrees. The prime time soap operas and drama series of the 1990s, such as Beverly Hills, 90210, Melrose Place, Party of Five, The OC, and Dawson's Creek, focused more on younger characters. In the 2000s, ABC began to revitalize the prime time soap opera format with shows such as Desperate Housewives, Grey's Anatomy, Brothers & Sisters, Ugly Betty, Private Practice, and more recently Revenge, Nashville, Scandal, Mistresses, and formerly Ringer, which its sister production company ABC Studios co-produced with CBS Television Studios for The CW. While not soaps in the traditional sense, these shows managed to appeal to wide audiences with their high drama mixed with humor, and are soap operas by definition. These successes led to NBC's launching serials, including Heroes and Friday Night Lights. The upstart MyNetworkTV, a sister network of Fox, launched a line of prime time telenovelas (a genre similar to soap operas in terms of content) upon its launch in September 2006, but discontinued its use of the format in August 2007 after disappointing ratings. On June 13, 2012, Dallas, a continuation of the 1978 original series premiered on the cable network, TNT. The revived series, which was canceled after three seasons in 2014, delivered solid ratings for the channel, only losing viewership after the show's most established star, Larry Hagman, died midway through the series. In 2012, Nick at Nite debuted a primetime soap opera, Hollywood Heights, which aired episodes five nights a week (on Monday through Fridays) in a manner similar to a daytime soap opera, instead of the once-a-week episode output common of other prime time soaps. The series, which was an adaptation of the Mexican telenovela Alcanzar una estrella, suffered from low ratings (generally receiving less than 1 million viewers) and was later moved to sister cable channel TeenNick halfway through its run to burn off the remaining episodes. In 2015, Fox debuted Empire, a prime time musical serial centering on the power struggle between family members within the titular recording company. Created by Lee Daniels and Danny Strong and led by Oscar nominees Terrence Howard and Taraji P. Henson, the drama premiered to high ratings. The show is strongly influenced by other works such as William Shakespeare's King Lear, James Goldman's The Lion in Winter and the 1980s soap opera Dynasty. Also in 2015, E! introduced The Royals, a series following the life and drama of a fictional English Royal family, which was also inspired by Dynasty (even featuring Joan Collins as the Queen's mother). In addition, ABC debuted a prime time soap opera Blood & Oil, following a young couple looking to make money off the modern-day Williston oil boom, premiering on September 27, 2015. The telenovela, a shorter-form format of serial melodrama, shares some thematic and especially stylistic similarity to the soap opera, enough that the colloquialism Spanish soap opera has arisen to describe the format. The chief difference between the two is length of series; while soap operas usually have indefinite runs, telenovelas typically have a central story arc with a prescribed ending within a year or two of the show's launch, requiring more concise storytelling. Spanish-language networks, chiefly Univision and Telemundo, have found success airing telenovelas for the growing U.S. Hispanic market. Both originally produced and imported Latin American dramas (as well as imported Turkish dramas since the 2020s) are popular features of the networks' daytime and primetime lineups, sometimes beating English-language networks in the ratings. Some web series are soap operas, such as Degrassi: In Session or Venice: The Series. In 2013, production company Prospect Park revived All My Children and One Life to Live for the web, retaining original creator Agnes Nixon as a consultant and keeping many of the same actors (Prospect Park purchased the rights to both series months after their cancellations by ABC in 2011, although it initially suspended plans to relaunch the soaps later that same year due to issues receiving approval from acting and production unions). Each show initially produced four half-hour episodes a week, but quickly cut back to two half-hour episodes each. In the midst of (though not directly related to) a lawsuit between Prospect Park and ABC, the experiment ended that same year, with both shows being canceled again. As of 2017, Turkey is the second largest exporter of television soap operas. In 2016, Turkish TV exports earned $350 million, making it the second largest drama exporter in the world behind the United States. Turkish soap operas have a large following across Asia, the Balkans, Eastern Europe, Latin America, the Middle East, and Africa. Soap operas in the UK began on radio and consequently were associated with the BBC. It had resisted soaps as antithetical to its quality image, but began broadcasting Front Line Family in April 1941 on its North American shortwave service to encourage American intervention on Britain's behalf in World War II. The BBC continues to broadcast the world's longest-running radio soap, The Archers, which first aired in May 1950, and has been running nationally since 1951. It is currently broadcast on BBC Radio 4 and continues to attract over five million listeners, or roughly 25% of the radio listening population of the UK at that time of the evening. In the UK, soap operas are one of the most popular genres, with most being broadcast during prime time. Most UK soap operas focus on everyday, working-class communities, influenced by the conventions of the kitchen sink drama. The most popular soap operas in the United Kingdom are Coronation Street, EastEnders, Emmerdale, Hollyoaks, Doctors, and the Australian produced Neighbours and Home and Away. The first three of these are consistently among the highest-rated shows on British television. Such is the magnitude of the popularity of the soap genre in the UK that all television serials in the country are reputedly enjoyed by members of the British Royal Family. King Charles III himself made cameo appearances in two of the UK's biggest serials during his time as Prince of Wales: Coronation Street and EastEnders, the latter alongside his wife Queen Camilla (then Duchess of Cornwall), in 2000 and 2022 respectively. Major events in British culture are often mentioned in the storyline, such as the Home Nations' participation at the World Cup and the death of Princess Diana. Since 1999, The British Soap Awards has been televised on ITV. The 1986 Christmas Day episode of EastEnders is often referred to as the highest-rated UK soap opera episode ever, with 30.15 million viewers (more than half the population at the time). The figure of 30.15 million was actually a combination of the original broadcast, which had just over 19 million viewers, and the Sunday omnibus edition with 10 million viewers. The combined 30.15 million audience figure makes the aforementioned Christmas Day 1986 episode of EastEnders the highest-rated single-channel broadcast in the history of UK television. Overall it ranks third behind the 1966 FIFA World Cup Final (32.3 million viewers) and Princess Diana's funeral in 1997 (32.1 million viewers) which were transmitted on both BBC One and ITV. An early television serial was The Grove Family on the BBC, which produced 148 episodes from 1954 to 1957. The programme was broadcast live and only a handful of recordings were retained in the archives. The UK's first twice-weekly serial was ITV's Emergency - Ward 10, running from 1957 until 1967. In the 1960s, Coronation Street revolutionised UK television and quickly became a British institution. On 17 September 2010, it became the world's longest-running television soap opera and was listed in Guinness World Records. The BBC also produced several serials: Compact was about the staff of a women's magazine; The Newcomers was about the upheaval caused by a large firm setting up a plant in a small town; United! contained 147 episodes and focused on a football team; 199 Park Lane (1965) was an upper class serial, which ran for only 18 episodes. None of these serials came close to making the same impact as Coronation Street. Indeed, most of the 1960s BBC serials were largely wiped. During the 1960s, Coronation Street's main rival was Crossroads, a daily serial that began in 1964 and aired on ITV in the early evening. Crossroads was set in a Birmingham motel and, although the program was popular, its purported low technical standard and bad acting were much mocked. By the 1980s, its ratings had begun to decline. Several attempts to revamp the program through cast changes and, later, expanding the focus from the motel to the surrounding community were unsuccessful. Crossroads was cancelled in 1988 (a new version of Crossroads was later produced, running from 2001 until 2003). A later rival to Coronation Street was ITV's Emmerdale Farm (later renamed Emmerdale), which began in 1972 in a daytime slot and was set in rural Yorkshire. Increased viewership resulted in Emmerdale being moved to a prime-time slot in the 1980s. Pobol y Cwm (People of the Valley) is a Welsh language serial that has been produced by the BBC since October 1974, and is the longest-running television soap opera produced by the broadcaster. Pobol y Cwm was originally broadcast on BBC Wales television from 1974 to 1982; it was then moved to the Welsh-language television station S4C when it opened in November 1982. The program was occasionally shown on BBC1 in London during periods of regional optout in the mid- to late 1970s. Pobol y Cwm was briefly shown in the rest of the UK in 1994 on BBC2, with English subtitles; it is consistently the most watched programme each week on S4C. Daytime soap operas were non-existent until the 1970s because there was virtually no daytime television in the UK. ITV introduced General Hospital, which later moved to a prime time slot. In 1980, Scottish Television debuted Take the High Road, which lasted for over twenty years. Later, daytime slots were filled with an influx of Australian soap operas such as The Sullivans (aired on ITV from 1977), The Young Doctors (from 1982), Sons and Daughters (from 1983), A Country Practice (from 1982), Richmond Hill (from 1988 to 1989) and eventually, Neighbours was acquired by the BBC in 1986, and Home and Away aired on ITV beginning in 1989. These achieved significant levels of popularity; Neighbours and Home and Away were moved to early-evening slots, helping begin the UK soap opera boom in the late 1980s. The day Channel 4 began operations in 1982 it launched its own soap, the Liverpool-based Brookside, which would redefine soaps over the next decade. The focus of Brookside was different from earlier soap operas in the UK; it was set in a middle-class new-build cul-de-sac, unlike Coronation Street and Emmerdale Farm, which were set in established working-class communities. The characters in Brookside were generally either people who had advanced themselves from inner-city council estates, or the upper middle-class who had fallen on hard times. Though Brookside was still broadcast in a pre-watershed slot (8.00 p.m. and 8.30 p.m. on weekdays, around 5.00 p.m. for the omnibus on Saturdays), it was more liberal than other soaps of the time: the dialogue regularly included expletives. This stemmed from the overall more liberal policy of the channel during that period. The soap was also heavily politicised. Bobby Grant (Ricky Tomlinson), a militant trade-unionist anti-hero, was the most overtly political character. Storylines were often more sensationalist than on other soaps (throughout the soap's history, there were two armed sieges on the street) and were staged with more violence (particularly, rape) often being featured. In 1985, the BBC's EastEnders debuted and became a near instant success with viewers and critics alike, with the first episode attracting over 17 million viewers. The Christmas Day 1986 episode was watched by 30.15 million viewers and contained a scene in which divorce papers were served to Angie Watts (Anita Dobson) by her husband, Queen Vic landlord Den (Leslie Grantham). A notable success in pioneering late-night broadcasting, in October 1984, Yorkshire Television began airing the cult Australian soap opera Prisoner, which originally ran from 1979 to 1986. It was eventually broadcast on all regions of the UK in differing slots, usually around 23:00 (but never before 22:30 in any region), under the title Prisoner: Cell Block H. It was probably most popular in the Midlands where Central Television consistently broadcast the serial three times a week from 1987 to 1991. Its airing in the UK was staggered, so different regions of the country saw it at a different pace. The program was immensely successful, regularly achieving 10 million viewers when all regions' ratings per episode were added together. Central bowed to fan pressure to repeat the soap, of which the first 95 episodes aired. Then, rival station Channel 5 also acquired rights to repeat the entire rerun of the program, starting in 1997. All 692 episodes have since been released on DVD in the UK. In 1992, the BBC made Eldorado to daily alternate with EastEnders. The programme was heavily criticised and only lasted one year. Nevertheless, soap operas gained increasing prominence on UK television schedules. In 1995, Channel 4 premiered Hollyoaks, a soap with a youth focus. When Channel 5 launched in March 1997, it debuted the soap opera Family Affairs, which was formatted as a week-daily soap, airing Monday through Fridays. Brookside's premise evolved during the 1990s, phasing out the politicised stories of the 1980s and shifting the emphasis to controversial and sensationalist stories such as child rape, sibling incest, religious cults and drug addiction, including the infamous 'body under the patio' storyline that ran from 1993 to 1995, and gave the serial its highest ratings ever with 9 million viewers. Coronation Street and Brookside began releasing straight-to-video features. The Coronation Street releases generally kept the pace and style of conventional programs episodes with the action set in foreign locations. The Brookside releases were set in the usual locations, but featured stories with adult content not allowed on television pre-watershed, with these releases given '18' certificates. Emmerdale Farm was renamed Emmerdale in 1989. The series was revamped in 1993 with many changes executed via the crash of a passenger jet that partially destroyed the village and killed several characters. This attracted criticism as it was broadcast near the fifth anniversary of the Lockerbie bombing. The storyline drew the soap its highest ever audience of 18 million viewers. The revamp was a success and Emmerdale grew in popularity. Throughout the 1990s, Brookside, Coronation Street, EastEnders and Emmerdale continued to flourish. Each increased the number of episodes that aired on a weekly basis by at least one, further defining soap operas as the leading genre in British television. Since 2000, new soap operas have continued to be developed. Daytime serial Doctors began in March 2000, preceding Neighbours on BBC One and had since become the BBC's flagship daytime series. The series was cancelled in October 2023, with the final episode set to be screened in December 2024. In 2002, as ratings for the Scottish serial High Road (formerly Take The High Road) continued to decline, BBC Scotland launched River City, which proved popular and effectively replaced High Road when it was cancelled in 2003. The long-running serial Brookside ended in November 2003 after 21 years on the air, leaving Hollyoaks as Channel 4's flagship serial. A new version of Crossroads featuring a mostly new cast was produced by Carlton Television for ITV in 2001. It did not achieve high ratings and was cancelled in 2003. In 2001, ITV also launched a new early-evening serial entitled Night and Day. This program too attracted low viewership and, after being shifted to a late night time slot, was cancelled in 2003. Family Affairs, which was broadcast opposite the racier Hollyoaks, never achieved significantly high ratings leading to several dramatic casting revamps and marked changes in style and even location over its run. By 2004, Family Affairs had a larger fan base and won its first awards, but was cancelled in late 2005. In 2008, ITV premiered The Royal Today, a daily spin-off of popular 1960s-based drama The Royal (itself a spin-off of Heartbeat), which had been running in a primetime slot since 2003. Just days later, soap opera parody programme Echo Beach premiered alongside its sister show, the comedy Moving Wallpaper. Both Echo Beach and The Royal Today ended after just one series due to low ratings. Radio soap opera Silver Street debuted on the BBC Asian Network in 2004. Poor ratings and criticism of the programme led to its cancellation in 2010. UK soap operas for many years usually only aired two nights a week. The exception was the original Crossroads, which began as a week-daily soap opera in the 1960s, but later had its number of weekly broadcasts reduced. In 1989, Coronation Street began airing three times a week. In 1996, it expanded to four episodes a week. Brookside premiered in 1982 with two episodes a week. In 1990 it expanded to three episodes a week. EastEnders increased its number of episodes a week in 1994 and Emmerdale did so in 1997. Family Affairs debuted as a weekdaily soap in 1997, producing five episodes a week its entire run. In 2004, Emmerdale began airing six episodes a week. In a January 2008 overhaul of the ITV network, the Sunday episodes of Coronation Street and Emmerdale were moved out of their slots. Coronation Street added a second episode on Friday evenings at 8:30 p.m. Emmerdale's Tuesday edition was extended to an hour, putting it in direct competition with EastEnders. In July 2009, the schedules of these serials were changed again. On 23 July 2009, Coronation Street moved from the Wednesday slot it held for 49 years, to Thursday evenings. Emmerdale reverted to running just one 30-minute episode on Tuesday evenings and the other 30-minute installment was moved to Thursday evenings. Coronation Street later returned to a Wednesday slot, to air Mondays, Wednesdays and Fridays at 19:30 and 20:30. Emmerdale airs at 19:00 every weeknight, and 20:00 on Thursdays. Later, Coronation Street (which began airing two episodes on Monday nights in 2002) produced five episodes a week. It was announced in June 2016 that starting late 2017, Coronation Street would air six episodes a week. Doctors aired five episodes a week until 2022, and four episodes from 2022 onwards, and is the only soap without a weekend omnibus repeat screening. Hollyoaks produces five episodes a week. The imported Neighbours screens as five new episodes a week. As of 2019, EastEnders produces four episodes a week. UK soap operas are shot on videotape in the studio using a multi-camera setup. In their early years, Coronation Street and Emmerdale used 16 mm film for footage shot on location. Since the 1980s, UK soap opera have routinely featured scenes shot outdoors in each episode. This footage is shot on videotape on a purpose-built outdoor set that represents the community that the soap focuses on. Hollyoaks and Family Affairs were taped on high-definition video, and used the filmizing process. Australia has had quite a number of well-known soap operas, some of which have gained cult followings in the United Kingdom, New Zealand and other countries. The majority of Australian television soap operas are produced for early evening or evening timeslots. They usually produce two or two and a half hours of new material each week, either arranged as four or five half-hour episodes a week, or as two one-hour episodes. Stylistically, these series most closely resemble UK soap operas in that they are nearly always shot on videotape, are mainly recorded in a studio and use a multi-camera setup. The original Australian serials were shot entirely in studio. During the 1970s occasional filmed inserts were used to incorporate sequences shot outdoors. Outdoor shooting later became commonplace and starting in the late 1970s, it became standard practice for some on-location footage to be featured in each episode of any Australian soap opera, often to capitalise on the attractiveness and exotic nature of these locations for international audiences. Most Australian soap operas focus on a mixed age range of middle-class characters and will regularly feature a range of locations where the various, disparate characters can meet and interact, such as the café, the surf club, the wine bar or the school. The genre began in Australia on radio, as it had in the United States and the United Kingdom. One such radio serial, Big Sister, featured actress Thelma Scott in the cast and aired nationally for five years beginning in 1942. Probably the best known Australian radio serial was the long-running soap opera Blue Hills, which was created by Gwen Meredith and ran from 1949 to 1976. With the advent of Australian television in 1956, daytime television serials followed. The first Australian television soap opera was Autumn Affair (1958) featuring radio personality and Blue Hills star Queenie Ashton making the transition to television. Each episode of this serial ran for 15 minutes and aired each weekday on the Seven Network. Autumn Affair failed to secure a sponsor and ended in 1959 after 156 episodes. It was followed by The Story of Peter Grey (1961), another Seven Network weekday series aired in a daytime slot in 15-minute installments. The Story of Peter Grey ran for 164 episodes. The first successful wave of Australian evening television soap operas started in 1967 with Bellbird, produced by the Australian Broadcasting Corporation. This rural-based serial screened in an early evening slot in 15-minute installments as a lead-in to the evening news. Bellbird was a moderate success but built up a consistent and loyal viewer base, especially in rural areas, and enjoyed a ten-year run. Motel (1968) was Australia's first half-hour soap opera; the daytime soap had a short run of 132 episodes. The first major soap opera hit in Australia was the sex-melodrama Number 96, a nighttime series produced by Cash Harmon Television for Network 10, which debuted March 1972. The program dealt with such topics as homosexuality, adultery, drug use, rape within marriage and racism, which had rarely been explored on Australian television programs before. The series became famous for its sex scenes and nudity and for its comedic characters, many of whom became cult heroes in Australia. By 1973, Number 96 had become Australia's highest-rated show. In 1974, the sexed-up antics of Number 96 prompted the creation of The Box, which rivaled it in terms of nudity and sexual situations and was scheduled in a nighttime slot. Produced by Crawford Productions, many critics considered The Box to be a more slickly produced and better written show than Number 96. The Box also aired on the Ten Network, programmed to run right after Number 96. For 1974 Number 96 was again the highest rating show on Australian television, and that year The Box occupied the number two spot. Also in 1974, the Reg Grundy Organisation created its first soap opera, and significantly Australia's first teen soap opera, Class of '74. With its attempts to hint at the sex and sin shown more openly on Number 96 and The Box, its high school setting and early evening timeslot, Class of '74 came under intense scrutiny from the Broadcasting Control Board, who vetted scripts and altered entire storylines. By 1975, both Number 96 and The Box, perhaps as a reaction to declining ratings for both shows, de-emphasised the sex and nudity shifting more towards comedic plots. Class of '74 was renamed Class of '75 and also added more slapstick comedy for its second year, but the revamped show's ratings declined, resulting in its cancellation in mid-1975. That year Cash Harmon's newly launched second soap The Unisexers failed in its early evening slot and was cancelled after three weeks; the Reg Grundy Organisation's second soap Until Tomorrow ran in a daytime slot for 180 episodes. A feature film version of Bellbird entitled Country Town was produced in 1971 by two of the show's stars, Gary Gray and Terry McDermott, without production involvement by the Australian Broadcasting Corporation. Number 96 and The Box also released feature film versions, both of which had the same title as the series, released in 1974 and 1975 respectively. As Australian television had broadcast in black and white until 1975, these theatrical releases all had the novelty of being in colour. The film versions of Number 96 and The Box also allowed more explicit nudity than could be shown on television at that time. In November 1976 The Young Doctors debuted on the Nine Network. This Grundy Organization series eschewed the adult drama of Number 96 and The Box, focusing more on relationship drama and romance. It became a popular success but received few critical accolades. A week later The Sullivans, a carefully produced period serial chronicling the effects of World War II on a Melbourne family, also debuted on Nine. Produced by Crawford Productions, The Sullivans became a ratings success, attracted many positive reviews, and won television awards. During this period Number 96 re-introduced nudity into its episodes, with several much-publicised full-frontal nude scenes, a cast revamp and a new range of shock storylines designed to boost the show's declining ratings. Bellbird experienced changes to its broadcast pattern with episodes screening in 60 minute blocks, and later in 30 minute installments. Bellbird, Number 96 and The Box, which had been experiencing declining ratings, were cancelled in 1977. Various attempts to revamp each of the shows with cast reshuffles or spectacular disaster storylines had proved only temporarily successful. The Young Doctors and The Sullivans continued to be popular. November 1977 saw the launch of successful soap opera/police procedural series Cop Shop (1977–1984) produced by Crawford Productions for Channel Seven. In early December 1977 Channel Ten debuted the Reg Grundy Organisation produced The Restless Years (1977–1981), a more standard soap drama focusing on several young school leavers. The Seven Network, achieving success with Cop Shop produced by Crawford Productions, had Crawfords produce Skyways, a series with a similar format but set in an airport, to compete with the Nine Network's popular talk show The Don Lane Show. Skyways, which debuted in July 1979, emphasised adult situations including homosexuality, marriage problems, adultery, prostitution, drug use and smuggling, crime, suicide, political intrigue, and murder, and featured some nudity. Despite this, the program achieved only moderate ratings and was cancelled in mid-1981. The Reg Grundy Organisation found major success with the women's-prison drama Prisoner (1979–1986) on Network Ten, and melodramatic family saga Sons and Daughters (1982–1987) on the Seven Network. Both shows achieved high ratings in their original runs, and unusually, found success in repeats after the programs ended. Grundy soap The Young Doctors and Crawford Productions' The Sullivans continued on the Nine Network until late 1982. Thereafter Nine attempted many new replacement soap operas produced by the Reg Grundy Organisation: Taurus Rising (1982), Waterloo Station (1983), Starting Out (1983) and Possession (1985), along with Prime Time (1986) produced by Crawford Productions. None of these programs were successful and most were cancelled after only a few months. The Reg Grundy Organisation also created Neighbours, a suburban-based daily serial devised as a gentle family drama with some comedic and lightweight situations, for the Seven Network in 1985. Produced in Melbourne at the studios of HSV-7, Neighbours achieved high ratings in Melbourne, Brisbane and Adelaide, but not in Sydney, where it aired at 5.30 p.m. placing it against the hit dating game show Perfect Match on Channel 10. The Seven Network's Sydney station ATN-7 quickly lost interest in Neighbours as a result of the low ratings in Sydney. HSV-7 in Melbourne lobbied heavily to keep Neighbours on the air, but ATN-7 managed to convince the rest of the network to cancel the show and instead keep ATN-7's own Sydney-based dramas A Country Practice and Sons and Daughters. After the network cancelled Neighbours, it was immediately picked up by Channel Ten, which revamped the cast and scripts slightly and aired the series in the 7.00 p.m. slot starting 20 January 1986. It initially attracted low audiences; however, after a concerted publicity drive, Ten managed to transform the series into a major success, turning several of its actors into major international stars. The show's popularity eventually declined and it was moved to the 6.30 p.m. slot in 1992. In January 2011 it moved to Eleven and ended after 8,903 episodes on 28 July 2022. In November 2022, Amazon Freevee revived the show with an order of 400 episodes to begin airing in 2023. It is Australia's longest-running soap opera. The success of Neighbours in the 1980s prompted the creation of somewhat similar suburban and family or teen-oriented soap operas such as Home and Away (1988–present) on Channel Seven and Richmond Hill (1988) on Channel Ten. Both proved popular, however Richmond Hill emerged as only a moderate success and was cancelled after one year to be replaced on Ten by E Street (1989–1993). Nine continued trying to establish a successful new soap opera, without success. After the failure of family drama Family and Friends in 1990, it launched the raunchier and more extreme Chances in 1991, which resurrected the sex and melodrama of Number 96 and The Box in an attempt to attract attention. Chances achieved only moderate ratings, and was moved to a late-night timeslot. It underwent several revamps that removed much of the original cast, and refocused the storylines to incorporate science-fiction and fantasy elements. The series continued in a late night slot until 1992, when it was cancelled due to low ratings despite the much-discussed fantasy storylines. Several Australian soap operas have also found significant international success. In the UK, starting in the mid-1980s, daytime broadcasts of The Young Doctors, The Sullivans, Sons and Daughters and Neighbours (which itself was subsequently moved to an early-evening slot) achieved significant success. Grundy's Prisoner began airing in the United States in 1979 and achieved high ratings in many regions there, however, the show ended its run in that country three years into its run. Prisoner also aired in late-night timeslots in the UK beginning in the late 1980s, achieving enduring cult success there. The show became so popular in that country that it prompted the creation of two stage plays and a stage musical based on the show, all of which toured the UK, among many other spin-offs. In the late 1990s, Channel 5 repeated Prisoner in the UK. Between 1998 and 2005, Channel 5 ran late-night repeats of Sons and Daughters. During the 1980s, the Australian attempts to emulate big-budget U.S. soap operas such as Dallas and Dynasty had resulted in the debuts of Taurus Rising and Return to Eden, two slick soap opera dramas with big budgets that were shot entirely on film. Though their middling Australian ratings resulted in the shows running only for a single season, both programs were successfully sold internationally. Other shows to achieve varying levels of international success include Richmond Hill, E Street, Paradise Beach (1993–1994), and Pacific Drive (1995–1997). Indeed, these last two series were designed specifically for international distribution. Channel Seven's Home and Away, a teen soap developed as a rival to Neighbours, has also achieved significant and enduring success on UK television. Something in the Air, a serial examining a range of characters in a small country town ran on the ABC from 2000 to 2002. Attempts to replicate the success of daily teen-oriented serials Neighbours and Home and Away saw the creation of Echo Point (1995) and Breakers (1999) on Network Ten. These programs foregrounded youthful attractive casts and appealing locations but the programs were not long-running successes and Neighbours and Home and Away remained the most visible and consistently successful Australian soap operas in production. In their home country, they both attracted respectable although not spectacular ratings in the early 2000s. By 2004, Neighbours was regularly attracting just under a million viewers per episode – considered at that time a low figure for Australian prime time television. By March 2007, the Australian audience for Neighbours had fallen to fewer than 700,000 a night. This prompted a revamp of the show's cast, its visual presentation, and a move away from the recently added action-oriented emphasis to refocus the show on the domestic storylines it is traditionally known for. During this period Neighbours and Home and Away continued to achieve significant ratings in the UK. This and other lucrative overseas markets, along with Australian broadcasting laws that enforce a minimum amount of domestic drama production on commercial television networks, help ensure that both programs remain in production. Both shows get higher total ratings in the UK than in Australia (the UK has three times the total population of Australia) and the UK channels make a major contribution to the production costs. It has been suggested that with their emphasis on the younger, attractive and charismatic characters, Neighbours and Home and Away have found success in the middle ground between glamorous, fantastic U.S. soaps with their wealthy but tragic heroes and the more grim, naturalistic UK soap operas populated by older, unglamorous characters. The casts of Neighbours and Home and Away are predominantly younger and more attractive than the casts of UK soaps, and without excessive wealth and glamour of the U.S. daytime serial, a middle-ground in which they have found their lucrative niche. Neighbours was carried in the United States on the Oxygen cable channel in March 2004; however it attracted few viewers, perhaps in part due to its scheduling opposite well-established and highly popular U.S. soap operas such as All My Children and The Young and the Restless, and was dropped by the network shortly afterwards due to low ratings. headLand made its debut on Channel Seven in November 2005, the series arose out of a proposed spinoff of Home and Away that was to have been produced in conjunction with Home and Away's UK broadcaster, Channel 5. The idea for the spin-off was scuttled after Five pulled out of the deal, which meant that the show could potentially air on a rival channel in the UK; as such, Five requested that the new show be developed as a standalone series and not be spun off from a series that it owned a stake in. The series premiered in Australia on November 15, 2005, but was not a ratings success and was cancelled two months later on January 23, 2006. The series broadcast on E4 and Channel 4 in the UK. Nickelodeon's H2O: Just Add Water appeared in July 2006 on Network Ten. Since Connie considered this mention as a torrid soap opera, this was mentioned in the Steven Universe episode "Love Letters". After losing the UK television rights to Neighbours to Five, the BBC commissioned a replacement serial Out of the Blue, which was produced in Australia. It debuted as part of BBC One's weekday afternoon schedule on 28 April 2008 but low ratings prompted its move to BBC Two on 19 May 2008. The series was cancelled after its first season. Neighbours' continued low ratings in Australia resulted in it being moved to Ten's new digital channel, Eleven on January 11, 2011. However, it continues to achieve reasonable ratings on Channel 5 in the United Kingdom, and as of March 2013 still reportedly achieved significant international sales. Neighbours was cancelled due to Channel 5, the UK broadcaster of the show, deciding to drop the programme – the money they were paying for the rights was providing the majority of its funding. It ended on 29 July 2022. Months after its series finale, Fremantle Australia, the programme's production company, announced on 17 November 2022, that production on the programme will restart in 2023 after the company agreed on a deal with Amazon Freevee. Amazon Freevee will air the programme for free in the UK and the US while Network 10 will retain the rights to the programme. Pioneering series Pukemanu aired over two years (1971–72) and was the NZBC's first continuing drama. It followed the goings-on of a North Island timber town. Close to Home is a New Zealand television soap opera that ran on TVNZ 1 from 1975 to 1983. At its peak in 1977 nearly one million viewers tuned in twice weekly to watch the series co-created by Michael Noonan and Tony Isaac (who had initially only agreed to make the show on the condition they would get to make The Governor). Gloss is a television drama series that screened from 1987 to 1990. The series is about a fictional publishing empire run by the Redfern family. Gloss was NZ's answer to US soap Dynasty, with the Carrington oil scions replaced by the wealthy Redferns and their Auckland magazine empire. It was a starting point for many actors who went on to many productions in New Zealand, Australia and around the world including Temuera Morrison, Miranda Harcourt, Peter Elliott, Lisa Chappell, Danielle Cormack and Kevin Smith. Many of them would go on to star in Shortland Street, which has been New Zealand's most popular soap since its debut in 1992. It airs on TVNZ 2. Radio New Zealand began airing its first radio soap You Me Now in September 2010. It is available for podcast on its website. Relatively few daily soap operas have been produced on English Canadian television, with most Canadian stations and networks that carry soap operas airing those imported from the United States or the United Kingdom. Notable daily soaps that did exist included Family Passions, Scarlett Hill, Strange Paradise, Metropia, Train 48 and the international co-production Foreign Affairs. Family Passions was an hour-long program, as is typical of American daytime soaps; all of the others ran half-hour episodes. Unlike American or British soap operas, the most influential of which have run for years or even decades, even daily Canadian soap operas have run for a few seasons at most. Short-run soaps, including 49th & Main and North/South, have also aired. Many of these were produced in an effort to comply with Canadian content regulations, which require a percentage of programming on Canadian television to originate from Canada. Notable prime time soap operas in Canada have included Riverdale, House of Pride, Paradise Falls, Lance et Compte ("He Shoots, He Scores"), Heartland, Loving Friends and Perfect Couples, and The City. The Degrassi franchise of youth dramas also incorporated some elements of the soap opera format. On French-language television in Quebec, the téléroman has been a popular mainstay of network programming since the 1950s. Notable téléromans have included Rue des Pignons, Les Belles Histoires des pays d'en haut, Diva, La famille Plouffe, and the soap opera parody Le Cœur a ses raisons. Unlike the season-based production in most countries, most of Indian television fiction tends to be regular-broadcasting soap opera. These started in the 1980s, as more and more people began to purchase television sets. At the beginning of the 21st century, soap operas became an integral part of Indian culture. Indian soap operas mostly concentrate on the conflict between love and arranged marriages occurring in India, and many includes family melodrama. Indian soap operas have multilingual production. Many soap operas produced in India are also broadcast overseas in the UK, Canada, the United States, and some parts of Europe, South Africa, Australia and South East Asia. They are often mass-produced under large production banners, with companies like Balaji Telefilms running different language versions of the same serial on different television networks or channels. The Australian serial The Restless Years was remade in the Netherlands as Goede tijden, slechte tijden (which debuted in 1990) and in Germany as Gute Zeiten, schlechte Zeiten (which has aired since 1992): both titles translate to "good times, bad times". These remakes are still airing, but have long since diverged from the original Australian storylines. The two shows are the highest-rated soap operas in their respective countries. A later Australian serial, Sons and Daughters, has inspired five remakes produced under license from the original producers and based, initially, on original story and character outlines. These are Verbotene Liebe (Germany, 1995–2015); Skilda världar (Sweden, 1996–2002); Apagorevmeni agapi (Greece, 1998); Cuori Rubati (Italy, 2002–2003) and Zabranjena ljubav (Croatia, 2004–2008). Both The Restless Years and Sons and Daughters were created and produced in Australia by the Reg Grundy Organisation. Another Australian soap opera reformatted for a European audience was E Street which ran on Network 10 in Australia from 1989 to 1993. Germany produced 37 episodes of Westerdeich ("Westside") in 1995 using scripts from 1989 episodes of E Street. It was also remade in Belgium as Wittekerke ("Whitechurch") and ran from 1993 to 2008. The Norwegian soap opera Hotel Cæsar aired on TV 2 from 1998 to 2017, and is the longest-running television drama in Scandinavia. Popular foreign soaps in the country include Days of Our Lives (broadcast on TV6 (Norway), The Bold and the Beautiful (TNT (Norway) and Home and Away (TV 2), all of which are subtitled. Serials have included Goede tijden, slechte tijden (1990–present), Onderweg naar Morgen (1994–2010) and Goudkust (1996–2001). In 2016 Goede tijden, slechte tijden spin-off Nieuwe Tijden started airing, but was ultimately cancelled by broadcaster RTL in 2018. Linear viewership for Goede tijden, slechte tijden, the country's most prominent soap opera, has decreased in recent years. However, due to a rising viewership on streaming platforms, RTL has decided to continue producing the show throughout the 2022/2023 television season. U.S. daytime serials As The World Turns and The Bold and the Beautiful have aired in the Netherlands; As the World Turns began airing in the country in 1990, with Dutch subtitles. In the 1980s, West German networks successfully added American daytime and primetime soap operas to their schedule before Das Erste introduced its first self-produced weekly soap with Lindenstraße, which was seen as a German counterpart to Coronation Street. Like in other countries, the soap opera met with negative reviews, but eventually proved critics wrong with nearly 13 million viewers tuning in each week. Even though the format proved successful, it was not until 1992 before Gute Zeiten, schlechte Zeiten became the first German daily soap opera. Early ratings were bad as were the reviews, but the RTL network was willing to give its first soap opera a chance; ratings would improve, climbing to 7 million viewers by 2002. Not long after Gute Zeiten, schlechte Zeiten, Das Erste introduced Marienhof, which aired twice a week. After successfully creating the first German daily soap, production company Grundy Ufa wanted to produce another soap for RTL. Like GZSZ, the format was based on an Australian soap opera from Reg Watson. But RTL did not like the plot idea about separated twins who meet each other for the first time after 20 years and fall in love without knowing that they are related. The project was then taken to Das Erste, which commissioned the program, titled Verbotene Liebe, which premiered on January 2, 1995. With the premiere of Verbotene Liebe, the network turned Marienhof into a daily soap as well. In the meanwhile, RTL debuted the Grundy Ufa–produced Unter uns in late 1994. ZDF started a business venture with Canada and co-produced the short-lived series Family Passions, starring actors such as Gordon Thomson, Roscoe Born, Dietmar Schönherr and a young Hayden Christensen. The daytime serial premiered on December 5, 1994, lasting 130 episodes. After its cancellation, the network debuted Jede Menge Leben. Even after a crossover with three soaps, Freunde fürs Leben, Forsthaus Falkenau and Unser Lehrer Doktor Specht, the soap was canceled after 313 episodes. Sat.1 tried to get into the soap business as well, after successfully airing the Australian soap opera Neighbours, which was dropped in 1995 due to the talk show phenomenon that took over most of the daytime schedules of German networks. The network first tried to tell a family saga with So ist das Leben! Die Wagenfelds, before failing with Geliebte Schwestern. RTL II made its own short-lived attempt with Alle zusammen – jeder für sich. The teen soap opera Schloss Einstein debuted on September 4, 1998, focusing on the life of a group of teenagers at the fictional titular boarding school near Berlin. As of July 2014, the series has produced over 815 episodes during the course of 17 seasons, a milestone in German television programming, and was renewed for an 18th season to debut in 2015. In 1999, after the lasting success of Gute Zeiten, schlechte Zeiten, Marienhof, Unter uns and Verbotene Liebe, ProSieben aired Mallorca – Suche nach dem Paradies, set on the Spanish island with the same name. After nine months, the network canceled the program due to low viewership and high production costs. Even though ratings had improved, the show ended its run in a morning timeslot. The soap opera became something of a cult classic, as its 200-episode run was repeated several times on free-to-air and pay television. In 2006, Alles was zählt became the last successful daily soap to make its debut, airing as a lead-in to Gute Zeiten, schlechte Zeiten and also produced by Grundy Ufa. Since Germany started to produce its own telenovelas, all soap operas faced declines in ratings. Unter uns was in danger of cancellation in 2009, but escaped such a fate due to budget cuts imposed by the show's producers and the firing of original cast member Holger Franke, whose firing and the death of his character outraged fans, resulting in a ratings spike in early 2010. After Unter uns was saved, Das Erste planned to make changes to its soap lineup. Marienhof had to deal with multiple issues in its storytelling, as well as in producing a successful half-hour show. Several changes were made within months, however Marienhof was canceled in June 2011. Verbotene Liebe was in danger of being cancelled as well, but convinced the network to renew it with changes that it made in both 2010 and 2011; the soap was later expanded to 45 minutes after Marienhof was canceled, and the network tried to decide on whether to revamp its lineup. While Gute Zeiten, schlechte Zeiten, Unter uns and Alles was zählt are currently the only daily soaps on the air after Verbotene Liebe has been cancelled and aired its last episode in June, 2015 due to low ratings, the telenovelas Sturm der Liebe and Rote Rosen are considered soaps by the press as well, thanks to the changing protagonists every season. In Belgium, the two major soap operas are Thuis ("Home") and Familie ("Family"), both prime time soap operas. Soap operas have been very popular in Flanders, the Dutch-speaking part of Belgium. Familie debuted in late 1991, and with more han 6,700 half-hour episodes, it has the highest episode total of any soap in Europe outside of the United Kingdom. The highest-rated soap opera is Thuis, which has aired on "één" since late 1995. Thuis is often one of the five most-watched Belgian shows and regularly garners over one million viewers (with 6.6 million Flemings in total). During the 1990s, foreign soap operas such as Neighbours and The Bold and the Beautiful were extremely popular, the latter having achieved a cult status in Belgium and airing in the middle of the decade during prime time. Both soaps still air today, along with other foreign soaps such as Days of Our Lives, Australia's Home and Away and Germany's "Sturm der Liebe". Vitaya unsuccessful attempted to air the Dutch soap opera "Goede Tijden, Slechte Tijden" in 2010. Other foreign soaps that previously aired on Belgian television include The Young and the Restless, EastEnders (both on VTM), "Port Charles" (at één, then known as TV1) and "Coronation Street" (on Vitaya). "Santa Barbara" aired during the 1990s on VTM for its entire run. In early 2000s, the only teen soap opera on Belgian television was Spring ("Jump" in English), which aired on the youth-oriented Ketnet and produced over 600 15-minute episodes from late 2002 until 2009, when it was cancelled after a steady decline in ratings following the departures of many of its original characters. The most successful soap opera in Italy is the evening series Un posto al sole ("A Place Under the Sun"), which had aired on Rai 3 since 1996 (whose format is based on the Australian soap opera Neighbours). Several other Italian soaps have been produced such as Ricominciare ("Starting Over"), Cuori rubati ("Stolen Hearts"), Vivere ("Living"), Sottocasa ("Downstairs"), Agrodolce ("Bittersweet") and Centovetrine ("Hundred Shop Windows"). The most popular Italian prime-time soap opera, Incantesimo ("Enchantment"), which ran from 1998 to 2008, became a daytime soap opera for the final two years of its run, airing five days a week on Rai 1. The same happened with Il paradiso delle signore (Woman's Paradise), a period drama, which ran from 2015 to 2017 in prime time, and became a daytime period soap opera from 2018. In the early years of RTÉ, the network produced several dramas but had not come close to launching a long-running serial. RTÉ's first television soap was Tolka Row, which was set in urban Dublin. For several years, both Tolka Row and The Riordans were produced by RTÉ; however, the urban soap was soon dropped in favor of the more popular rural soap opera The Riordans, which premiered in 1965. Executives from Yorkshire Television visited during on-location shoots for The Riordans in the early 1970s and in 1972, debuted Emmerdale Farm (now Emmerdale), based on the successful format of the Irish soap opera. In the late 1970s, The Riordans was controversially dropped. The creator of that series would then go on to produce the second of his "Agri-soap" trilogy Bracken, starring Gabriel Byrne, whose character had appeared in the last few seasons of The Riordans. Bracken was soon replaced by the third "Agri-soap" Glenroe, which ran until 2001. As RTÉ wanted a drama series for its Sunday night lineup rather than a soap opera, On Home Ground (2001–2002), The Clinic (2002–2009) and RAW (2010–2013) replaced the agri-soaps of the previous decades. In 1989, RTÉ decided to produce its first Dublin-based soap opera since the 1960s. Fair City, which is set in the fictional city of Carrickstown, initially aired one night a week during the 1989–90 season, and similar to its rural soaps, much of the footage was filmed on location – in a suburb of Dublin City. In 1992, RTÉ made a major investment into the series by copying the houses used in the on-location shoots for an on-site set in RTÉ's Headquarters in Dublin 4. By the early 1990s, it was airing two nights a week for 35 weeks a year. With competition from the UK soap operas, RTÉ expanded Fair City to three nights a week for most of the year and one night a week during the summer in 1996, later expanding to four nights a week and two nights during the summer. Until the early 2000s, the series produced four episodes a week, airing all 52 weeks of the year. Fair City airs Sundays, Tuesdays and Thursdays at 8.00 p.m. GMT on RTÉ One; however, after rival network TV3 moved Coronation Street to Thursday night, the Wednesday night episode of Fair City began airing at 7:30 p.m. each week. TG4 produce the Irish language soap Ros na Rún ("Headland of the Secrets" or "Headland of the Sweethearts"); set in the fictional village of Ros Na Rún, located outside Galway and near Spiddal, it centres on the domestic and professional lives of the town's residents. It is modeled on an average village in the West of Ireland, but with its own distinct personality – with a diverse population that share secrets, romances and friendships among other things. While the core community has remained the same, the look and feel of Ros Na Rún has changed and evolved over the years to incorporate the changing face of rural Ireland. It has an established a place not only in the hearts and minds of the Irish speaking public, but also the wider Irish audience. The program has dealt with many topics, including domestic violence, infidelity, theft, arson, abortion, homosexuality, adoption, murder, rape, drugs, teen pregnancy and paedophilia. It runs twice a week for 35 weeks of the year, currently airing Tuesday and Thursday nights. Ros na Rún is the single largest independent production commissioned in the history of Irish broadcasting. Prior to TG4's launch, it originally aired on RTÉ One in the early 1990s. Although Ireland has access to international soaps (such as Coronation Street, Emmerdale, EastEnders, Home and Away, Hollyoaks and Neighbours), Fair City continues to outperform them all, and is Ireland's most popular soap opera, with the show peaking at over 700,000 viewers. January 2015 Red Rock has broadcast on TV3. Red Rock airs twice a week on Wednesday and Thursday nights. The series is base in a fishing village in Dublin. The soaps centres around the local Garda station but also includes stories from the village. RTÉ Radio produced its first radio soap, Kennedys of Castleross, which ran from April 13, 1955, to 1975. In 1979 RTÉ long running TV soap The Riordans moved to Radio until December 24, 1985. In the mid-1980s, RTÉ debuted a new radio soap, Harbour Hotel, which ran until the mid-1990s. The network later ran two short-lived radio soaps, Konvenience Korner and Riverrun, which were followed in 2004 by Driftwood. RTÉ does not run any radio soaps, however RTÉ Radio 1 continues to air radio dramas as part of its nighttime schedule. In Greece, there have been several soap operas. An early serial was Sti skia tou hrimatos ("Money Shadows"), which ran from 1990 to 1991. September 1991 saw the debut of Lampsi ("the Shining"), from creator Nicos Foskolos. The series would become Greece's longest-running soap opera. After the success of Lampsi came the short lived To galazio diamandi ("Blue Diamond") and Simphonia siopis ("Omertà"). Lampsi was canceled in June 2005 due to declining ratings. It was replaced by Erotas ("Love"), a soap that ran from 2005 to 2008. After that series ended, ANT1 abandoned the soap opera genre and focused on comedy series and weekly dramas. Greece's second longest-running soap is Kalimera Zoi ("Goodmorning Life"), which ran from September 1993 until its cancellation in June 2006 due to low ratings. Mega Channel began producing soap operas in 1990 with the prime time serial I Dipsa ("The Thirst"), which ran for 102 episodes. Other daytime soaps have included Paralliloi dromoi (1992–1994) and its successor Haravgi ("Daylight", 1994–1995), both of which were cancelled due to low viewership; as well as the serials Apagorevmeni Agapi ("Forbidden Love"), which ran from 1998 to 2006; Gia mia thesi ston Ilio ("A Spot Under the Sun"), which ran from 1998 to 2002; Filodoxies ("Expectations"), which ran from 2002 to 2006; and Vera Sto Deksi ("Ring on the Right Hand"), which ran from 2004 to 2006 and proved to be a successful competitor to Lampsi, causing that show's ratings to decline. Ta Mistika Tis Edem ("Edem Secrets"), which was created by the producers of Vera Sto Deksi, debuted in 2008 and has eclipsed that show's success. Its ratings place it consistently among the three highest-rated daytime programs. YENED (which was renamed ERT2 in 1982) was responsible for the first Greek soap operas I Kravgi Ton Likon and Megistanes. ERT also produced the long-running soap O Simvoleografos. Since 2000 and with the introduction of private television, ERT produced additional daily soap operas, which included Pathos ("Passion"), Erotika tis Edem ("Loving in Eden") and Ta ftera tou erota ("The Wings of Love"). These failed to achieve high ratings and were canceled shortly after their premiere. Alpha produced Kato apo tin Acropoli ("Under the Acropolis"), which ran for 2½ years. In 2022, Alpha produce a new soap opera Paradeisos based on the Italian soap Il Paradiso Delle Signore. The first daytime soap opera produced by a Cyprus channel was LOGOs TV's Odos Den Ksehno ("'Don't Forget' Street"), which ran from January to December 1996. It was followed by To Serial, which also ran for one year from September 1997 to June 1998. CyBC created the third weekdaily soap, Anemi Tou Pathous ("Passion Winds"), running from January 2000 to June 2004, which was replaced by I Platia ("The Square") from September 2004 to July 2006. Epikindini Zoni ran from 2009 to 2010, and was cancelled after 120 episodes. Vimata Stin Ammo made its debut in September 2010 until 2014 and was followed by Halkina Hronia (2017-2022). Sigma TV first commissioned the weekdaily comedic soap Sto Para Pente, which aired from September 1998 to June 2004, and was the longest weekday show in Cyprus television history, before it was surpassed by Se Fonto Kokkino, which ran from September 2008 to July 2012 and then by Galateia (2016-2020). Other Sigma TV weekday shows include Akti Oniron (which ran from 1999 to 2001), Vourate Geitonoi (which ran from 2001 to 2005, and was the most successful weekday series, achieving ratings shares of up to 70% of all television households in the country), Oi Takkoi (which ran from 2002 to 2005), S' Agapo (which ran from 2001 to 2002), Vasiliki (which ran from 2005 to 2006), Vendetta (which ran from September 2005 to December 2006), 30 kai Kati (which ran from 2006 to 2007), Mila Mou (which ran from September 2007 to January 2009), 7 ouranoi ke dinnefa alites (2012-2015) and Galateia (2016-2020) ANT1 Cyprus aired the soap I Goitia Tis Amartias in 2002, which was soon canceled. Dikse Mou To Filo Sou followed from 2006 to 2009, along with Gia Tin Agapi Sou, which ran from 2008 to 2009 and itself was followed by Panselinos, which has aired 2009 to 2011. The longest-running weekly show on Cyprus television is Istories Tou Horkou ("Villages Stories", which premiered on CyBC in March 1996 and ran until its cancellation in June 2006; it was revived in September 2010 but was cancelled again in March 2011 due to very low ratings), followed by Manolis Ke Katina ("Manolis and Katina", which ran from 1995 to 2004). The most controversial of these series was To Kafenio ("The Coffee Shop"), which premiered on CyBC on 1993 as a weekly series, before moving to MEGA Channel Cyprus six years later in 1999 as a weekday show and then moved to ANT1 Cyprus on 2000, which canceled the show one year later. There were plans to move the show back to CyBC as a weekly series in 2001, with the original cast, however, this plan was never realised. The most successful weekly shows in Cyprus currently are ANT1's Eleni I Porni ("Eleni, The Whore"), which premiered in October 2010 and CyBC's Stin Akri Tu Paradisou ("At The Heaven's Edge"), which premiered in 2007. The most successful weekdaily soap was Aigia Fuxia, which aired on ANT1 Cyprus from 2008 to 2010. The only daily Finnish soap opera so far is Salatut elämät (Secret Lives), which has achieved popularity in Finland since its 1999 debut on MTV3. It focuses on the lives of people along the imaginary Pihlajakatu street in Helsinki. The show has also spawned several Internet spin-off series and a film based on the show that was released in 2012. Other soap-like shows in Finland are YLE shows Uusi päivä (which has aired from 2010 to 2018) and Kotikatu (which ran from 1995 to 2012), however these programs did not adhere to a five-episode-a-week schedule. In February 2022, MBC launched the first Egyptian daily soap opera, Downtown West El Balad. In Cairo, two brothers became enemies after the death of their father because the eldest son was excluded from the inheritance. 190 episodes have already been produced. In Latin America, for many years, primetime (as well as part of daytime) programming, for the most part, has been traditionally composed of telenovelas. However, throughout the years, there have been cases where a number of television programs tended to "mix" the concepts of television series and telenovela, such as, for example, a telenovela that lasted several seasons to end. With this "overlap", many people consider that these shows could be more accurately described as "soap operas". With this being said, the two most notable Latin American examples of TV programs that could fit on the definition of a "soap opera" are Chiquititas (in both Argentina and Brazil) and Malhação (only in Brazil). Chiquititas was first broadcast in Argentina by Telefe in 1995 and soon became a national hit, especially among children. In regards to the audience, all eight seasons (the final season ended in 2006) of Chiquititas guaranteed the first place in the Argentine TV ratings for Telefe. Throughout the years, Chiquititas had a number of spin-offs not only in Argentina, but also in Brazil, Mexico and Portugal. In 1997, Silvio Santos, founder and owner of the Brazilian television network SBT, seeing the good ratings of Chiquititas in Argentina, decided to make a partnership with Telefe, and thus, SBT started to broadcast Chiquititas in Brazil, but in the format of "remake", with the use of the Portuguese language instead of Spanish, with the use of dubbing when singing the soundtrack songs (unlike the Argentine version, on which the actors themselves sung the songs), with a Brazilian cast and with slight modifications in regards to its plot (the Brazilian version was set in the city of São Paulo instead of Buenos Aires, although many scenes of the Brazilian adaptation were actually filmed at the same Telefe studios in Buenos Aires where the Argentine version was also recorded, due to the aforementioned partnership between Telefe and SBT). The Brazilian version of Chiquititas, which lasted five seasons and ended in 2001, was successful in the ratings as well, in a slightly smaller scale compared to the Argentine version, and despite the success of Malhação (see below), the soap opera was one of the most known TV programs of the late 1990s in Brazil, enough to put Chiquititas also in the imaginary of many Brazilian children (a proof of this is that the casting process for the third season of Chiquititas in 1999 reunited about 15,000 children in the city of São Paulo alone, a record number not seen even in any Brazilian telenovela). In 2013, SBT decided to make a second adaptation of Chiquititas, which lasted two seasons (the final season ended in 2015), but unlike the first version, which resembled more like its Argentine counterpart, the second version, produced only by SBT, is different not only because the soundtrack is entirely sung by the actors themselves (as well as on the Argentine version, on which the actors did not dub the songs, but unlike the first Brazilian adaptation). Despite the fact that the ratings of the 2013 version of Chiquititas were smaller, the soap opera was not considered a failure by the critics. SBT executives evaluated the ratings as being "satisfactory", and some fans consider the 2013 version to be a small "revival" of the 1997 version of Chiquititas. Malhação has been transmitted by Rede Globo on almost every week since 1995 and has become the most successful Brazilian soap opera in the ratings. On each one of the 27 seasons shown as of 2021, the soap opera stayed in the first place on the ratings (like the Argentine version of Chiquititas). Moreover, Malhação also had a number of spin-offs being produced in Brazil. However, unlike Chiquititas, Malhação is more focused on teenagers, with more mature issues like teenage pregnancy, sexual relationships and the use of illicit drugs being discussed on its plot. Another interesting topic is that Malhação is considered by some fans as being the "entrance door" to many rookie actors who obtain the first opportunity of working on Rede Globo, because history has shown that a good acting in Malhação increases the possibility of being "promoted" to the primetime telenovelas (also broadcast by Rede Globo). In fact, estimates indicate that hundreds of actors participate in the casting process of Malhação each year, proving that many aspiring actors want to appear in this soap opera to further progress their careers. With the advent of internet television and mobile phones, several soap operas have also been produced specifically for these platforms, including EastEnders: E20, a spin-off of the established EastEnders. For those produced only for the mobile phone, episodes may generally consist of about six or seven pictures and accompanying text. On September 13, 2011, TG4 launched a new 10-part online series titled, Na Rúin (an Internet spin-off of Ros na Rún). The miniseries took on the theme of a mystery; the viewer had to read Rachel and Lorcán's blogs as well as watch video diaries detailing each character's thoughts to solve the mystery of missing teenage character Ciara. Due to the massive number of episodes typically produced for a long-running soap opera (into the tens of thousands for some) and the fact many episodes are lost over time, home video release (in VHS, DVD or Blu-ray) of daily soap operas is generally considered impractical and impossible beyond occasional retrospective releases or highlights. A notable exception is the 1966–1971 series Dark Shadows, which has had its entire run of 1,225 episodes (with an audio recreation of its sole missing episode) released to home video. In the case of American "primetime soap operas" this generally does not apply as typically such series produce far fewer episodes (generally on par with that of other genres), allowing home video release. In motion pictures, the 1982 comedy Tootsie has the lead character impersonating a woman in order to gain acting work on a long running television soap opera. Several scenes parody the production of soaps, their outrageous storylines and idiosyncratic stylistic elements. The 1991 comedy Soapdish stars Sally Field as an aging soap opera actress on the fictional series The Sun Also Sets who pines over her own neuroses and misfortunes, such as her live-in boyfriend who leaves her to go back to his wife, and the incidents of backstabbing and scheming behind the scenes, some of which are more interesting than the stories on the program. Another 1991 comedy, Delirious, stars John Candy as a soap opera writer who, after a head injury, has a dream experience of being in his own creation. The dream experience is an increasingly outrageous exaggeration of soap opera plot elements. On television, several soap opera parodies have been produced:
{{Infobox television | alt_name = Y&R | first_aired = {{Start date|1973|3|26}} | genre = [[Soap opera]] | opentheme = "[[Nadia's Theme]]"<br>by [[Barry De Vorzon]] and<br>[[Perry Botkin Jr.]] | country = United States | language = English | num_seasons = 52 | creator = [[William J. Bell]]<br>[[Lee Phillip Bell]] | network = CBS | runtime = 30 minutes (1973-1980)<br>60 minutes (1980-present) | camera = [[Videotape]]<br>[[Multi-camera]] | executive_producer = Anthony Morina<br>Josh Griffith<br>(and [[others]]) | writer = [[Josh Griffith]] | director = [[Sally McDonald]]<br>[[Owen Renfroe]]<br>[[Casey Childs]]<br>[[Conal O'Brien]]<br>[[Michael Eilbaum]]<br>[[Steven Wilford]]<br>[[See below]] | num_episodes = 13,010 | last_aired = present }} '''''The Young and the Restless''''' is a very popular [[United States|America]]n [[soap opera]]. The show is the highest-rated soap opera, which means that more people watch it than any other soap opera. It aired its first episode on March 26, 1973. It is set in [[fiction]]al [[Genoa City]], [[Wisconsin]]. In the beginning, the show was about the Brooks family, who were very rich, and the Foster family, who were poor. Today, very few members of either family are still on the show. Today, the show tells stories about two [[cosmetics]] companies: '''Newman Enterprises''' and '''Jabot Cosmetics'''. Newman's president is Victor Newman (played by the actor Eric Braeden) and Jabot's president is Jack Abbott (played by Peter Bergman). These two men hate each other very much and would like to see the other fail. This hatred is the root of much of the story on the show. ==List of actors/characters on the show== {| class=wikitable ![[Actor]]!![[Fictional character|Role]] |- | Hunter Allan || Noah Newman |- | Peter Bergman || Jack Abbott |- | Vail Bloom || Heather Stevens |- | Eric Braeden || Victor Newman |- | Tracey E. Bregman || Lauren Fenmore Baldwin |- | Bryton || Devon Hamilton |- | Hayley Erin || Abby Carlton |- | Sharon Case || Sharon Abbott |- | [[Judith Chapman]] || Gloria Bardwell |- | [[Jeanne Cooper]] || Katherine Chancellor |- | Doug Davidson || Paul Williams |- | Michael Muhney || Adam Wilson |- | Matthew and Riley Esham || Fenmore Baldwin |- | Adrienne Frantz || Amber Moore |- | Daniel Goddard || Cane Ashby |- | Michael Graziadei || Daniel Romalotti |- | Amelia Heinle || Victoria Newman Hellstrom |- | Elizabeth Hendrickson || Chloe Mitchell |- | Christel Khalil || Lily Winters |- | Christian LeBlanc || Michael Baldwin |- | Kate Linder || Esther Valentine |- | Thad Luckinbill || J.T. Hellstrom |- | Joshua Morrow || Nicholas Newman |- | Emily O'Brien || Jana Hawkes |- | Nia Peeples || Karen Taylor |- | Eyal Podell || Professor Adrian Korbel |- | Greg Rikaart || Kevin Fisher |- | Melody Thomas Scott || Nikki Newman |- | Ted Shackelford || Jeffrey Bardwell |- | Michelle Stafford || Phyllis Summers Newman |- | Kristoff St. John || Neil Winters |- | Jess Walton || Jill Foster Abbott |- | Patty Weaver || Gina Roma |} == Other websites == *{{imdb title|0069658}} {{tv-stub}} {{DEFAULTSORT:Young and the Restless, The}} [[Category:1973 American television series debuts]] [[Category:1970s American television series]] [[Category:1980s American television series]] [[Category:1990s American television series]] [[Category:2000s American television series]] [[Category:2010s American television series]] [[Category:American soap operas]] [[Category:Television series set in Wisconsin]] [[Category:CBS network shows]] [[Category:English-language television programs]]
The Young and the Restless (often abbreviated as Y&R) is an American television soap opera created by William J. Bell and Lee Phillip Bell for CBS. The show is set in fictional Genoa City (not the real-life similarly named Genoa City, Wisconsin). First broadcast on March 26, 1973, The Young and the Restless was originally broadcast as half-hour episodes, five times a week. The show expanded to one-hour episodes on February 4, 1980. In 2006, the series began airing previous episodes weeknights on SOAPnet until 2013, when it moved to TVGN (now Pop). As of July 1, 2013, Pop still airs previous episodes on weeknights. The series is also syndicated internationally. The Young and the Restless originally focused on two core families: the wealthy Brooks family and the working class Foster family. After a series of recasts and departures in the early 1980s, all the original characters except Jill Foster Abbott were written out. Bell replaced them with new core families, the Abbotts and the Williamses. Over the years, other families such as the Newman family, the Barber/Winters family, and the Baldwin-Fishers were introduced. Despite these changes, one of its most enduring storylines was the four-decade feud between Jill Abbott and Katherine Chancellor, the longest rivalry on any American soap opera. Since its television debut, The Young and the Restless has won 11 Daytime Emmy Awards for Outstanding Drama Series. It is also currently the highest-rated daytime drama on American television, a rank it has held for 34 years as of the end of the 2021–22 season. As of 2008, it had appeared at the top of the weekly Nielsen ratings in that category for more than 1,000 weeks since 1988. As of December 12, 2013, according to Nielsen ratings, The Young and the Restless marked an unprecedented 1,300 weeks, or 25 years, as the highest-rated daytime drama. The serial is also a sister series to The Bold and the Beautiful, as several actors have crossed over between shows. The series aired its 12,500th episode on May 1, 2022. Currently, the series has been renewed by CBS to run through the 2023–2024 television season. Well-known celebrities that have made an appearance on the show include Katy Perry, Lionel Richie, Wayne Gretzky, Il Divo, Enrique Iglesias, and many others. Some well-known celebrities such as Eva Longoria, David Hasselhoff, Tom Selleck, and Paul Walker got their jumpstart on The Young and the Restless. To compete with the youthful ABC soap operas, All My Children, One Life to Live, and General Hospital, CBS executives wanted a new daytime serial that was youth oriented. William J. Bell and Lee Phillip Bell created The Young and the Restless in 1972 for the network under the working title, The Innocent Years! "We were confronted with the very disturbing reality that young America had lost much of its innocence," Bell said. "Innocence as we had known and lived it all our lives had, in so many respects, ceased to exist." They changed the title of the series to The Young and the Restless because they felt it "reflected the youth and mood of the early seventies." The Bells named the fictional setting for the show after the real Genoa City, Wisconsin, a community on U.S. Route 12 in Wisconsin along the Illinois-Wisconsin state line located between their then-home in Chicago and their annual summer vacation spot in Lake Geneva. The Young and the Restless began airing on March 26, 1973, replacing the canceled soap opera, Where the Heart Is. Bell worked as head writer from the debut of the series until his retirement in 1998. He wrote from his home in Chicago while production took place in Los Angeles, California. Originally, Bell wanted to shoot the series in New York City; however, CBS executives felt that Los Angeles would be more cost effective. John Conboy acted as the show's first executive producer, staying in the position until 1982. Bell and H. Wesley Kenney became co-executive producers that year until Edward Scott took over in 1989. Bell then became senior executive producer. Other executive producers included David Shaughnessy, John F. Smith, Lynn Marie Latham, Josh Griffith, Maria Arena Bell, and Paul Rauch. In the mid-1980s, Bell and his family moved to Los Angeles to create a new soap opera. During this time, his three children, William Jr., Bradley Bell, and Lauralee Bell, each became involved in soap operas. Lauralee Bell worked as an actress on The Young and the Restless. Bradley Bell co-created The Bold and the Beautiful with his father. William Bell Jr. became involved in the family's production companies as president of Bell Dramatic Serial Co. and Bell-Phillip Television Productions Inc. "It's worked out very well for us because we really all worked in very different aspects of the show," William Bell Jr. said. "With my father and I, it was a great kind of partnership and pairing in the sense that he had a total control of the creative side of the show and I didn't have even the inclination to interject in what he was doing." After William J. Bell's 1998 retirement, a number of different head writers took over the position, including Kay Alden, Trent Jones, John F. Smith, Lynn Marie Latham, Scott Hamner, Josh Griffith, Maria Arena Bell, and Hogan Sheffer. In 2012, former General Hospital executive producer Jill Farren Phelps was hired as the new executive producer of the soap, replacing Bell. Griffith was also named the sole head writer. On August 15, 2013, it was speculated and reported by several online sources that Griffith had resigned as head-writer of the serial. Further speculation adds that Shelly Altman may take over as the new scribe, alongside Tracey Thomson or Jean Passanante may be brought aboard as co-head scribe. On September 12, 2013, it was announced that Passanante and Altman were named head writers of the show, with Thomson promoted to co-head writer. On September 18, 2014, former All My Children, Santa Barbara and General Hospital head writer Charles Pratt Jr. was named as the new head writer of the show. Passanante, Altman and Thomson have been demoted to breakdown writers. Pratt was also named as co-executive producer sharing the credit with Phelps. On June 7, 2016, Serial Scoop announced that Phelps had been terminated from her position as executive producer; a replacement was not named at the time of their reporting. The following morning, Sony Pictures Television confirmed to several list of soap opera media outlets that Phelps had been let go from her position; British television producer Mal Young was announced as Phelps' replacement. Phelps' last appearance as executive producer was July 12, 2016, while Young's first appearance occurred the following day on July 13. On September 13, 2016, it was announced that Pratt was named as executive producer and show-runner of Lee Daniels' Star. The same day, Daytime Confidential revealed that former Generations and Days of Our Lives head writer Sally Sussman, who previously had positions with the show, such as Associate Head Writer, was in-talks to replace Pratt as Head Writer. On September 15, 2016, it was confirmed that Sussman was named as the soap's new head writer. On September 21, 2016, Daytime Confidential reported that after ten years since leaving the soap, Alden had been re-hired to be a story consultant under Sussman's regime. Sussman's tenure as head writer began taping on October 20, 2016, and began airing on December 7, 2016. On June 20, 2017, CBS announced its decision to renew the serial for three years. On July 31, 2017, it was announced that both Alden and Sussman would retire from their positions; Young was named as Sussman's successor as head writer. Sussman last aired as head writer on October 24, 2017. Young's tenure as head writer aired on October 25, 2017. In December 2018, Young announced his decision to leave the serial, citing that it was a "good time to move on", and cited his desire to pursue his own project. Anthony Morina was announced as executive producer, while Griffith was named co-executive producer and head writer. On January 30, 2020, CBS announced it had renewed the serial through 2024. In a statement, CBS Entertainment president Kelly Kahl stated: "It's a remarkable achievement and a testament to the extraordinary cast, gifted writers, talented producers and supremely passionate fans, as well as our tremendous partnership with [Y&R studio] Sony Pictures Television." On March 20, 2020, after 32 years and over 1,500 consecutive weeks, The Young and the Restless was no longer the number-one soap opera in the United States, having been dethroned by The Bold and the Beautiful, which took 33 years since its 1987 debut to attain that position. The Young and the Restless then reclaimed the number-one spot the following week. On April 20, 2020, CBS announced plans to begin airing a week of vintage episodes, following the soap's shutdown, due to the COVID-19 pandemic; production on the soap would resume in the summer of 2020. On December 1 of the same year, the serial aired its 12,000th episode. On August 19, 2021, it was reported that actress Briana Thomas had filed a lawsuit against Sony Pictures Television and CBS Studios, alleging sexual harassment on the set from showrunner Tony Morina. On September 29, 2022, a day before the premiere of The Young and the Restless’ 50th Season, the show announced that they would be producing showcasts, an audio form of the show’s episodes. Episodes of the showcast will be available on Apple Podcasts, Spotify, Google Podcasts, and Stitcher. Taped at CBS Television City, studios 41 and 43 in Hollywood since its debut on March 26, 1973, the show was packaged by the distribution company Columbia Pictures Television, which has now been replaced by Sony Pictures Television. The Young and the Restless originally aired as a half-hour series on CBS and was the first soap opera to focus on the visual aspects of production, creating "a look that broke with the visual conventions of the genre." Similar to the radio serials that had preceded them, soap operas at the time primarily focused on dialogue, characters, and story, with details like sets as secondary concerns. The Young and the Restless stood out by using unique lighting techniques and camera angles, similar to Hollywood-style productions. The style of videotaping included using out-of-the-ordinary camera angles and a large number of facial close-ups with bright lighting on the actors' faces. Conboy said he used lighting to create "artistic effects". Those effects made the series look dark, shadowy, and moody. The Young and the Restless' look influenced the taping styles of other soap operas. When H. Wesley Kenney replaced Conboy as executive producer, he balanced the lighting of the scenes. Due to the success of the series, CBS and its affiliates pressured Bell to lengthen the series from 30 minutes to a full hour. Bell attributed the show's fall from number one in the Nielsen ratings to this change, since the lengthening of the show led to the departure of a number of cast members. "The issue of performing in a one-hour show had not been part of their contracts," Bell said. This forced the show to recast multiple main characters and eventually phase out the original core families in favor of new ones. The show expanded to one hour on February 4, 1980, replacing the long-running serial Love of Life. On June 8, 1981, it moved to 12:30 p.m. Eastern, the slot occupied by Search for Tomorrow since its premiere in 1951, which ultimately led to the latter show moving to NBC after a disastrous experiment in another timeslot, alienating that program's loyal viewers. It airs 11:00 a.m. on most stations in the Central, Mountain and Pacific time zones, usually as a lead-in to the local noon news for most CBS stations outside the Eastern Time Zone (though some stations in the Central Time Zone opt to air it at 11:30 a.m.). Exteriors used in the late 1980s and early 1990s (and reused years later) included locations in and around Pittsburgh, Pennsylvania, including Allegheny General Hospital, One Oxford Centre, the Duquesne Club, Hampton Township, Allegheny County, Pennsylvania and the state correctional institution - Pittsburgh. Phillip Chancellor died in the Richland, Pennsylvania area, where the police chief was not told and believed the accident really happened. On June 27, 2001, The Young and the Restless became the first daytime soap opera to be broadcast in high-definition. In September 2011, its sister soap The Bold and the Beautiful became the last soap to make the transition from Standard-definition television to High-definition television before One Life to Live ended its ABC run on January 13, 2012, and began its TOLN run online on April 29, 2013. On April 24, 2006, Soapnet began airing same-day episodes of the series. The final airing on SoapNet was on June 28, 2013. The soap has moved from SoapNet to TV Guide Network. The same day episodes begin airing on TVGN (now Pop) weeknights on July 1, 2013. Co-creators William J. Bell and Lee Phillip Bell centered The Young and the Restless around two core families, the wealthy Brooks' and the poor Fosters. Bell borrowed this technique of soap opera building from his mentor, Irna Phillips. While casting for the series, Bell and executive producer John Conboy auditioned 540 actors for the 13 main characters. They assembled the youngest group of actors ever cast on a soap opera at the time, hiring mostly unknown actors whom they considered "glamorous model types". Chemistry between actors also factored into the criteria for casting. The stories focused on the younger characters, with an emphasis in fantasy. The fantasy element was reflected in the love story between Jill Foster and the millionaire Phillip Chancellor II; the Leslie Brooks, Brad Elliot, and Lorie Brooks love triangle; and Snapper Foster's romance with Chris Brooks. Sexuality also played a major role in the stories. Formerly, soap operas did not delve into the sexual side of their romances. Bell changed that, first during his time as head writer of Days of Our Lives and again on The Young and the Restless. William Gray Espy's Snapper Foster is considered the "first to discover sex on a soap opera." During the story, the character is engaged to Chris Brooks (Trish Stewart) and having a sexual relationship with Sally McGuire (Lee Crawford). Other plots reflected sexual themes as well. For the first time in the genre, the dialogue and the story situations included explicit sexual themes such as premarital intercourse, impotence, incest, and rape. The first two rape storylines that would be told on the serial were controversial at the time as they reflected a more introspective and analytic storytelling style, the first time rape storylines would be addressed in this manner in the genre. The first, in 1973–74, revolved around the rape of Chris Brooks and the aftermath, in which she entertained (and, eventually, rejected) the idea that she was perhaps at fault for her attack. The second, in 1976, involved Chris's sister Peggy (Pamela Peters Solow) and was meant to serve as a cut-and-dried story in which no viewer could justify this attack, committed out of the blue by an authority figure. The series also explored social issues. Jennifer Brooks underwent the first mastectomy on a soap opera. Other social issue storylines included bulimia, alcoholism, and cancer. Lesbianism was also touched on with Katherine Chancellor, who flirts with Jill while drunk in 1974 and has a brief relationship with Joann Curtis (Kay Heberle) in 1977. When the series lengthened from a half-hour to an hour in 1980, multiple cast members who portrayed characters from the original core families departed because their contracts only bound them to performing in a half-hour show. A number of the characters were recast until one of the few remaining original actors, Jaime Lyn Bauer, who portrayed Lorie Brooks, decided to leave. When she announced her intention not to renew her contract, Bell decided to replace the original core families. "As I studied the remaining cast, I realized I had two characters- Paul Williams, played by Doug Davidson, and Jack Abbott, played by Terry Lester- both of whom had a relatively insignificant presence on the show," Bell said. "They didn't have families. Hell, they didn't even have bedrooms. But these became the two characters I would build our two families around." The characters from the Abbott and Williams families were integrated into the series while the Brooks and Foster families, with the exception of Jill, were phased out. The continuity of the feud between Jill and Katherine, which began in the early years of the show, smoothed the transition. The relationship between the two characters remained a central theme throughout the series and became the longest lasting rivalry in daytime history. Another character introduced in the 1980s was Eric Braeden's Victor Newman. Originally, the character was "a despicable, contemptible, unfaithful wife abuser" who was intended to be killed off. Braeden's tenure on the show was meant to last between eight and twelve weeks. "When I saw Eric Braeden's first performance- the voice, the power, the inner strength- I knew immediately that I didn't want to lose this man," Bell said. "He was exactly what the show needed. Not the hateful man we saw on-screen, but the man he could and would become." Bell rewrote the story to save the character and put Braeden on contract. Victor's romance with Nikki Reed became a prominent plot in the series. With the success of another iconic character, Kimberlin Brown's Sheila Carter, Bell successfully crossed her over from The Young and the Restless to his second soap, The Bold and the Beautiful, in 1992. The success of the crossover was due to the creativity of Bell, as the nefarious character of Sheila was presumed to have died in a fire on The Young and the Restless. In the 1990s, core black characters were introduced with the Barber and Winters families. Victoria Rowell (Drucilla Barber) and Tonya Williams (Olivia Winters) were cast as the nieces of the Abbotts' maid, Mamie Johnson, in 1990. The brothers Neil Winters (Kristoff St. John) and Malcolm Winters (Shemar Moore) were introduced as love interests for Olivia and Drucilla. The Young and the Restless became popular among black viewers, which Williams and St. John attributed to the writing for the black characters. "I play a CEO at a major corporation, that's something we don't see that often," St. John said. "And the show doesn't use the old African-American stereotypes that we have been seeing on TV, like the hustler, the pimp, the drug dealer. We have come a long way." Though the characters held prominent positions in the fictional work place of Genoa City, they had little interaction with other characters outside of their jobs. The serial has won 165 Daytime Emmys, among 360 nominations. The following list summarizes 82 Daytime Emmy awards won by The Young and the Restless: In Canada, the Global Television Network airs new episodes a day ahead of the US broadcast. Most Global stations use The Young and the Restless as a late-afternoon lead-in for their local newscasts, but times vary by market. It also airs on NTV in Newfoundland and Labrador which airs the program on a same-day-as-CBS basis. In Belize, Channel 5 Great Belize Television airs the soap, while rival Channel 7 Tropical Vision Limited also airs the soap. In Jamaica, the show formerly aired on CVMTV. In Trinidad and Tobago, the show airs weekdays on CBS and has been airing in Trinidad and Tobago since the 1980s. In 1988, 70 percent of Trinidadians and Tobagonians who had access to a television watched daily episodes of The Young and the Restless, a series that emphasized family problems, sexual intrigue, and gossip. In Australia, The Young and the Restless airs after Days of Our Lives on Arena. It previously aired on the Nine Network from April 1, 1974, to February 23, 2007, before joining the W lineup from April 2, 2007, to August 17, 2012. On July 20, 2012, it was announced that the show would move to Arena on August 20 that year after W rebranded as SoHo. Episodes are approximately one week behind those airing in the US at present. In New Zealand, The Young and the Restless aired alongside Days of Our Lives on TVNZ 1 from August 25, 1975, to April 1988 where it moved to TVNZ 2, but it returned to TVNZ 1 from 2005 to November 6, 2009. The soap was approximately four seasons behind the CBS season due to being preempted by holiday and sporting programming. In the Philippines, aired from 1987 to 1989 on ABS-CBN. In the United Kingdom, The Young and The Restless has aired on many TV channels starting in 1990, when episodes from 1987 debuted on Galaxy in a regular weekday timeslot, 14:30-15:30 (and repeated in the early hours of the following morning). When BSB merged with Sky in November 1990 to form BskyB, the soap moved to Sky's flagship channel Sky One and aired at 11:00-12:00 weekdays until the end of 1992. When BskyB's original three-year contract to air the soap expired, Sky chose not to renew it. In 1993, Y&R was picked up by satellite channel, UK Living (then known as simply Living, now Sky Witness) in a primetime timeslot 20:00-21:00, picking up where Sky left off. This lasted until late 1995. In 2007, Zone Romantica, now CBS Drama, began broadcasting the show weekdays [four years behind US]. In 2009, when CBS went into partnership with and took over the Chellozone Channels the show was relegated to just one showing a day [in favor of repeating shows such as Dynasty and Dallas in the daytime]. It was attracting around 18–20,000 viewers at midnight in the last week of its broadcast in the UK in August 2010. In Italy The Young and the Restless aired as Febbre d'amore from February 1983 to February 1984 on Italia 1, from October 1984 to June 1995 on Rete 4, in the summer of 1995 on Canale 5 and from April 1998 to October 2009 again on Rete 4. They were US episodes shown from autumn 1979 to March 1986, from November 1989 to December 1993 and from November 1998 to January 2007. In France, the show airs as Les Feux de l'Amour on free-to-air channel TF1. In Turkey, the show aired on the public broadcast network TRT under the name of Yalan Rüzgarı (meaning The Wind of Lies), utilising the abbreviation of the name of the original show. "Nadia's Theme" has been the theme music of The Young and the Restless since the show's debut in 1973. The melody, originally titled "Cotton's Dream", was composed by Barry De Vorzon and Perry Botkin Jr. for the 1971 theatrical film Bless the Beasts and Children. The melody was later renamed "Nadia's Theme" after the American Broadcasting Company television network's sports summary program Wide World of Sports lent the music for a montage of Romanian gymnast Nadia Comăneci's routines during the 1976 Summer Olympics; despite the title, Nadia never performed her floor exercises using this piece of music. Instead, she used a piano arrangement of a medley of the songs "Yes Sir, That's My Baby" and "Jump in the Line". Botkin wrote a rearranged version of the piece specifically for The Young and the Restless' debut. The song remained unchanged, save for a slight remix in 1988 and a three-year stint in the early 2000s, when an alternate, jazzier arrangement of that tune was used, using portions of the longer closing version of the original theme. An LP album was published in 1976 by A&M Records. The track list contains two titles of the French composer Michel Colombier, Rainbow and Emmanuel, a success track which he wrote in memory of the death of his son. In late September and early October 2012, upon the show's 10,000th episode, the current form of opening credits were updated. In the years prior, fans criticized them for their lack of updates and cast additions (some contract players, such as Adrienne Frantz, and Kimberlin Brown came and went without being added). In mid-February 2017, the opening title sequence was updated in anticipation of the soaps's 44th anniversary. In March 2023, the opening title sequence was updated in celebration of the soap's 50th anniversary. The Young and the Restless entered CBS' daytime lineup at 12 noon/11 a.m. Central in March 1973, a timeslot where two popular game shows, NBC's Jeopardy! (original version, hosted by Art Fleming) and ABC's Password (revival of the 1960s hit with Allen Ludden as host), vied for the top spot in the ratings. Quite a number of CBS affiliates in the Eastern Time Zone had for years been running local newscasts at noon, despite CBS giving them a 30-minute break one hour later, at 1 p.m., for that purpose. With that scenario, at first, The Young and the Restless was at the bottom of the ratings, inheriting Where the Heart Is' low numbers and affiliate clearances. However, the youthful and sexually-driven storylines, which appealed perfectly to CBS' desired key demographic of younger women, helped it to rise rapidly, surging to ninth by 1974–75 and third by 1975–76. This eventually led to both Jeopardy!'s and Password's cancellations during 1975; a succession of shows (of varying formats) on both networks failed to make any significant impact for the next five years, other than perhaps ABC's The $20,000 Pyramid, which aired at noon/11 from 1978 to 1980. Apart from a dip to sixth in 1980-81 (its first full season at a full hour, when the first half went against ABC's then-top-rated Family Feud in the Central Time Zone or the second half against the first part of All My Children in the Eastern), it has been one of the five highest-rated soaps since the 1975–76 season. After Feud's decline beginning in 1984 (it would be canceled the next year) and the eventual plateauing of AMC and NBC's Days of Our Lives, the latter of which did undergo a brief resurgence in the mid-1980s, Y&R widened the ratings lead to the point where ABC and NBC never caught up again, both eventually ceding the Noon/11 time slot back to their local stations by the 1990s. By 1988–89 it had dethroned long-time leader General Hospital as the top-rated soap, a position it held until 2020. In 2010, it marked its 1,000th consecutive week n the #1 spot for daytime dramas. During the week of December 2, 2013, the series celebrated its 25th consecutive year as the number one daytime drama. The Young and the Restless airs every weeknight on Pop, where it averaged 362,000 viewers from July to September 2013. On the week ending April 6, 2012, The Young and the Restless was watched by a new low of an average of 3,960,000 viewers for the week, beating its previous low of 4.209 million in October 2011, as well as being the only week to date below 4 million viewers. Currently, the show is still the most-watched daytime drama; and for the season 2011–12, has a household rating of 3.5, and 1.5 for the Women 18–49 demographic. As of 2008, the Tuesday episodes of The Young and the Restless on average is the most-watched daytime drama showing.
{{about|the American state of Georgia|the country in Europe also called Georgia|Georgia (country)|other meanings of "Georgia"|Georgia}} {{Infobox U.S. state | name = Georgia | official_name = State of Georgia | image_flag = Flag of the State of Georgia.svg | flag_link = Flag of Georgia (U.S. state) | image_seal = Seal of Georgia.svg | seal_link = Seal of Georgia (U.S. state) | image_map = Georgia in United States.svg | nickname = Peach State, Empire State of the South | motto = "Wisdom, Justice, Moderation" | anthem = "[[Georgia on My Mind]]" | Former = Province of Georgia | Tree = Live Oak | seat = [[Atlanta]] | LargestCity = capital<!-- Set this to "capital" if the capital and the largest city are the same. --> | LargestMetro = [[Atlanta metropolitan area|Atlanta]] | Governor = {{nowrap|[[Brian Kemp]] ([[Republican Party (United States)|R]])}} | Lieutenant Governor = {{nowrap|[[Burt Jones]] (R)}} | Legislature = [[Georgia General Assembly]] | Upperhouse = [[Georgia State Senate|Senate]] | Lowerhouse = [[Georgia House of Representatives|House of Representatives]] | Judiciary = [[Supreme Court of Georgia (U.S. state)|Supreme Court of Georgia]] | Senators = {{ublist|{{nowrap|[[Jon Ossoff]] ([[Democratic Party (United States)|D]])}} | {{nowrap|[[Raphael Warnock]] (D)}}}} | Representative = 8 Republicans<br />6 Democrats | area_total_km2 = 153,909 | area_total_sq_mi = 59,425 | area_land_km2 = 149,976 | area_land_sq_mi = 57,906 | area_water_km2 = 3,933 | area_water_sq_mi = 1,519 | area_water_percent = 2.6 | area_rank = 24th | width_mi = 230 | width_km = 370 | length_mi = 298 | length_km = 480 | elevation_m = 180 | elevation_ft = 600 | elevation_max_m = 1,458 | elevation_max_ft = 4,784 | elevation_max_point = [[Brasstown Bald]]<ref name=USGS>{{cite web|url=http://egsc.usgs.gov/isb/pubs/booklets/elvadist/elvadist.html|title=Elevations and Distances in the United States|publisher=[[United States Geological Survey]]|year=2001|access-date=October 21, 2011|url-status=dead|archive-url=https://web.archive.org/web/20120722022527/http://egsc.usgs.gov/isb/pubs/booklets/elvadist/elvadist.html|archive-date=July 22, 2012}}</ref><ref name= NAVD88>Elevation adjusted to [[North American Vertical Datum of 1988]].</ref> | elevation_min_point = Atlantic Ocean<ref name=USGS /> | elevation_min_ft = 0 | elevation_min_m = 0 | population_demonym = Georgian | population_rank = 8th | OfficialLang = [[English language|English]] | Languages = [[English language|English]] <br /> [[Spanish language|Spanish]] (7.42%) <br /> Other (2.82%) | population_as_of = 2020 | 2010Pop = 10,711,908<ref name="Bureau 2021">{{cite web | last=Bureau | first=US Census | title=2020 Census Apportionment Results | website=The United States Census Bureau | date=2021-04-26 | url=https://www.census.gov/data/tables/2020/dec/2020-apportionment-data.html | access-date=2021-04-27}}</ref> | population_density_rank = 18th | 2000DensityUS = 185.2 | 2000Density = 71.5 | MedianHouseholdIncome = $56,183<ref>{{cite web|url=http://kff.org/other/state-indicator/median-annual-income/?currentTimeframe=0|work=The Henry J. Kaiser Family Foundation|title=Median Annual Household Income|access-date=December 9, 2016}}</ref> | IncomeRank = 33rd | AdmittanceOrder = 4th | AdmittanceDate = January 2, 1788 | timezone1 = [[Eastern Time Zone|Eastern]] | utc_offset1 = −05:00 | timezone1_DST = [[Eastern Daylight Time|EDT]] | utc_offset1_DST = −04:00 | Latitude = 30.356–34.985° N | Longitude = 80.840–85.605° W | iso_code = US-GA | postal_code = GA | TradAbbreviation = Ga. | website = https://www.georgia.gov | Capital = Atlanta | Representatives = }} {{Infobox U.S. state symbols <!--Source:[List of Georgia state symbols]--> |Name = Georgia |Flag = Flag of the State of Georgia.svg |Flag link = Flag of Georgia (U.S. state) |Seal = Seal of Georgia.svg |Mammal = [[North Atlantic right whale]] |Bird = [[Brown Thrasher]] |Insect = [[Honeybee]] |Fish = [[Largemouth Bass]] |Fossil = [[Shark tooth]] |Flower = [[Cherokee Rose]] |Tree = [[Quercus virginiana|Southern Live Oak]] |Gemstone = [[Quartz]] |Reptile = [[Gopher tortoise]] |Dog breed = "[[Pet adoption|Adoptable Dog]]" |Food = [[Grits]], [[Peach]] |Route marker = Georgia 8.svg }} '''Georgia''' is a [[U.S. state|state]] in the southeastern part of the [[United States]]. It is bordered by [[Florida]] to the south, [[Alabama]] to the west, [[Tennessee]] and [[North Carolina]] to the north, and [[South Carolina]] to the east. All of Georgia's coastline is on the [[Atlantic Ocean]], which borders Georgia to the southeast. The capital of Georgia is [[Atlanta]]. Georgia was established in [[1732]] and became a state in [[1788]]. It was the last of the original 13 colonies and eventually joined in the growing rebellion against [[Britain]]. Georgia was named after [[George II of Great Britain]]. == History == Georgia was initially inhabited by [[Native Americans]] before the arrival of Europeans. The largest tribes were the [[Cherokee]] and the Creek. == Geography == There are 5 regional habitats of Georgia. They are the [[Georgia mountains|Appalachian Plateau]] region, the [[Okeefenokee marsh/swamp area|Ridge and Valley]] region, the [[the Georgia coastal region|Blue Ridge]] region, the [[the Georgia Piedmont|Piedmont]] region, and the [[Atlantic Plain]] region. === Climate === Much of Georgia is a [[humid subtropical climate]], with hot and humid summers. The northern parts of Georgia receive more rain than any other areas in the state. Winters are usually mild. == Cities in Georgia == The [[Capital (city)|capital]] and largest city in Georgia is [[Atlanta, Georgia|Atlanta]]. The [[1996]] [[Olympic Games]] took place there, and the book and movie ''[[Gone with the Wind]]'' was set in and around Atlanta in the [[1860s]]. Other big cities in the state include [[Augusta, Georgia|Augusta]], [[Columbus, Georgia|Columbus]], [[Savannah, Georgia|Savannah]], and [[Macon, Georgia|Macon]]. Georgia has a total of 159 counties. <gallery perrow="5"> Image:Cnncenter.jpg|<center>CNN Center</center> Image:Georgia Aquarium Jan 2006.jpg|<center>Georgia Aquarium</center> Image:Atlanta cityscape.jpg|<center>Atlanta</center> </gallery>{{notelist}} ==Related pages== * [[List of counties in Georgia (U.S. state)|List of counties in Georgia]] * [[List of rivers of Georgia (U.S. state)|List of rivers of Georgia]] == References == {{commonscat|Georgia (U.S. state)}} {{reflist}} {{United States}} {{Georgia (U.S. state)}} {{US-geo-stub}} [[Category:Georgia (U.S. state)| ]] [[Category:1788 establishments in the United States]]
Georgia is a state in the Southeastern region of the United States, bordered to the north by Tennessee and North Carolina; to the northeast by South Carolina; to the southeast by the Atlantic Ocean; to the south by Florida; and to the west by Alabama. Georgia is the 24th-largest state in area and 8th most populous of the 50 United States. Its 2020 population was 10,711,908, according to the U.S. Census Bureau. Atlanta, a "beta(+)" global city, is both the state's capital and its largest city. The Atlanta metropolitan area, with a population of more than 6 million people in 2021, is the 8th most populous metropolitan area in the United States and contains about 57% of Georgia's entire population. Other major metropolitan areas in the state include Augusta, Savannah, Columbus, and Macon. Founded in 1732 as the Province of Georgia and first settled in 1733, Georgia became a British royal colony in 1752. It was the last and southernmost of the original Thirteen Colonies to be established. Named after King George II of Great Britain, the Georgia Colony covered the area from South Carolina south to Spanish Florida and west to French Louisiana at the Mississippi River. On January 2, 1788, Georgia became the fourth state to ratify the United States Constitution. From 1802 to 1804, western Georgia was split to form the Mississippi Territory, which later was admitted as the U.S. states of Alabama and Mississippi. Georgia declared its secession from the Union on January 19, 1861, and was one of the original seven Confederate States. Following the Civil War, it was the last state to be restored to the Union, on July 15, 1870. In the post-Reconstruction era of the late 19th century, Georgia's economy was transformed as a group of prominent politicians, businessmen, and journalists, led by Henry W. Grady, espoused the "New South" philosophy of sectional reconciliation and industrialization. During the mid-20th century, several people from Georgia, most notably Martin Luther King Jr., were prominent leaders during the civil rights movement. Atlanta was selected as host of the 1996 Summer Olympics, which marked the 100th anniversary of the modern Olympic Games. Since 1945, Georgia has seen substantial population and economic growth as part of the broader Sun Belt phenomenon. From 2007 to 2008, 14 of Georgia's counties ranked among the nation's 100 fastest-growing. Georgia is defined by a diversity of landscapes, flora, and fauna. The state's northernmost regions include the Blue Ridge Mountains, part of the larger Appalachian Mountain system. The Piedmont plateau extends from the foothills of the Blue Ridge south to the Fall Line, an escarpment to the Coastal Plain defining the state's southern region. Georgia's highest point is Brasstown Bald at 4,784 feet (1,458 m) above sea level; the lowest is the Atlantic Ocean. With the exception of some high-altitude areas in the Blue Ridge, the entirety of the state has a humid subtropical climate. Of the states entirely east of the Mississippi River, Georgia is the largest in land area. Before settlement by Europeans, Georgia was inhabited by the mound building cultures. The British colony of Georgia was founded by James Oglethorpe on February 12, 1733. The colony was administered by the Trustees for the Establishment of the Colony of Georgia in America under a charter issued by (and named for) King George II. The Trustees implemented an elaborate plan for the colony's settlement, known as the Oglethorpe Plan, which envisioned an agrarian society of yeoman farmers and prohibited slavery. The colony was invaded by the Spanish in 1742, during the War of Jenkins' Ear. In 1752, after the government failed to renew subsidies that had helped support the colony, the Trustees turned over control to the crown. Georgia became a crown colony, with a governor appointed by the king. The Province of Georgia was one of the Thirteen Colonies that revolted against British rule in the American Revolution by signing the 1776 Declaration of Independence. The State of Georgia's first constitution was ratified in February 1777. Georgia was the 10th state to ratify the Articles of Confederation on July 24, 1778, and was the 4th state to ratify the United States Constitution on January 2, 1788. After the Creek War (1813–1814), General Andrew Jackson forced the Muscogee (Creek) tribes to surrender land to the state of Georgia, including in the Treaty of Fort Jackson (1814), surrendering 21 million acres in what is now southern Georgia and central Alabama, and the Treaty of Indian Springs (1825). In 1829, gold was discovered in the North Georgia mountains leading to the Georgia Gold Rush and establishment of a federal mint in Dahlonega, which continued in operation until 1861. The resulting influx of white settlers put pressure on the government to take land from the Cherokee Nation. In 1830, President Andrew Jackson signed the Indian Removal Act, sending many eastern Native American nations to reservations in present-day Oklahoma, including all of Georgia's tribes. Despite the Supreme Court's ruling in Worcester v. Georgia (1832) that U.S. states were not permitted to redraw Indian boundaries, President Jackson and the state of Georgia ignored the ruling. In 1838, his successor, Martin Van Buren, dispatched federal troops to gather the tribes and deport them west of the Mississippi. This forced relocation, known as the Trail of Tears, led to the death of more than four thousand Cherokees. In early 1861, Georgia joined the Confederacy (with secessionists having a slight majority of delegates) and became a major theater of the Civil War. Major battles took place at Chickamauga, Kennesaw Mountain, and Atlanta. In December 1864, a large swath of the state from Atlanta to Savannah was destroyed during General William Tecumseh Sherman's March to the Sea. 18,253 Georgian soldiers died in service, roughly one of every five who served. In 1870, following the Reconstruction era, Georgia became the last Confederate state to be restored to the Union. With white Democrats having regained power in the state legislature, they passed a poll tax in 1877, which disenfranchised many poor black (and some white) people, preventing them from registering. In 1908, the state established a white primary; with the only competitive contests within the Democratic Party, it was another way to exclude black people from politics. They constituted 46.7% of the state's population in 1900, but the proportion of Georgia's population that was African American dropped thereafter to 28%, primarily due to tens of thousands leaving the state during the Great Migration. According to the Equal Justice Initiative's 2015 report on lynching in the United States (1877–1950), Georgia had 531 deaths, the second-highest total of these extralegal executions of any state in the South. The overwhelming number of victims were black and male. Political disfranchisement persisted through the mid-1960s, until after Congress passed the Voting Rights Act of 1965. An Atlanta-born Baptist minister who was part of the educated middle class that had developed in Atlanta's African-American community, Martin Luther King Jr., emerged as a national leader in the civil rights movement. King joined with others to form the Southern Christian Leadership Conference (SCLC) in Atlanta in 1957 to provide political leadership for the Civil Rights Movement across the South. The civil rights riots of the 1956 Sugar Bowl would also take place in Atlanta after a clash between Georgia Tech's president Blake R. Van Leer and Governor Marvin Griffin. On February 5, 1958, during a training mission flown by a B-47, a Mark 15 nuclear bomb, also known as the Tybee Bomb, was lost off the coast of Tybee Island near Savannah. The bomb was thought by the Department of Energy to lie buried in silt at the bottom of Wassaw Sound. By the 1960s, the proportion of African Americans in Georgia had declined to 28% of the state's population, after waves of migration to the North and some immigration by whites. With their voting power diminished, it took some years for African Americans to win a state-wide office. Julian Bond, a noted civil rights leader, was elected to the state House in 1965, and served multiple terms there and in the state senate. Atlanta Mayor Ivan Allen Jr. testified before Congress in support of the Civil Rights Act, and Governor Carl Sanders worked with the Kennedy administration to ensure the state's compliance. Ralph McGill, editor and syndicated columnist at the Atlanta Constitution, earned admiration by writing in support of the Civil Rights Movement. In 1970, newly elected Governor Jimmy Carter declared in his inaugural address that the era of racial segregation had ended. In 1972, Georgians elected Andrew Young to Congress as the first African American Congressman since the Reconstruction era. In 1980, construction was completed on an expansion of what is now named Hartsfield–Jackson Atlanta International Airport (ATL). The busiest and most efficient airport in the world, it accommodates more than a hundred million passengers annually. Employing more than 60,000 people, the airport became a major engine for economic growth. With the advantages of cheap real estate, low taxes, right-to-work laws and a regulatory environment limiting government interference, the Atlanta metropolitan area became a national center of finance, insurance, technology, manufacturing, real estate, logistics, and transportation companies, as well as the film, convention, and trade show businesses. As a testament to the city's growing international profile, in 1990 the International Olympic Committee selected Atlanta as the site of the 1996 Summer Olympics. Taking advantage of Atlanta's status as a transportation hub, in 1991 UPS established its headquarters in the suburb of Sandy Springs. In 1992, construction finished on Bank of America Plaza, the tallest building in the U.S. outside of New York or Chicago. Beginning from the Atlantic Ocean, the state's eastern border with South Carolina runs up the Savannah River, northwest to its origin at the confluence of the Tugaloo and Seneca Rivers. It then continues up the Tugaloo (originally Tugalo) and into the Chattooga River, its most significant tributary. These bounds were decided in the 1797 Treaty of Beaufort, and tested in the U.S. Supreme Court in the two Georgia v. South Carolina cases in 1923 and 1989. The border then takes a sharp turn around the tip of Rabun County, at latitude 35°N, though from this point it diverges slightly south (due to inaccuracies in the original survey, conducted in 1818). This northern border was originally the Georgia and North Carolina border all the way to the Mississippi River, until Tennessee was divided from North Carolina, and the Yazoo companies induced the legislature of Georgia to pass an act, approved by the governor in 1795, to sell the greater part of Georgia's territory presently comprising Alabama and Mississippi. The state's western border runs in a straight line south-southeastward from a point southwest of Chattanooga, to meet the Chattahoochee River near West Point. It continues downriver to the point where it joins the Flint River (the confluence of the two forming Florida's Apalachicola River); the southern border goes almost due east and very slightly south, in a straight line to the St. Mary's River, which then forms the remainder of the boundary back to the ocean. The water boundaries are still set to be the original thalweg of the rivers. Since then, several have been inundated by lakes created by dams, including the Apalachicola/Chattahoochee/Flint point now under Lake Seminole. An 1818 survey erroneously placed Georgia's border with Tennessee one mile (1.6 km) south of the intended location of the 35th parallel north. State legislators still dispute this placement, as correction of this inaccuracy would allow Georgia access to water from the Tennessee River. Georgia consists of five principal physiographic regions: The Cumberland Plateau, Ridge-and-Valley Appalachians, Blue Ridge Mountains, Piedmont, and the Atlantic coastal plain. Each region has its own distinctive characteristics. For instance, the region, which lies in the northwest corner of the state, includes limestone, sandstone, shale, and other sedimentary rocks, which have yielded construction-grade limestone, barite, ocher, and small amounts of coal. The state of Georgia has approximately 250 tree species and 58 protected plants. Georgia's native trees include red cedar, a variety of pines, oaks, hollies, cypress, sweetgum, scaly-bark and white hickories, and sabal palmetto. East Georgia is in the subtropical coniferous forest biome and conifer species as other broadleaf evergreen flora make up the majority of the southern and coastal regions. Yellow jasmine and mountain laurel make up just a few of the flowering shrubs in the state. White-tailed deer are found in nearly all counties of Georgia. The northern mockingbird and brown thrasher are among the 160 bird species that live in the state. Reptiles include the eastern diamondback, copperhead, and cottonmouth snakes as well as alligators; amphibians include salamanders, frogs and toads. There are about 79 species of reptile and 63 amphibians known to live in Georgia. The Argentine black and white tegu is currently an invasive species in Georgia. It poses a problem to local wildlife by chasing down and killing many native species and dominating habitats. The most popular freshwater game fish are trout, bream, bass, and catfish, all but the last of which are produced in state hatcheries for restocking. Popular saltwater game fish include red drum, spotted seatrout, flounder, and tarpon. Porpoises, whales, shrimp, oysters, and blue crabs are found inshore and offshore of the Georgia coast. The majority of the state is primarily a humid subtropical climate. Hot and humid summers are typical, except at the highest elevations. The entire state, including the North Georgia mountains, receives moderate to heavy precipitation, which varies from 45 inches (1,100 mm) in central Georgia to approximately 75 inches (1,900 mm) around the northeast part of the state. The degree to which the weather of a certain region of Georgia is subtropical depends on the latitude, its proximity to the Atlantic Ocean or Gulf of Mexico, and the elevation. The latter factor is felt chiefly in the mountainous areas of the northern part of the state, which are farther away from the ocean and can be 4,500 feet (1,400 m) above sea level. The USDA plant hardiness zones for Georgia range from zone 6b (no colder than −5 °F (−21 °C)) in the Blue Ridge Mountains to zone 8b (no colder than 15 °F (−9 °C) ) along the Atlantic coast and Florida border. The highest temperature ever recorded is 112 °F (44 °C) in Louisville on July 24, 1952, while the lowest is −17 °F (−27 °C) in northern Floyd County on January 27, 1940. Georgia is one of the leading states in frequency of tornadoes, though they are rarely stronger than EF1. Although tornadoes striking the city are very rare, an EF2 tornado hit downtown Atlanta on March 14, 2008, causing moderate to severe damage to various buildings. With a coastline on the Atlantic Ocean, Georgia is also vulnerable to hurricanes, although direct hurricane strikes were rare during the 20th century. Georgia often is affected by hurricanes that strike the Florida Panhandle, weaken over land, and bring strong tropical storm winds and heavy rain to the interior, a recent example being Hurricane Michael, as well as hurricanes that come close to the Georgia coastline, brushing the coast on their way north without ever making landfall. Hurricane Matthew of 2016 and Hurricane Dorian of 2019 did just that. Due to anthropogenic climate change, the climate of Georgia is warming. This is already causing major disruption, for example, from sea level rise (Georgia is more vulnerable to it than many other states because its land is sinking) and further warming will increase it. Atlanta, located in north-central Georgia at the Eastern Continental Divide, has been Georgia's capital city since 1868. It is the most populous city in Georgia, with a 2020 U.S. census population of just over 498,000. The state has seventeen cities with populations over 50,000, based on official 2020 U.S. census data. Along with the rest of the Southeast, Georgia's population continues to grow rapidly, with primary gains concentrated in urban areas. The U.S. Census Bureau lists fourteen metropolitan areas in the state. The population of the Atlanta metropolitan area added 1.23 million people (24%) between 2000 and 2010, and Atlanta rose in rank from the eleventh-largest metropolitan area in the United States to the ninth-largest. The Atlanta metropolitan area is the cultural and economic center of the Southeast; its official population in 2020 was over 6 million, or 57% of Georgia's total population. The United States Census Bureau reported Georgia's official population to be 10,711,908 as of the 2020 United States census. This was an increase of 1,024,255, or 10.57% over the 2010 figure of 9,687,653 residents. Immigration resulted in a net increase of 228,415 people, and migration within the country produced a net increase of 378,258 people. As of 2010, the number of illegal immigrants living in Georgia more than doubled to 480,000 from January 2000 to January 2009, according to a federal report. That gave Georgia the greatest percentage increase among the 10 states with the biggest undocumented immigrant populations during those years. Georgia has banned sanctuary cities. In 2018, The top countries of origin for Georgia's immigrants were Mexico, India, Jamaica, Korea, and Guatemala. There were 743,000 veterans in 2009. According to HUD's 2022 Annual Homeless Assessment Report, there were an estimated 10,689 homeless people in Georgia. In the 1980 census, 1,584,303 people from Georgia claimed English ancestry out of a total state population of 3,994,817, making them 40% of the state, and the largest ethnic group at the time. Today, many of these same people claiming they are of "American" ancestry are actually of English descent, and some are of Scots-Irish descent; however, their families have lived in the state for so long, in many cases since the colonial period, that they choose to identify simply as having "American" ancestry or do not in fact know their own ancestry. Their ancestry primarily goes back to the original thirteen colonies and for this reason many of them today simply claim "American" ancestry, though they are of predominantly English ancestry. Historically, about half of Georgia's population was composed of African Americans who, before the American Civil War, were almost exclusively enslaved. The Great Migration of hundreds of thousands of blacks from the rural South to the industrial North from 1914 to 1970 reduced the African American population. Georgia had the second-fastest-growing Asian population growth in the U.S. from 1990 to 2000, more than doubling in size during the ten-year period. Indian people and Chinese people are the largest Asian groups in Georgia. In addition, according to census estimates, Georgia ranks third among the states in terms of the percent of the total population that is African American (after Mississippi and Louisiana) and third in numeric Black population after New York and Florida. Georgia also has a sizeable Latino population. Many are of Mexican descent. Georgia is the state with the third-lowest percentage of older people (65 or older), at 12.8 percent (as of 2015). The colonial settlement of large numbers of Scottish American, English American and Scotch-Irish Americans in the mountains and Piedmont, and coastal settlement by some English Americans and African Americans, have strongly influenced the state's culture in food, language and music. The concentration of African slaves repeatedly "imported" to coastal areas in the 18th century from rice-growing regions of West Africa led to the development of Gullah-Geechee language and culture in the Low Country among African Americans. They share a unique heritage in which many African traditions of food, religion and culture were retained. In the creolization of Southern culture, their foodways became an integral part of Low Country cooking. As of 2011, 58.8% of Georgia's population younger than 1 were minorities (meaning they had at least one parent who was not non-Hispanic white) compared to other states like California with 75.1%, Texas with 69.8%, and New York with 55.6%. The largest European ancestry groups as of 2011 were: English 8.1%, Irish 8.1%, and German 7.2%. As of 2021, 85.62% (8,711,102) of Georgia residents age 5 and older spoke English at home as a primary language, while 7.82% (795,646) spoke Spanish, and 6.55% (666,849) spoke languages other than English or Spanish at home, with the most common of which were Vietnamese, Chinese, and Korean. In total, 14.38% (1,462,495) of Georgia's population age 5 and older spoke a mother language other than English. According to the Pew Research Center, the composition of religious affiliation in Georgia was 70% Protestant, 9% Catholic, 1% Mormon, 1% Jewish, 0.5% Muslim, 0.5% Buddhist, and 0.5% Hindu. Atheists, deists, agnostics, and other unaffiliated people make up 13% of the population. Overall, Christianity was the dominant religion in the state, as part of the Bible Belt. According to the Association of Religion Data Archives in 2010, the largest Christian denominations by number of adherents were the Southern Baptist Convention with 1,759,317; the United Methodist Church with 619,394; and the Roman Catholic Church with 596,384. Non-denominational Evangelical Protestant had 566,782 members, the Church of God (Cleveland, Tennessee) has 175,184 members, and the National Baptist Convention, USA, Inc. has 172,982 members. The Presbyterian Church (USA) is the largest Presbyterian body in the state, with 300 congregations and 100,000 members. The other large body, Presbyterian Church in America, had at its founding date 14 congregations and 2,800 members; in 2010 it counted 139 congregations and 32,000 members. The Roman Catholic Church is noteworthy in Georgia's urban areas, and includes the Archdiocese of Atlanta and the Diocese of Savannah. Georgia is home to the second-largest Hindu temple in the United States, the BAPS Shri Swaminarayan Mandir Atlanta, located in the suburb city of Lilburn. Georgia is home to several historic synagogues including The Temple (Atlanta), Congregation Beth Jacob (Atlanta), and Congregation Mickve Israel (Savannah). Chabad and the Rohr Jewish Learning Institute are also active in the state. By the 2022 Public Religion Research Institute's study, 71% of the population were Christian; throughout its Christian population, 60% were Protestant and 8% were Catholic. Jehovah's Witnesses and Mormons collectively made up 3% of other Christians according to the study. Judaism, Islam, Buddhism, and Hinduism collectively formed 4% of the state's non-Christian population; New Age spirituality was 2% of the religious population. Approximately 23% of the state was irreligious. Georgia's 2018 total gross state product was $602 billion. For years Georgia as a state has had the highest credit rating by Standard & Poor's (AAA) and is one of only 15 states with a AAA rating. If Georgia were a stand-alone country, it would be the 28th-largest economy in the world, based on data from 2005. There are 16 Fortune 500 companies and 26 Fortune 1000 companies with headquarters in Georgia, including Home Depot, UPS, Coca-Cola, TSYS, Delta Air Lines, Aflac, Southern Company, and Elevance Health Atlanta boasts the world's busiest airport, as measured both by passenger traffic and by aircraft traffic. Also, the Port of Savannah is the fourth-largest seaport and fastest-growing container seaport in North America, importing and exporting a total of 2.3 million TEUs per year. Atlanta has a significant effect on the state of Georgia, the Southeastern United States, and beyond. It has been the site of growth in finance, insurance, technology, manufacturing, real estate, service, logistics, transportation, film, communications, convention and trade show businesses and industries, while tourism is important to the economy. Atlanta is a global city, also called world city or sometimes alpha city or world center, as a city generally considered to be an important node in the global economic system. For the five years through November 2017, Georgia has been ranked the top state (number 1) in the nation to do business, and has been recognized as number 1 for business and labor climate in the nation, number 1 in business climate in the nation, number 1 in the nation in workforce training and as having a "Best in Class" state economic development agency. In 2016, Georgia had a median annual income per person of between $50,000 and $59,999, which is in inflation-adjusted dollars for 2016. The U.S. median annual income for the entire nation is $57,617. This lies within the range of Georgia's median annual income. Widespread farms produce peanuts, corn, and soybeans across middle and south Georgia. The state is the number one producer of pecans in the world, thanks to Naomi Chapman Woodroof regarding peanut breeding, with the region around Albany in southwest Georgia being the center of Georgia's pecan production. Gainesville in northeast Georgia touts itself as the Poultry Capital of the World. Georgia is in the top five blueberry producers in the United States. Major products in the mineral industry include a variety of clays, stones, sands and the clay palygorskite, known as attapulgite. While many textile jobs moved overseas, there is still a textile industry located around the cities of Rome, Columbus, Augusta, Macon and along the I-75 corridor between Atlanta and Chattanooga, Tennessee. Historically it started along the fall line in the Piedmont, where factories were powered by waterfalls and rivers. It includes the towns of Cartersville, Calhoun, Ringgold and Dalton In November 2009, the South Korean automaker Kia Corporation began production in Georgia. The first Kia plant built in the U.S., Kia Motors Manufacturing Georgia, is located in West Point. Rivian, an electric vehicle manufacturer, plans to begin production at a facility in Social Circle in 2024. Industrial products include textiles and apparel, transportation equipment, food processing, paper products, chemicals and products, and electric equipment. Georgia was ranked the number 2 state for infrastructure and global access by Area Development magazine. The Georgia Ports Authority owns and operates four ports in the state: Port of Savannah, Port of Brunswick, Port Bainbridge, and Port Columbus. The Port of Savannah is the third-busiest seaport in the United States, importing and exporting a total of 2.3 million TEUs per year. The Port of Savannah's Garden City Terminal is the largest single container terminal in North America. Several major companies including Target, IKEA, and Heineken operate distribution centers in close proximity to the Port of Savannah. Hartsfield–Jackson Atlanta International Airport moves over 650,000 tons of cargo annually through three cargo complexes (2 million square feet or 200,000 square meters of floor space). It has nearby cold storage for perishables; it is the only airport in the Southeast with USDA-approved cold-treatment capabilities. Delta Air Lines also offers an on-airport refrigeration facility for perishable cargo, and a 250-acre Foreign Trade Zone is located at the airport. Georgia is a major railway hub, has the most extensive rail system in the Southeast, and has the service of two Class I railroads, CSX and Norfolk Southern, plus 24 short-line railroads. Georgia is ranked the No. 3 state in the nation for rail accessibility. Rail shipments include intermodal, bulk, automotive and every other type of shipment. Georgia has an extensive interstate highway system including 1,200 miles (1,900 kilometers) of interstate highway and 20,000 miles (32,000 kilometers) of federal and state highways that facilitate the efficient movement of more than $620 billion of cargo by truck each year. Georgia's six interstates connect to 80 percent of the U.S. population within a two-day truck drive. More than $14 billion in funding has been approved for new roadway infrastructure. Southern Congressmen have attracted major investment by the U.S. military in the state. The several installations include Moody Air Force Base, Fort Stewart, Hunter Army Airfield, Naval Submarine Base Kings Bay, Fort Moore, Robins Air Force Base, Fort Eisenhower, Marine Corps Logistics Base Albany, Dobbins Air Reserve Base, Coast Guard Air Station Savannah and Coast Guard Station Brunswick. These installations command numerous jobs and business for related contractors. Georgia's electricity generation and consumption are among the highest in the United States, with natural gas being the primary electrical generation fuel, followed by coal. The state also has two nuclear power facilities, Plant Hatch and Plant Vogtle, which contribute almost one fourth of Georgia's electricity generation, and two additional nuclear reactors are being built at Vogtle as of 2022. In 2013, the generation mix was 39% gas, 35% coal, 23% nuclear, 3% hydro and other renewable sources. The leading area of energy consumption is the industrial sector because Georgia "is a leader in the energy-intensive wood and paper products industry". Solar generated energy is becoming more in use with solar energy generators currently installed ranking Georgia 15th in the country in installed solar capacity. In 2013, $189 million was invested in Georgia to install solar for home, business and utility use representing a 795% increase over the previous year. Georgia has a progressive income tax structure with six brackets of state income tax rates that range from 1% to 6%. In 2009, Georgians paid 9% of their income in state and local taxes, compared to the U.S. average of 9.8% of income. This ranks Georgia 25th among the states for total state and local tax burden. The state sales tax in Georgia is 4% with additional percentages added through local options (e.g. special-purpose local-option sales tax or SPLOST), but there is no sales tax on prescription drugs, certain medical devices, or food items for home consumption. The state legislature may allow municipalities to institute local sales taxes and special local taxes, such as the 2% SPLOST tax and the 1% sales tax for MARTA serviced counties. Excise taxes are levied on alcohol, tobacco, and motor fuel. Owners of real property in Georgia pay property tax to their county. All taxes are collected by the Georgia Department of Revenue and then properly distributed according to any agreements that each county has with its cities. The Georgia Film, Music and Digital Entertainment Office promotes filming in the state. Since 1972, seven hundred film and television projects have been filmed on location in Georgia. Georgia overtook California in 2016 as the state with the most feature films produced on location. In fiscal year 2017, film and television production in Georgia had an economic impact of $9.5 billion. Atlanta has been called the "Hollywood of the South". Television shows like Stranger Things, The Walking Dead, and The Vampire Diaries are filmed in the state. Movies too, such as Passengers, Forrest Gump, Contagion, Hidden Figures, Sully, Baby Driver, The Hunger Games: Catching Fire, Captain America: Civil War, Black Panther, Birds of Prey and many more, were filmed around Georgia. In the Atlanta area, World of Coke, Georgia Aquarium, Zoo Atlanta and Stone Mountain are important tourist attractions. Stone Mountain is Georgia's "most popular attraction"; receiving more than four million tourists per year. The Georgia Aquarium, in Atlanta, was the largest aquarium in the world in 2010 according to Guinness World Records. Callaway Gardens, in western Georgia, is a family resort. The area is also popular with golfers. The Savannah Historic District attracts more than eleven million tourists each year. The Golden Isles is a string of barrier islands off the Atlantic coast of Georgia near Brunswick that includes beaches, golf courses and the Cumberland Island National Seashore. Several sites honor the lives and careers of noted American leaders: the Little White House in Warm Springs, which served as the summer residence of President Franklin Delano Roosevelt while he was being treated for polio; President Jimmy Carter's hometown of Plains and the Carter Presidential Center in Atlanta; the Martin Luther King Jr. National Historical Park in Atlanta, which is the final resting place of Martin Luther King Jr. and Coretta Scott King; and Atlanta's Ebenezer Baptist Church, where Dr. King preached. Authors have grappled with Georgia's complex history. Popular novels related to this include Margaret Mitchell's Gone with the Wind, Olive Ann Burns' Cold Sassy Tree, and Alice Walker's The Color Purple. A number of noted authors, poets and playwrights have lived in Georgia, such as James Dickey, Flannery O'Connor, Sidney Lanier, Frank Yerby and Lewis Grizzard. Well-known television shows set in Atlanta include, from Tyler Perry Studios, House of Payne and Tyler Perry's Meet the Browns, The Real Housewives of Atlanta, the CBS sitcom Designing Women, Matlock, the popular AMC series The Walking Dead, FX comedy drama Atlanta, Lifetime's Drop Dead Diva, Rectify and numerous HGTV original productions. The Dukes of Hazzard, a 1980s TV show, was set in the fictional Hazzard County, Georgia. The first five episodes were shot on location in Conyers and Covington, Georgia as well as some locations in Atlanta. Production was then moved to Burbank, California. Also filmed in Georgia is The Vampire Diaries, using Covington as the setting for the fictional Mystic Falls. A number of notable musicians in various genres of popular music are from Georgia. Among them are Ray Charles (whose many hits include "Georgia on My Mind", now the official state song), and Gladys Knight (known for her Georgia-themed song, "Midnight Train to Georgia"). Rock groups from Georgia include the Atlanta Rhythm Section, The Black Crowes, and The Allman Brothers. The city of Athens sparked an influential rock music scene in the 1980s and 1990s. Among the groups achieving their initial prominence there were R.E.M., Widespread Panic, and the B-52's. Since the 1990s, various hip-hop and R&B musicians have included top-selling artists such as Outkast, Usher, Ludacris, TLC, B.o.B., and Ciara. Atlanta is mentioned in a number of these artists' tracks, such as Usher's "A-Town Down" reference in his 2004 hit "Yeah!" (which also features Atlanta artists Lil Jon and Ludacris), Ludacris' "Welcome to Atlanta", Outkast's album "ATLiens", and B.o.B.'s multiple references to Decatur, Georgia, as in his hit song "Strange Clouds". Two movies, both set in Atlanta, won Oscars for Best Picture: Gone with the Wind (1939) and Driving Miss Daisy (1989). Other films set in Georgia include Deliverance (1972), Parental Guidance (2012), and Vacation (2015). Sports in Georgia include professional teams in nearly all major sports, Olympic Games contenders and medalists, collegiate teams in major and small-school conferences and associations, and active amateur teams and individual sports. The state of Georgia has teams in four major professional leagues—the Atlanta Braves of Major League Baseball, the Atlanta Falcons of the National Football League, the Atlanta Hawks of the National Basketball Association, and Atlanta United FC of Major League Soccer. The Georgia Bulldogs (Southeastern Conference), Georgia Tech Yellow Jackets (Atlantic Coast Conference), Georgia State Panthers and Georgia Southern Eagles (Sun Belt Conference) are Georgia's NCAA Division I FBS football teams, having won multiple national championships between them. The Georgia Bulldogs and the Georgia Tech Yellow Jackets have a historical rivalry in college football known as Clean, Old-Fashioned Hate, and the Georgia State Panthers and the Georgia Southern Eagles have recently developed their own rivalry. The 1996 Summer Olympics took place in Atlanta. The stadium that was built to host various Olympic events was converted to Turner Field, home of the Atlanta Braves through 2016. Atlanta will serve as a host city for the 2026 FIFA World Cup. The Masters golf tournament, the first of the PGA Tour's four "majors", is held annually the second weekend of April at the Augusta National Golf Club. The RSM Classic is a golf tournament on the PGA Tour, played in the autumn in Saint Simons Island, Georgia. The Atlanta Motor Speedway hosts the Dixie 500 NASCAR Cup Series stock car race and Road Atlanta the Petit Le Mans endurance sports car race. Atlanta's Georgia Dome hosted Super Bowl XXVIII in 1994 and Super Bowl XXXIV in 2000. The dome has hosted the NCAA Final Four Men's Basketball National Championship in 2002, 2007, and 2013. It hosted WWE's WrestleMania XXVII in 2011, an event which set an attendance record of 71,617. The venue was also the site of the annual Chick-fil-A Peach Bowl post-season college football games. Since 2017, they have been held at the Mercedes-Benz Stadium along with the FIRST World Championships. Professional baseball's Ty Cobb was the first player inducted into the Baseball Hall of Fame. He was from Narrows, Georgia and was nicknamed the "Georgia Peach". The Mercedes-Benz Stadium hosted Super Bowl LIII in 2018 and the CFP National Championship in the same year, the SEC Championship Game in 2017, the MLS All-Star Game in 2018, the MLS Cup in 2018, and the record-setting friendly fixture between Mexico Men's National Football Team and Honduras Men's National Football Team. The people of Georgia have been named 'Peaches' after the amount of peaches grown and distributed from Georgia. Georgia's major fine art museums include the High Museum of Art and the Michael C. Carlos Museum, both in Atlanta; the Georgia Museum of Art on the campus of the University of Georgia in Athens; Telfair Museum of Art and the SCAD Museum of Art in Savannah; and the Morris Museum of Art in Augusta. The state theatre of Georgia is the Springer Opera House located in Columbus. The Atlanta Opera brings opera to Georgia stages. The Atlanta Symphony Orchestra is the most widely recognized orchestra and largest arts organization in the southeastern United States. There are a number of performing arts venues in the state, among the largest are the Fox Theatre, and the Alliance Theatre at the Woodruff Arts Center, both on Peachtree Street in Midtown Atlanta as well as the Cobb Energy Performing Arts Centre, located in Northwest Atlanta. Georgia county and city public school systems are administered by school boards with members elected at the local level. As of 2013, all but 19 of 181 boards are elected from single-member districts. Residents and activist groups in Fayette County sued the board of commissioners and school board for maintaining an election system based on at-large voting, which tended to increase the power of the majority and effectively prevented minority participation on elected local boards for nearly 200 years. A change to single-member districts has resulted in the African-American minority being able to elect representatives of its choice. Georgia high schools (grades nine through twelve) are required to administer a standardized, multiple choice End of Course Test, or EOCT, in each of eight core subjects: algebra, geometry, U.S. history, economics, biology, physical science, ninth grade literature and composition, and American literature. The official purpose of the tests is to assess "specific content knowledge and skills". Although a minimum test score is not required for the student to receive credit in the course, completion of the test is mandatory. The EOCT score accounts for 15% of a student's grade in the course. The Georgia Milestone evaluation is taken by public school students in the state. In 2020, because of the ongoing COVID-19 pandemic, the Georgia State BOE agreed to state superintendent Richard Woods' proposal to change the weight of the EOCT test to only count for 0.01% of the Student's course grade. This change is currently only in effect for the 2020–21 school year. Georgia has 85 public colleges, universities, and technical colleges in addition to more than 45 private institutes of higher learning. Among Georgia's public universities is the flagship research university, the University of Georgia, founded in 1785 as the country's oldest state-chartered university and the birthplace of the American system of public higher education. The University System of Georgia is the presiding body over public post-secondary education in the state. The System includes 29 institutions of higher learning and is governed by the Georgia Board of Regents. Georgia's workforce of more than 6.3 million is constantly refreshed by the growing number of people who move there along with the 90,000 graduates from the universities, colleges and technical colleges across the state, including the highly ranked University of Georgia, Georgia Institute of Technology, Georgia State University and Emory University. The HOPE Scholarship, funded by the state lottery, is available to all Georgia residents who have graduated from high school or earned a General Educational Development certificate. The student must maintain a 3.0 or higher grade point average and attend a public college or university in the state. The Georgia Historical Society, an independent educational and research institution, has a research center located in Savannah. The research center's library and archives hold the oldest collection of materials related to Georgia history in the nation. The Atlanta metropolitan area is the ninth largest media market in the United States as ranked by Nielsen Media Research. The state's other top markets are Savannah (95th largest), Augusta (115th largest), and Columbus (127th largest). There are 48 television broadcast stations in Georgia including TBS, TNT, TCM, Cartoon Network, CNN and Headline News, all founded by notable Georgia resident Ted Turner. The Weather Channel also has its headquarters in Atlanta. By far, the largest daily newspaper in Georgia is the Atlanta Journal-Constitution with a daily readership of 195,592 and a Sunday readership of 397,925. Other large dailies include The Augusta Chronicle, the Columbus Ledger-Enquirer, The Telegraph (formerly The Macon Telegraph) and the Savannah Morning News. WSB-AM in Atlanta was the first licensed radio station in the southeastern United States, signing on in 1922. Georgia Public Radio has been in service since 1984 and, with the exception of Atlanta, it broadcasts daily on several FM (and one AM) stations across the state. Georgia Public Radio reaches nearly all of Georgia (with the exception of the Atlanta area, which is served by WABE). WSB-TV in Atlanta is the state's oldest television station, having begun operations in 1948. WSB the first television service in Georgia, and the South. As with all other U.S. states and the federal government, Georgia's government is based on the separation of legislative, executive, and judicial power. Executive authority in the state rests with the governor, currently Brian Kemp (Republican). Both the Governor of Georgia and lieutenant governor are elected on separate ballots to four-year terms of office. Unlike the federal government, but like many other U.S. States, most of the executive officials who comprise the governor's cabinet are elected by the citizens of Georgia rather than appointed by the governor. Legislative authority resides in the General Assembly, composed of the Senate and House of Representatives. The Lieutenant Governor presides over the Senate, while members of the House of Representatives select their own Speaker. The Georgia Constitution mandates a maximum of 56 senators, elected from single-member districts, and a minimum of 180 representatives, apportioned among representative districts (which sometimes results in more than one representative per district); there are currently 56 senators and 180 representatives. The term of office for senators and representatives is two years. The laws enacted by the General Assembly are codified in the Official Code of Georgia Annotated. State judicial authority rests with the state Supreme Court and Court of Appeals, which have statewide authority. In addition, there are smaller courts which have more limited geographical jurisdiction, including Superior Courts, State Courts, Juvenile Courts, Magistrate Courts and Probate Courts. Justices of the Supreme Court and judges of the Court of Appeals are elected statewide by the citizens in non-partisan elections to six-year terms. Judges for the smaller courts are elected to four-year terms by the state's citizens who live within that court's jurisdiction. Georgia consists of 159 counties, second only to Texas, with 254. Georgia had 161 counties until the end of 1931, when Milton and Campbell were merged into the existing Fulton. Some counties have been named for prominent figures in both American and Georgian history, and many bear names with Native American origin. Counties in Georgia have their own elected legislative branch, usually called the Board of Commissioners, which usually also has executive authority in the county. Several counties have a sole Commissioner form of government, with legislative and executive authority vested in a single person. Georgia is the only state with current Sole Commissioner counties. Georgia's Constitution provides all counties and cities with "home rule" authority. The county commissions have considerable power to pass legislation within their county, as a municipality would. Georgia recognizes all local units of government as cities, so every incorporated town is legally a city. Georgia does not provide for townships or independent cities, though there have been bills proposed in the Legislature to provide for townships; it does allow consolidated city-county governments by local referendum. All of Georgia's second-tier cities except Savannah have now formed consolidated city-county governments by referendum: Columbus (in 1970), Athens (1990), Augusta (1995), and Macon (2012). (Augusta and Athens have excluded one or more small, incorporated towns within their consolidated boundaries; Columbus and Macon eventually absorbed all smaller incorporated entities within their consolidated boundaries.) The small town of Cusseta adopted a consolidated city-county government after it merged with unincorporated Chattahoochee County in 2003. Three years later, in 2006, the town of Georgetown consolidated with the rest of Quitman County. There is no true metropolitan government in Georgia, though the Atlanta Regional Commission (ARC) and Georgia Regional Transportation Authority do provide some services, and the ARC must approve all major land development projects in the Atlanta metropolitan area. Georgia voted Republican in six consecutive presidential elections from 1996 to 2016, a streak that was broken when the state went for Democratic candidate Joe Biden in 2020. Until 1964, Georgia's state government had the longest unbroken record of single-party dominance, by the Democratic Party, of any state in the Union. This record was established largely due to the disenfranchisement of most blacks and many poor whites by the state in its constitution and laws in the early 20th century. Some elements, such as requiring payment of poll taxes and passing literacy tests, prevented blacks from registering to vote; their exclusion from the political system lasted into the 1960s and reduced the Republican Party to a non-competitive status in the early 20th century. White Democrats regained power after Reconstruction due in part to the efforts of some using intimidation and violence, but this method came into disrepute. In 1900, shortly before Georgia adopted a disfranchising constitutional amendment in 1908, blacks comprised 47% of the state's population. The whites dealt with this problem of potential political power by the 1908 amendment, which in practice disenfranchised blacks and poor whites, nearly half of the state population. It required that any male at least 21 years of age wanting to register to vote must also be of good character and able to pass a test on citizenship, be able to read and write provisions of the U.S. and Georgia constitutions, or own at least forty acres of land or $500 in property. Any Georgian who had fought in any war from the American Revolution through the Spanish–American War was exempted from these additional qualifications. More importantly, any Georgian descended from a veteran of any of these wars also was exempted. Because, by 1908, many white Georgia males were grandsons of veterans or owned the required property, the exemption and the property requirement basically allowed only well-to-do whites to vote. The qualifications of good character, citizenship knowledge, literacy (all determined subjectively by white registrars), and property ownership were used to disqualify most blacks and poor whites, preventing them from registering to vote. The voter rolls dropped dramatically. In the early 20th century, Progressives promoted electoral reform and reducing the power of ward bosses to clean up politics. Their additional rules, such as the eight-box law, continued to effectively close out people who were illiterate. White one-party rule was solidified. For more than 130 years, from 1872 to 2003, Georgians nominated and elected only white Democratic governors, and white Democrats held the majority of seats in the General Assembly. Most of the Democrats elected throughout these years were Southern Democrats, who were fiscally and socially conservative by national standards. This voting pattern continued after the segregationist period. Legal segregation was ended by passage of federal legislation in the 1960s. According to the 1960 census, the proportion of Georgia's population that was African American was 28%; hundreds of thousands of blacks had left the state in the Great Migration to the North and Midwest. New white residents arrived through migration and immigration. Following support from the national Democratic Party for the civil rights movement and especially civil rights legislation of 1964 and 1965, most African-American voters, as well as other minority voters, have largely supported the Democratic Party in Georgia. In 2002, incumbent moderate Democratic Governor Roy Barnes was defeated by Republican Sonny Perdue, a state legislator and former Democrat. While Democrats retained control of the State House, they lost their majority in the Senate when four Democrats switched parties. They lost the House in the 2004 election. Republicans then controlled all three partisan elements of the state government. Even before 2002, the state had become increasingly supportive of Republicans in Presidential elections. It has supported a Democrat for president only four times since 1960. In 1976 and 1980, native son Jimmy Carter carried the state; in 1992, the former Arkansas governor Bill Clinton narrowly won the state; and in 2020, Joe Biden narrowly carried the state. Generally, Republicans are strongest in the predominantly white suburban (especially the Atlanta suburbs) and rural portions of the state. Many of these areas were represented by conservative Democrats in the state legislature well into the 21st century. One of the most conservative of these was U.S. Congressman Larry McDonald, former head of the John Birch Society, who died when the Soviet Union shot down KAL 007 near Sakhalin Island. Democratic candidates have tended to win a higher percentage of the vote in the areas where black voters are most numerous, as well as in the cities among liberal urban populations (especially Atlanta and Athens), and the central and southwestern portion of the state. The ascendancy of the Republican Party in Georgia and in the South in general resulted in Georgia U.S. House of Representatives member Newt Gingrich being elected as Speaker of the House following the election of a Republican majority in the House in 1994. Gingrich served as Speaker until 1999, when he resigned in the aftermath of the loss of House seats held by members of the GOP. Gingrich mounted an unsuccessful bid for president in the 2012 election, but withdrew after winning only the South Carolina and Georgia primaries. In 2008, Democrat Jim Martin ran against incumbent Republican Senator Saxby Chambliss. Chambliss failed to acquire the necessary 50 percent of votes due to a Libertarian Party candidate receiving the remainder of votes. In the runoff election held on December 2, 2008, Chambliss became the second Georgia Republican to be reelected to the U.S. Senate. In the 2018 elections, the governorship remained under control by a Republican (by 54,723 votes against a Democrat, Stacey Abrams), Republicans lost eight seats in the Georgia House of Representatives (winning 106), while Democrats gained ten (winning 74), Republicans lost two seats in the Georgia Senate (winning 35 seats), while Democrats gained two seats (winning 21), and five Democrat U.S. Representatives were elected with Republicans winning nine seats (one winning with just 419 votes over the Democratic challenger, and one seat being lost). In the three presidential elections up to and including 2016, the Republican candidate has won Georgia by approximately five to eight points over the Democratic nominee, at least once for each election being narrower than margins recorded in some states that have flipped within that timeframe, such as Michigan, Ohio and Wisconsin. This trend led to the state narrowly electing Democrat Joe Biden for president in 2020, and it coming to be regarded as a swing state. In a 2020 study, Georgia was ranked as 49th on the "Cost of Voting Index" with only Texas ranking higher. In 2022, Georgia swung substantially back to the right towards Republicans with incumbent Republican Governor Brian Kemp winning reelection by 7.5% over Democrat Stacey Abrams with a raw vote margin of over 300,000 votes in the 2022 Georgia gubernatorial election. The largest amount since the early 2000s, and every other Republican statewide getting elected by a 5–10% margin of victory. During the 1960s and 1970s, Georgia made significant changes in civil rights and governance. As in many other states, its legislature had not reapportioned congressional districts according to population from 1931 to after the 1960 census. Problems of malapportionment in the state legislature, where rural districts had outsize power in relation to urban districts, such as Atlanta's, were corrected after the U.S. Supreme Court ruling in Wesberry v. Sanders (1964). The court ruled that congressional districts had to be reapportioned to have essentially equal populations. A related case, Reynolds v. Sims (1964), required state legislatures to end their use of geographical districts or counties in favor of "one man, one vote"; that is, districts based upon approximately equal populations, to be reviewed and changed as necessary after each census. These changes resulted in residents of Atlanta and other urban areas gaining political power in Georgia in proportion to their populations. From the mid-1960s, the voting electorate increased after African Americans' rights to vote were enforced under civil rights law. Economic growth through this period was dominated by Atlanta and its region. It was a bedrock of the emerging "New South". From the late 20th century, Atlanta attracted headquarters and relocated workers of national companies, becoming more diverse, liberal and cosmopolitan than many areas of the state. In the 21st century, many conservative Democrats, including former U.S. Senator and governor Zell Miller, decided to support Republicans. The state's then-socially conservative bent resulted in wide support for measures such as restrictions on abortion. In 2004, a state constitutional amendment banning same-sex marriages was approved by 76% of voters. However, after the United States Supreme Court issued its ruling in Obergefell v. Hodges, all Georgia counties came into full compliance, recognizing the rights of same-sex couples to marry in the state. In presidential elections, Georgia voted solely Democratic in every election from 1900 to 1960. In 1964, it was one of only a handful of states to vote for Republican Barry Goldwater over Democrat Lyndon B. Johnson. In 1968, it did not vote for either of the two parties, but rather the American Independent Party and its nominee, Alabama Governor George Wallace. In 1972, the state returned to Republicans as part of a landslide victory for Richard Nixon. In 1976 and 1980, it voted for Democrat and former Georgia governor Jimmy Carter. The state returned to Republicans in 1984 and 1988, before going Democratic once again in 1992. For every election between that year and 2020, Georgia voted heavily Republican, in line with many of its neighbors in the Deep South. In 2020, it voted Democratic for the first time in 28 years, aiding Joe Biden in his defeat of incumbent Republican Donald Trump. Prior to 2020, Republicans in state, federal and congressional races had seen decreasing margins of victory, and many election forecasts had ranked Georgia as a "toss-up" state, or with Biden as a very narrow favorite. Concurrent with the 2020 presidential election were two elections for both of Georgia's United States Senate seats (one of which being a special election due to the resignation of Senator Johnny Isakson, and the other being regularly scheduled). After no candidate in either race received a majority of the vote, both went to January 5, 2021, run-offs, which Democrats Jon Ossoff and Raphael Warnock won. Ossoff is the state's first Jewish senator, and Warnock is the state's first Black senator. Biden's, Ossoff's, and Warnock's wins were attributed to the rapid diversification of the suburbs of Atlanta and increased turnout of younger African American voters, particularly around the suburbs of Atlanta and in Savannah, Georgia. There are 48 state parks, 15 historic sites, and numerous wildlife preserves under supervision of the Georgia Department of Natural Resources. Other historic sites and parks are supervised by the National Park Service and include the Andersonville National Historic Site in Andersonville; Appalachian National Scenic Trail; Chattahoochee River National Recreation Area near Atlanta; Chickamauga and Chattanooga National Military Park at Fort Oglethorpe; Cumberland Island National Seashore near St. Marys; Fort Frederica National Monument on St. Simons Island; Fort Pulaski National Monument in Savannah; Jimmy Carter National Historic Site near Plains; Kennesaw Mountain National Battlefield Park near Kennesaw; Martin Luther King Jr. National Historical Park in Atlanta; Ocmulgee National Monument at Macon; Trail of Tears National Historic Trail; and the Okefenokee Swamp in Waycross, Georgia Outdoor recreational activities include hiking along the Appalachian Trail; Civil War Heritage Trails; rock climbing and whitewater kayaking. Other outdoor activities include hunting and fishing. Transportation in Georgia is overseen by the Georgia Department of Transportation, a part of the executive branch of the state government. Georgia's major Interstate Highways are I-20, I-75, I-85, and I-95. On March 18, 1998, the Georgia House of Representatives passed a resolution naming the portion of Interstate 75, which runs from the Chattahoochee River northward to the Tennessee state line the Larry McDonald Memorial Highway. Larry McDonald, a Democratic member of the House of Representatives, had been on Korean Air Lines Flight 007 when it was shot down by the Soviets on September 1, 1983. Georgia's primary commercial airport is Hartsfield–Jackson Atlanta International Airport (ATL), the world's busiest airport. In addition to Hartsfield–Jackson, there are eight other airports serving major commercial traffic in Georgia. Savannah/Hilton Head International Airport is the second-busiest airport in the state as measured by passengers served, and is the only additional international airport. Other commercial airports (ranked in order of passengers served) are located in Augusta, Columbus, Albany, Macon, Brunswick, Valdosta, and Athens. The Georgia Ports Authority manages two deepwater seaports, at Savannah and Brunswick, and two river ports, at Bainbridge and Columbus. The Port of Savannah is a major U.S. seaport on the Atlantic coast. The Metropolitan Atlanta Rapid Transit Authority (MARTA) is the principal rapid transit system in the Atlanta metropolitan area. Formed in 1971 as strictly a bus system, MARTA operates a network of bus routes linked to a rapid transit system consisting of 48 miles (77 km) of rail track with 38 train stations. MARTA operates almost exclusively in Fulton and DeKalb counties, with bus service to two destinations in Cobb county and the Cumberland Transfer Center next to the Cumberland Mall, and a single rail station in Clayton County at Hartsfield–Jackson Atlanta International Airport. MARTA also operates a separate paratransit service for disabled customers. As of 2009, the average total daily ridership for the system (bus and rail) was 482,500 passengers. The state has 151 general hospitals, more than 15,000 doctors and almost 6,000 dentists. The state is ranked forty-first in the percentage of residents who engage in regular exercise. Jimmy Carter, from Plains, Georgia, was President of the United States from 1977 to 1981. Martin Luther King Jr. was born in Atlanta in 1929. He was a civil rights movement leader who protested for equal rights and against racial discrimination. He won the Nobel Peace Prize in 1964. Blake R. Van Leer played an important role in the civil rights movement, Georgia's economy and was president of Georgia Tech. Mordecai Sheftall, the highest ranking Jewish officer in the American Revolution, was born and lived his life in Georgia. Naomi Chapman Woodruff, originally from Idaho, was responsible for developing a peanut breeding program in Georgia which lead to a harvest of nearly five times the typical amount. 33°N 83°W / 33°N 83°W / 33; -83 (State of Georgia)
{{Infobox Philosopher |region = Western Philosophy |era = [[German philosophy|18th-century philosophy]] |color = #B0C4DE |image = Immanuel Kant portrait c1790.jpg |caption = Immanuel Kant |name = Immanuel Kant |birth = 22 April 1724<br />[[Königsberg]], [[Kingdom of Prussia]] |death = {{death date and age|df=yes|1804|02|12|1724|04|22}}<br /> [[Königsberg]], [[Kingdom of Prussia]] |school_tradition = [[Kantianism]], [[Age of Enlightenment|enlightenment philosophy]] |main_interests = [[Epistemology]], [[Metaphysics]], [[Ethics]] |influences = [[Christian Wolff (philosopher)|Wolff]], [[Alexander Baumgarten|Baumgarten]], [[Johannes Nikolaus Tetens|Tetens]], [[Francis Hutcheson (philosopher)|Hutcheson]], [[Sextus Empiricus|Empiricus]], [[Michel de Montaigne|Montaigne]], [[David Hume|Hume]], [[René Descartes|Descartes]], [[Nicolas Malebranche|Malebranche]], [[Spinoza]], [[Gottfried Leibniz|Leibniz]], [[John Locke|Locke]], [[George Berkeley|Berkeley]], [[Jean-Jacques Rousseau|Rousseau]], [[Isaac Newton|Newton]], [[Emanuel Swedenborg]] |influenced = [[Johann Gottlieb Fichte|Fichte]], [[Friedrich Wilhelm Joseph Schelling|Schelling]], [[Georg Wilhelm Friedrich Hegel|Hegel]], [[Arthur Schopenhauer|Schopenhauer]], [[Friedrich Nietzsche|Nietzsche]], [[Charles Peirce|Peirce]], [[Edmund Husserl|Husserl]], [[Martin Heidegger|Heidegger]], [[Ludwig Wittgenstein|Wittgenstein]], [[Jean-Paul Sartre|Sartre]], [[Ernst Cassirer|Cassirer]], [[Jürgen Habermas|Habermas]], [[John Rawls|Rawls]], [[Noam Chomsky|Chomsky]], [[Robert Nozick|Nozick]], [[Karl Popper]], [[Søren Kierkegaard|Kierkegaard]], [[C. G. Jung|Jung]], [[John R. Searle|Searle]], [[Michel Foucault]], [[Hannah Arendt]], [[Karl Marx]], [[Giovanni Gentile]], [[Karl Jaspers]], [[Friedrich Hayek|Hayek]], [[Henri Bergson|Bergson]], [[Ørsted]], [[A.J. Ayer]], [[Ralph Waldo Emerson|Emerson]], [[Otto Weininger|Weininger]], [[P.F. Strawson]], [[John McDowell]] |notable_ideas = [[Categorical imperative]], [[Transcendental idealism]], [[Synthetic proposition|Synthetic a priori]], [[Noumenon]], [[Sapere aude]], [[Nebular hypothesis]] |signature = Immanuel Kant signature.svg}} '''Immanuel Kant''' (22 April 1724 – 12 February 1804) was a [[Germany|German]] [[philosophy|philosopher]] born in [[Königsberg]], [[East Prussia]]. Kant studied philosophy at the [[University of Königsberg]], and later became a professor of philosophy. He called his system "[[transcendental idealism]]". Kant's writing about [[epistemology]], [[metaphysics]], [[ethics]], and [[aesthetics]] have made him one of the most influential figures in the history of philosophy. Today the town Königsberg is part of [[Russia]], and is renamed [[Kaliningrad]]. In Kant's time, it was the second largest city in the kingdom of [[Prussia]]. == Life == Immanuel Kant was born on April 22, 1724. In 1740 he entered the [[University of Königsberg]]<ref>{{cite web|url=http://www.britannica.com/topic/Albertus-University-of-Konigsberg|title=Albertus University of Königsberg|publisher=Encyclopædia Britannica|accessdate=2016-02-22}}</ref> and studied the philosophy of [[Gottfried Leibniz]] and his follower Christian Wolff. He studied there until 1746 when his father died, then left Königsberg to take up a job as tutor. He became the tutor of [[Count Kayserling]] and his family. In 1755 Kant became a lecturer and stayed in this position until 1770. He was made the second librarian of the Royal Library in 1766. Kant was eventually given the Chair of [[Logic]] and [[Metaphysics]] at the University of Königsberg. In his entire life Kant never travelled more than seventy miles from the city of Königsberg. Kant died on February 12, 1804 with the final words "Es ist gut" ("It is good").<ref>{{cite web|url=http://www.britannica.com/biography/Immanuel-Kant/Last-years|title=Immanuel Kant: Last years|publisher=Encyclopædia Britannica|accessdate=2016-02-22}}</ref> === University === After finishing his study in the university, Kant hoped to be a teacher of philosophy, but it was very difficult. He could have lived a life of private lecturer as interested in [[physics]], both [[astronomical object]]s (such as planets and stars) and the earth. He wrote some papers about this, but he became more interested in [[metaphysics]]. He wanted to learn the nature of human experience: how humans could know something, and what their [[knowledge]] was based on. === First doubts === Under the strong influence of the philosophical system of Leibniz and Wolff, Kant began to doubt the basic answers of past philosophers. Then, Kant read a Scottish philosopher, [[David Hume]]. Hume had tried to make clear what our experience had been, and had reached a very strong opinion called "[[skeptic]]ism", that there was nothing to make our experience sure. Kant was very shocked by Hume, and saw the theory he had learned in a new point of view. He began to try finding a third way other than the two that Kant called "skepticism" and "dogmaticism". Kant read another thinker, named [[Jean-Jacques Rousseau]], who lived in [[Geneva]], [[Switzerland]], and wrote in [[French language|French]]. His thought on human beings, especially on [[morals]], human [[freedom]] and perpetual [[peace]], impressed Kant. == Philosophy == Some scholars like to include Kant as one of the German [[Idealism|idealists]], but Kant himself did not belong to that group and would not have agreed that he was an idealist. The most-known work of Kant is the book ''[[Critique of Pure Reason]]'' (''Kritik der reinen Vernunft'') that Kant published in 1781. Kant called his way of thought "critique", not philosophy. Kant said that critique was a preparation for establishment of real philosophy. According to Kant, people should know what human [[reason]] can do and which limits it has. In the ''Critique of Pure Reason'' Kant wrote about several limits of human reason, to both feeling and thinking something. For [[sensation]], there are two limits inside of human perception: space and time. There are no physical objects, but the limits of our mind that are at work whenever we sense and experience things. For thinking, he said there are twelve categories or pure rational concepts, divided into four fields: [[quantity]], [[quality]], relation and modality. Kant thought human reason applied those ideas to everything automatically whenever it had a thought about anything it could experience. == Ideology == Is what we think only our [[fantasy]], and not really there? Kant said "No", although without those sensual and rational limitations, we can think nothing. Kant was convinced there would be something we could not know directly beyond our limits, and even with our limits, we could know many things. It cannot be a personal fantasy either, since those limitations were common to all human reason before our particular experience. Kant called what we could not know directly ''Ding an sich'' -- "thing itself". We can think "thing itself" but cannot have any experience about it, nor know it. [[God]], the [[eternity]] of [[soul]], [[life]] after [[death]], such things belong to "thing itself", so they were not the right objects of philosophy according to Kant, although people had liked to discuss them from ancient times. == Books == Kant wrote two other books named ''Critique'': ''Critique of the practical reason'' (1788) and ''Critique of the Judgement'' (1790). In ''Critique of the practical reason'' Kant wrote about the problem of [[freedom]] and God. It was his main work of ethics. In ''Critique of the Judgement'' Kant wrote about [[beauty]] and [[teleology]], or the problem if there was a purpose in general, if the world, a living creature had a reason to exist, and so on. In both books, Kant said we could not answer those problems, because they were concerned with "thing itself". == Influence == Kant had a great influence on other thinkers. In the 19th century, German philosophers like [[Fichte]], [[Schelling]], [[Georg Wilhelm Friedrich Hegel|Hegel]], [[Arthur Schopenhauer|Schopenhauer]] and writers like [[Herder]], [[Friedrich Schiller|Schiller]], and [[Johann Wolfgang von Goethe|Goethe]] were influenced by Kant. In the early 20th century Kant's ideas were very influential on one group of German philosophers. They became known as the new-Kantians. One of them, [[Windelband]], said, "every philosophy before Kant poured into Kant, and every philosophy after Kant pours from Kant". Kant has influenced many modern thinkers, including [[Hannah Arendt]], and [[John Rawls]]. == References == {{reflist}} == Other websites == {{Commons|Immanuel Kant}} * [http://www.hkbu.edu.hk/~ppp/ksp1/KSPglos.html Stephen Plaquist's Glossary of Kantian Terminology] * [http://www.rsrevision.com/Alevel/ethics/kant/index.htm Kant's Ethical Theory] Kantian ethics explained, applied and evaluated * {{gutenberg author|id=Kant, |name=Immanuel Kant}} * [http://www.korpora.org/Kant/ All works of Kant] {{Webarchive|url=https://web.archive.org/web/20141101095455/http://www.korpora.org/Kant/ |date=2014-11-01 }} {{in lang|de}} * [http://www.manchester.edu/kant Kant in the Classroom] {{Webarchive|url=https://web.archive.org/web/20090903132310/http://www.manchester.edu/kant/ |date=2009-09-03 }} (background information for Kant's lectures) * [http://www.intratext.com/Catalogo/Autori/AUT216.HTM Immanuel Kant's works]: text, concordances and frequency list {{Metaphysics}} {{Social and political philosophy}} {{authority control}} {{DEFAULTSORT:Kant, Immanuel}} [[Category:1724 births]] [[Category:1804 deaths]] [[Category:18th-century German philosophers]] [[Category:19th-century German philosophers]] [[Category:Continental philosophers]] [[Category:German academics]] [[Category:Prussian writers]] [[Category:People from former German territories]] [[Category:People from Kaliningrad]] [[Category:Prussian philosophers]]
Immanuel Kant (22 April 1724 – 12 February 1804) was a German philosopher and one of the central Enlightenment thinkers. Born in Königsberg, Kant's comprehensive and systematic works in epistemology, metaphysics, ethics, and aesthetics have made him one of the most influential and controversial figures in modern Western philosophy. In his doctrine of transcendental idealism, Kant argued space and time are mere "forms of intuition" that structure all experience and that the objects of experience are mere "appearances". The nature of things as they are in themselves is unknowable to us. In an attempt to counter the philosophical doctrine of skepticism, he wrote the Critique of Pure Reason (1781/1787), his best-known work. Kant drew a parallel to the Copernican revolution in his proposal to think of the objects of experience as conforming to our spatial and temporal forms of intuition and the categories of our understanding, so that we have a priori cognition of those objects. These claims have proved especially influential in the social sciences, particularly sociology and anthropology, which regard human activities as pre-oriented by cultural norms. Kant believed that reason is the source of morality, and that aesthetics arises from a faculty of disinterested judgment. Kant's religious views were deeply connected to his moral theory. Their exact nature, however, remains in dispute. He hoped that perpetual peace could be secured through an international federation of republican states and international cooperation. His cosmopolitan reputation, however, is called into question by his promulgation of scientific racism for much of his career, although he altered his views on the subject in the last decade of his life. Immanuel Kant was born on 22 April 1724 into a Prussian German family of Lutheran Protestant faith in Königsberg, East Prussia (since 1946 the Russian city of Kaliningrad). His mother, Anna Regina Reuter (1697–1737), was born in Königsberg to a father from Nuremberg. Her surname is sometimes erroneously given as Porter. Kant's father, Johann Georg Kant (1682–1746), was a German harness-maker from Memel, at the time Prussia's most northeastern city (now Klaipėda, Lithuania). It is possible that the Kants got their name from the village of Kantvainiai (German: Kantwaggen – today part of Priekulė) and were of Kursenieki origin. Baptized Emanuel, Kant later changed the spelling of his name to Immanuel after learning Hebrew. He was the fourth of nine children (six of whom reached adulthood). The Kant household stressed the pietist values of religious devotion, humility, and a literal interpretation of the Bible. The young Immanuel's education was strict, punitive and disciplinary, and focused on Latin and religious instruction over mathematics and science. In his later years, Kant lived a strictly ordered life. It was said that neighbors would set their clocks by his daily walks. He never married but seems to have had a rewarding social life; he was a popular teacher as well as a modestly successful author, even before starting on his major philosophical works. Kant showed a great aptitude for study at an early age. He first attended the Collegium Fridericianum, from which he graduated at the end of the summer of 1740. In 1740, aged 16, he enrolled at the University of Königsberg, where he would later remain for the rest of his professional life. He studied the philosophy of Gottfried Leibniz and Christian Wolff under Martin Knutzen (Associate Professor of Logic and Metaphysics from 1734 until he died in 1751), a rationalist who was also familiar with developments in British philosophy and science and introduced Kant to the new mathematical physics of Isaac Newton. Knutzen dissuaded Kant from the theory of pre-established harmony, which he regarded as "the pillow for the lazy mind". He also dissuaded Kant from idealism, the idea that reality is purely mental, which most philosophers in the 18th century regarded negatively. The theory of transcendental idealism that Kant later included in the Critique of Pure Reason was developed partially in opposition to traditional idealism. Kant had contacts with students, colleagues, friends and diners who frequented the local Masonic lodge. His father's stroke and subsequent death in 1746 interrupted his studies. Kant left Königsberg shortly after August 1748; he would return there in August 1754. He became a private tutor in the towns surrounding Königsberg, but continued his scholarly research. In 1749, he published his first philosophical work, Thoughts on the True Estimation of Living Forces (written in 1745–1747). Kant is best known for his work in the philosophy of ethics and metaphysics, but he made significant contributions to other disciplines. In 1754, while contemplating on a prize question by the Berlin Academy about the problem of Earth's rotation, he argued that the Moon's gravity would slow down Earth's spin and he also put forth the argument that gravity would eventually cause the Moon's tidal locking to coincide with the Earth's rotation. The next year, he expanded this reasoning to the formation and evolution of the Solar System in his Universal Natural History and Theory of the Heavens. In 1755, Kant received a license to lecture in the University of Königsberg and began lecturing on a variety of topics including mathematics, physics, logic, and metaphysics. In his 1756 essay on the theory of winds, Kant laid out an original insight into the Coriolis force. In 1756, Kant also published three papers on the 1755 Lisbon earthquake. Kant's theory, which involved shifts in huge caverns filled with hot gases, though inaccurate, was one of the first systematic attempts to explain earthquakes in natural rather than supernatural terms. In 1757, Kant began lecturing on geography making him one of the first lecturers to explicitly teach geography as its own subject. Geography was one of Kant's most popular lecturing topics and, in 1802, a compilation by Friedrich Theodor Rink of Kant's lecturing notes, Physical Geography, was released. After Kant became a professor in 1770, he expanded the topics of his lectures to include lectures on natural law, ethics, and anthropology, along with other topics. In the Universal Natural History, Kant laid out the nebular hypothesis, in which he deduced that the Solar System had formed from a large cloud of gas, a nebula. Kant also correctly deduced that the Milky Way was a large disk of stars, which he theorized formed from a much larger spinning gas cloud. He further suggested that other distant "nebulae" might be other galaxies. These postulations opened new horizons for astronomy, for the first time extending it beyond the solar system to galactic and intergalactic realms. From then on, Kant turned increasingly to philosophical issues, although he continued to write on the sciences throughout his life. In the early 1760s, Kant produced a series of important works in philosophy. The False Subtlety of the Four Syllogistic Figures, a work in logic, was published in 1762. Two more works appeared the following year: Attempt to Introduce the Concept of Negative Magnitudes into Philosophy and The Only Possible Argument in Support of a Demonstration of the Existence of God. By 1764, Kant had become a notable popular author, and wrote Observations on the Feeling of the Beautiful and Sublime; he was second to Moses Mendelssohn in a Berlin Academy prize competition with his Inquiry Concerning the Distinctness of the Principles of Natural Theology and Morality (often referred to as "The Prize Essay"). In 1766 Kant wrote a critical piece on Emanuel Swedenborg's Dreams of a Spirit-Seer. In 1770, Kant was appointed Full Professor of Logic and Metaphysics at the University of Königsberg. In defense of this appointment, Kant wrote his inaugural dissertation On the Form and Principles of the Sensible and the Intelligible World This work saw the emergence of several central themes of his mature work, including the distinction between the faculties of intellectual thought and sensible receptivity. To miss this distinction would mean to commit the error of subreption, and, as he says in the last chapter of the dissertation, only in avoiding this error does metaphysics flourish. It is often claimed that Kant was a late developer, that he only became an important philosopher in his mid-50's after rejecting his earlier views. While it is true that Kant wrote his greatest works relatively late in life, there is a tendency to underestimate the value of his earlier works. Recent Kant scholarship has devoted more attention to these "pre-critical" writings and has recognized a degree of continuity with his mature work. At age 46, Kant was an established scholar and an increasingly influential philosopher, and much was expected of him. In correspondence with his ex-student and friend Markus Herz, Kant admitted that, in the inaugural dissertation, he had failed to account for the relation between our sensible and intellectual faculties. He needed to explain how we combine what is known as sensory knowledge with the other type of knowledge—that is, reasoned knowledge—these two being related, but having very different processes. Kant also credited David Hume with awakening him from a "dogmatic slumber" in which he had unquestioningly accepted the tenets of both religion and natural philosophy. Hume, in his 1739 Treatise on Human Nature, had argued that we only know the mind through a subjective, essentially illusory series of perceptions. Ideas such as causality, morality, and objects are not evident in experience, so their reality may be questioned. Kant felt that reason could remove this skepticism, and he set himself to solving these problems. Although fond of company and conversation with others, Kant isolated himself, and resisted friends' attempts to bring him out of his isolation. When Kant emerged from his silence in 1781, the result was the Critique of Pure Reason. Kant countered Hume's empiricism by claiming that some knowledge exists inherently in the mind, independent of experience. He drew a parallel to the Copernican revolution in his proposal that worldly objects can be intuited a priori, and that intuition is consequently distinct from objective reality. He acquiesced to Hume somewhat by defining causality as a "regular, constant sequence of events in time, and nothing more". Although now uniformly recognized as one of the greatest works in the history of philosophy, this Critique disappointed Kant's readers upon its initial publication. The book was long, over 800 pages in the original German edition, and written in a convoluted style. Kant was quite upset with its reception. His former student, Johann Gottfried Herder criticized it for placing reason as an entity worthy of criticism instead of considering the process of reasoning within the context of language and one's entire personality. Similar to Christian Garve and Johann Georg Heinrich Feder, he rejected Kant's position that space and time possessed a form that could be analyzed. Additionally, Garve and Feder also faulted Kant's Critique for not explaining differences in perception of sensations. Its density made it, as Herder said in a letter to Johann Georg Hamann, a "tough nut to crack", obscured by "all this heavy gossamer". Its reception stood in stark contrast to the praise Kant had received for earlier works, such as his Prize Essay and shorter works that preceded the first Critique. Recognizing the need to clarify the original treatise, Kant wrote the Prolegomena to any Future Metaphysics in 1783 as a summary of its main views. Shortly thereafter, Kant's friend Johann Friedrich Schultz (1739–1805), a professor of mathematics, published Explanations of Professor Kant's Critique of Pure Reason (Königsberg, 1784), which was a brief but very accurate commentary on Kant's Critique of Pure Reason. Kant's reputation gradually rose through the latter portion of the 1780s, sparked by a series of important works: the 1784 essay, "Answer to the Question: What is Enlightenment?"; 1785's Groundwork of the Metaphysics of Morals (his first work on moral philosophy); and, from 1786, Metaphysical Foundations of Natural Science. But Kant's fame ultimately arrived from an unexpected source. In 1786, Karl Leonhard Reinhold published a series of public letters on Kantian philosophy. In these letters, Reinhold framed Kant's philosophy as a response to the central intellectual controversy of the era: the pantheism controversy. Friedrich Jacobi had accused the recently deceased Gotthold Ephraim Lessing (a distinguished dramatist and philosophical essayist) of Spinozism. Such a charge, tantamount to atheism, was vigorously denied by Lessing's friend Moses Mendelssohn, leading to a bitter public dispute among partisans. The controversy gradually escalated into a debate about the values of the Enlightenment and the value of reason. Reinhold maintained in his letters that Kant's Critique of Pure Reason could settle this dispute by defending the authority and bounds of reason. Reinhold's letters were widely read and made Kant the most famous philosopher of his era. Kant published a second edition of the Critique of Pure Reason in 1787, heavily revising the first parts of the book. Most of his subsequent work focused on other areas of philosophy. He continued to develop his moral philosophy, notably in 1788's Critique of Practical Reason (known as the second Critique), and 1797's Metaphysics of Morals. The 1790 Critique of the Power of Judgment (the third Critique) applied the Kantian system to aesthetics and teleology. In 1792, Kant's attempt to publish the Second of the four Pieces of Religion within the Bounds of Bare Reason, in the journal Berlinische Monatsschrift, met with opposition from the King's censorship commission, which had been established that same year in the context of the French Revolution. Kant then arranged to have all four pieces published as a book, routing it through the philosophy department at the University of Jena to avoid the need for theological censorship. This insubordination earned him a now-famous reprimand from the King. When he nevertheless published a second edition in 1794, the censor was so irate that he arranged for a royal order that required Kant never to publish or even speak publicly about religion. Kant then published his response to the King's reprimand and explained himself in the preface of The Conflict of the Faculties. He also wrote a number of semi-popular essays on history, religion, politics, and other topics. These works were well received by Kant's contemporaries and confirmed his preeminent status in eighteenth-century philosophy. There were several journals devoted solely to defending and criticizing Kantian philosophy. Despite his success, philosophical trends were moving in another direction. Many of Kant's most important disciples and followers (including Reinhold, Beck, and Fichte) transformed the Kantian position. The progressive stages of revision of Kant's teachings marked the emergence of German idealism. Kant opposed these developments and publicly denounced Fichte in an open letter in 1799. It was one of his final acts expounding a stance on philosophical questions. In 1800, a student of Kant named Gottlob Benjamin Jäsche (1762–1842) published a manual of logic for teachers called Logik, which he had prepared at Kant's request. Jäsche prepared the Logik using a copy of a textbook in logic by Georg Friedrich Meier entitled Excerpt from the Doctrine of Reason, in which Kant had written copious notes and annotations. The Logik has been considered of fundamental importance to Kant's philosophy, and the understanding of it. The great 19th-century logician Charles Sanders Peirce remarked, in an incomplete review of Thomas Kingsmill Abbott's English translation of the introduction to Logik, that "Kant's whole philosophy turns upon his logic." Also, Robert Schirokauer Hartman and Wolfgang Schwarz wrote in the translators' introduction to their English translation of the Logik, "Its importance lies not only in its significance for the Critique of Pure Reason, the second part of which is a restatement of fundamental tenets of the Logic, but in its position within the whole of Kant's work." Kant's health, long poor, worsened. He died at Königsberg on 12 February 1804, uttering "Es ist gut" (It is good) before expiring. His unfinished final work was published as Opus Postumum. Kant always cut a curious figure in his lifetime for his modest, rigorously scheduled habits, which have been referred to as clocklike. However, Heinrich Heine noted the magnitude of "his destructive, world-crushing thoughts" and considered him a sort of philosophical "executioner", comparing him to Robespierre with the observation that both men "represented in the highest the type of provincial bourgeois. Nature had destined them to weigh coffee and sugar, but Fate determined that they should weigh other things and placed on the scales of the one a king, on the scales of the other a god." When his body was transferred to a new burial spot, his skull was measured during the exhumation and found to be larger than the average German male's with a "high and broad" forehead. His forehead has been an object of interest ever since it became well known through his portraits: "In Döbler's portrait and in Kiefer's faithful if expressionistic reproduction of it—as well as in many of the other late eighteenth- and early nineteenth-century portraits of Kant—the forehead is remarkably large and decidedly retreating." Kant's mausoleum adjoins the northeast corner of Königsberg Cathedral in Kaliningrad, Russia. The mausoleum was constructed by the architect Friedrich Lahrs and was finished in 1924, in time for the bicentenary of Kant's birth. Originally, Kant was buried inside the cathedral, but in 1880 his remains were moved to a neo-Gothic chapel adjoining the northeast corner of the cathedral. Over the years, the chapel became dilapidated and was demolished to make way for the mausoleum, which was built on the same location. The tomb and its mausoleum are among the few artifacts of German times preserved by the Soviets after they captured the city. Today, many newlyweds bring flowers to the mausoleum. Artifacts previously owned by Kant, known as Kantiana, were included in the Königsberg City Museum. However, the museum was destroyed during World War II. A replica of the statue of Kant that in German times stood in front of the main University of Königsberg building was donated by a German entity in the early 1990s and placed in the same grounds. After the expulsion of Königsberg's German population at the end of World War II, the University of Königsberg where Kant taught was replaced by the Russian-language Kaliningrad State University, which appropriated the campus and surviving buildings. In 2005, the university was renamed Immanuel Kant State University of Russia. The name change was announced at a ceremony attended by President Vladimir Putin of Russia and Chancellor Gerhard Schröder of Germany, and the university formed a Kant Society, dedicated to the study of Kantianism. The university was again renamed in the 2010s, to Immanuel Kant Baltic Federal University. Like many of his contemporaries, Kant was greatly impressed with the scientific advances made by Newton and others. This new evidence of the power of human reason, however, called into question for many the traditional authority of politics and religion. In particular, the modern mechanistic view of the world called into question the very possibility of morality; for, if there is no agency, there cannot be any responsibility. The aim of Kant's critical project is to secure human autonomy, the basis of religion and morality, from this threat of mechanism—and to do so in a way that preserves the advances of modern science. In the Critique of Pure Reason, Kant summarizes his philosophical concerns in the following three questions: The Critique of Pure Reason focuses upon the first question and opens a conceptual space for an answer to the second question. It argues that even though we cannot, strictly know that we are free, we can—and for practical purposes, must—think of ourselves as free. In Kant's own words, "I had to deny knowledge in order to make room for faith." Our rational faith in morality is further developed in The Groundwork of the Metaphysics of Morals and The Critique of Practical Reason. The Critique of the Power of Judgment argues we may rationally hope for the harmonious unity of the theoretical and practical domains treated in the first two Critiques on the basis, not only of its conceptual possibility, but also on the basis of our affective experience of natural beauty and, more generally, the organization of the natural world. In Religion within the Bounds of Mere Reason, Kant endeavors to complete his answer to this third question. These works all place the active, rational human subject at the center of the cognitive and moral worlds. In brief, Kant argues that the mind itself necessarily makes a constitutive contribution to knowledge, that this contribution is transcendental rather than psychological, and that to act autonomously is to act according to rational moral principles. Kant's 1781 (revised 1787) book the Critique of Pure Reason has often been cited as the most significant volume of metaphysics and epistemology in modern philosophy. In the first Critique, and later on in other works as well, Kant frames the "general" and "real problem of pure reason" in terms of the following question: "How are synthetic judgments a priori possible?" To parse this claim, it is necessary to define some terms. First, Kant makes a distinction in terms of the source of the content of knowledge: Second, he makes a distinction in terms of the form of knowledge: An analytic proposition is true by nature of strictly conceptual relations. All analytic propositions are a priori (it is analytically true that no analytic proposition could be a posteriori). By contrast, a synthetic proposition is one the content of which includes something new. The truth or falsehood of a synthetic statement depends upon something more than what is contained in its concepts. The most obvious form of synthetic proposition is a simple empirical observation. Philosophers such as David Hume believed that these were the only possible kinds of human reason and investigation, which he called "relations of ideas" and "matters of fact". Establishing the synthetic a priori as a third mode of knowledge would allow Kant to push back against Hume's skepticism about such matters as causation and metaphysical knowledge more generally. This is because, unlike a posteriori cognition, a priori cognition has "true or strict...universality" and includes a claim of "necessity". Kant himself regards it as uncontroversial that we do have synthetic a priori knowledge—most obviously, that of mathematics. That 7 + 5 = 12, he claims, is a result not contained in the concepts of seven, five, and the addition operation. Yet, although he considers the possibility of such knowledge to be obvious, Kant nevertheless assumes the burden of providing a philosophical proof that we have a priori knowledge in mathematics, the natural sciences, and metaphysics. It is the twofold aim of the Critique both to prove and to explain the possibility of this knowledge. Before turning to Kant's arguments in the body of the Critique, there are two more distinctions from its introductory sections that must be introduced. "There are", Kant says, "two stems of human cognition, which may perhaps arise from a common but to us unknown root, namely sensibility and understanding, through the first of which objects are given to us, but through the second of which they are thought." Kant's term for the object of sensibility is intuition, and his term for the object of the understanding is concept. In general terms, the former is a non-discursive representation of a particular object, and the latter is a discursive (or mediate) representation of a general type of object. The conditions of possible experience require both intuitions and concepts, that is, the affection of the receptive sensibility and the actively synthesizing power of the understanding. Thus the statement: "Thoughts without content are empty, intuitions without concepts are blind." Kant's basic strategy in the first half of his book will be to argue that some intuitions and concepts are pure—that is, are contributed entirely by the mind, independent of anything empirical. Knowledge generated on this basis, under certain conditions, can be synthetic a priori. This insight is known as Kant's "Copernican revolution", because, just as Copernicus advanced astronomy by way of a radical shift in perspective, so Kant here claims do the same for metaphysics. The second half of the Critique is the explicitly critical part. In this "transcendental dialectic", Kant argues that many of the claims of traditional rationalist metaphysics violate the criteria he claims to establishing the first, "constructive" part of his book. As Kant observes, "human reason, without being moved by the mere vanity of knowing it all, inexorably pushes on, driven by its own need to such questions that cannot be answered by any experiential use of reason". It is the project of "the critique of pure reason" to establish the limits as to just how far reason may legitimately so proceed. The section of the Critique entitled "The transcendental aesthetic" advances Kant's famous thesis of transcendental idealism. Something is "transcendental" if it is a necessary condition for the possibility of experience, and "idealism" denotes some form of mind-dependence that must be further specified. (The correct interpretation of Kant's own specification remains controversial.) The thesis, then, states that human beings only experience and know appearances, not things-in-themselves, because space and time are nothing but the subjective forms of intuition that we ourselves contribute to experience. Nevertheless, although Kant says that space and time are "transcendentally ideal"—the pure forms of human sensibility, rather than part of nature or reality as it exists in-itself—he also claims that they are "empirically real", by which he means "that 'everything that can come before us externally as an object' is in both space and time, and that our internal intuitions of ourselves are in time". However we may interpret Kant's doctrine, he clearly wishes to distinguish his position from the subjective idealism of Berkeley. Paul Guyer, although critical of many of Kant's arguments in this section, nevertheless writes of the "Transcendental Aesthetic" that it "not only lays the first stone in Kant's constructive theory of knowledge; it also lays the foundation for both his critique and his reconstruction of traditional metaphysics. It argues that all genuine knowledge requires a sensory component, and thus that metaphysical claims that transcend the possibility of sensory confirmation can never amount to knowledge." One interpretation, known as the "two-world" interpretation, regards Kant's position as a statement of epistemological limitation, that we are not able to transcend the bounds of our own mind, meaning that we cannot access the "thing-in-itself". However, Kant also speaks of the thing in itself or transcendent object as a product of the (human) understanding as it attempts to conceive of objects in abstraction from the conditions of sensibility. Following this line of thought, some interpreters have argued that the thing in itself does not represent a separate ontological domain but simply a way of considering objects by means of the understanding alone; this is known as the "two-aspect" view. Following the "Transcendental Analytic" is the "Transcendental Logic". Whereas the former was concerned with the contributions of the sensibility, the latter is concerned, first, with the contributions of the understanding ("Transcendental Analytic") and, second, with the faculty of reason as the source of both metaphysical errors and genuine regulatory principles ("Transcendental Dialectic"). The "Transcendental Analytic" is further divided into two sections. The first, "Analytic of Concepts", is concerned with establishing the universality and necessity of the pure concepts of the understanding (i.e., the categories). This section contains Kant's famous "transcendental deduction". The second, "Analytic of Principles", is concerned with the application of those pure concepts in empirical judgments. This second section is longer than the first and is further divided into many sub-sections. The "Analytic of Concepts" argues for the universal and necessary validity of the pure concepts of the understanding, or the categories, e.g., the concepts of substance and causation. These twelve basic categories define what it is to be a thing in general—that is, they articulate the necessary conditions according to which something is a possible object of experience. These, in conjunction with the a priori forms of intuition, are the basis of all synthetic a priori cognition. According to Guyer and Wood, "Kant's idea is that just as there are certain essential features of all judgments, so there must be certain corresponding ways in which we form the concepts of objects so that judgments may be about objects." Kant provides two central lines of argumentation in support of his claims about the categories. The first, known as the "metaphysical deduction", proceeds analytically from a table of the Aristotelian logical functions of judgment. As Kant was aware, however, this assumes precisely what the skeptic rejects, namely, the existence of synthetic a priori cognition. For this reason, Kant also supplies a synthetic argument that does not depend upon the assumption in dispute. This argument, provided under the heading "Transcendental Deduction of the Pure Concepts of the Understanding", is widely considered to be both the most important and the most difficult of Kant's arguments in the Critique. Kant himself said that it is the one that cost him the most labor. Frustrated by its confused reception in the first edition of his book, he rewrote it entirely for the second edition. The "Transcendental Deduction" gives Kant's argument that these pure concepts apply universally and necessarily to the objects that are given in experience. According to Guyer and Wood, "He centers his argument on the premise that our experience can be ascribed to a single identical subject, via what he calls the 'transcendental unity of apperception,' only if the elements of experience given in intuition are synthetically combined so as to present us with objects that are thought through the categories." Kant's principle of apperception is that "The I think must be able to accompany all my representations; for otherwise something would be represented in me that could not be thought at all, which is as much as to say that the representation would either be impossible or else at least would be nothing for me." The necessary possibility of the self-ascription of the representations of self-consciousness, identical to itself through time, is an a priori conceptual truth that cannot be based on experience. This, however, is only a bare sketch of one of the arguments that Kant presents. Kant's deduction of the categories in the "Analytic of Concepts", if successful, demonstrates its claims about the categories only in an abstract way. The task of the "Analytic of Principles" is to show both that they must universally apply to objects given in actual experience (i.e., manifolds of intuition) and how it is they do so. In the first book of this section on the "schematism", Kant connects each of the purely logical categories of the understanding to the temporality of intuition to show that, although non-empirical, they do have purchase upon the objects of experience. The second book continues this line of argument in four chapters, each associated with one of the category groupings. In some cases, it adds a connection to the spatial dimension of intuition to the categories it analyzes. The fourth chapter of this section, "The Analogies of Experience", marks a shift from "mathematical" to "dynamical" principles, that is, to those that deal with relations among objects. Some commentators consider this the most significant section of the Critique. The analogies are three in number: The fourth section of this chapter, which is not an analogy, deals with the empirical use of the modal categories. That was the end of the chapter in the A edition of the Critique. The B edition, however, includes one more short section, "The Refutation of Idealism". In this section, by analysis of the concept of self-consciousness, Kant argues that his transcendental idealism is a "critical" or "formal" idealism that does not deny the existence of reality apart from our subjective representations. The final chapter of "The Analytic of Principles" distinguishes phenomena, of which we have can have genuine knowledge, from noumena, a term which refers to objects of pure thought that we cannot know, but to which we may still refer "in a negative sense". An Appendix to the section further develops Kant's criticism of Leibnizian-Wolffian rationalism by arguing that its "dogmatic" metaphysics confuses the "mere features of concepts through which we think things...[with] features of the objects themselves". Against this, Kant reasserts his own insistence upon the necessity of a sensible component in all genuine knowledge. The second of the two Divisions of "The Transcendental Logic", "The Transcendental Dialectic", contains the "negative" portion of Kant's Critique, which builds upon the "positive" arguments of the preceding "Transcendental Analytic" to expose the limits of metaphysical speculation. In particular, it is concerned to demonstrate as spurious the efforts of reason to arrive at knowledge independent of sensibility. This endeavor, Kant argues, is doomed to failure, which he claims to demonstrate by showing that reason, unbounded by sense, is always capable of generating opposing or otherwise incompatible conclusions. Like "the light dove, in free flight cutting through the air, the resistance of which it feels", reason "could get the idea that it could do even better in airless space". Against this, Kant claims that, absent epistemic friction, there can be no knowledge. Nevertheless, Kant's critique is not entirely destructive. He presents the speculative excesses of traditional metaphysics as inherent in our very capacity of reason. Moreover, he argues that its products are not without some (carefully qualified) regulative value. Kant calls the basic concepts of metaphysics "ideas". They are different from the concepts of understanding in that they are not limited by the critical stricture limiting knowledge to the conditions of possible experience and its objects. "Transcendental illusion" is Kant's term for the tendency of reason to produce such ideas. Although reason has a "logical use" of simply drawing inferences from principles, in "The Transcendental Dialectic", Kant is concerned with its purportedly "real use" to arrive at conclusions by way of unchecked regressive syllogistic ratiocination. The three categories of relation, pursued without regard to the limits of possible experience, yield the three central ideas of traditional metaphysics: Although Kant denies that these ideas can be objects of genuine cognition, he argues that they are the result of reason's inherent drive to unify cognition into a systematic whole. Leibnizian-Wolffian metaphysics was divided into four parts: ontology, psychology, cosmology, and theology. Kant replaces the first with the positive results of the first part of the Critique. He proposes to replace the following three with his later doctrines of anthropology, the metaphysical foundations of natural science, and the critical postulation of human freedom and morality. In the second of the two Books of "The Transcendental Dialectic", Kant undertakes to demonstrate the contradictory nature of unbounded reason. He does this by developing contradictions in each of the three metaphysical disciplines that he contends are, in fact, pseudo-sciences. This section of the Critique is long and Kant's arguments are extremely detailed. In this context, it not possible to do much more than enumerate the topics of discussion. The first chapter addresses what Kant terms the paralogisms—i.e., false inferences—that pure reason makes in the metaphysical discipline of rational psychology. He argues that one cannot take the mere thought of "I" in the proposition "I think" as the proper cognition of "I" as an object. In this way, he claims to debunk various metaphysical theses about the substantiality, unity, and self-identity of the soul. The second chapter, which is the longest, takes up the topic Kant calls the antinomies of pure reason—that is, the contradictions of reason with itself—in the metaphysical discipline of rational cosmology. (Originally, Kant had thought that all transcendental illusion could be analyzed in antinomic terms.) He presents four cases in which he claims reason is able to prove opposing theses with equal plausibility: Kant further argues in each case that his doctrine of transcendental idealism is able to resolve the antinomy. The third chapter examines fallacious arguments about God in rational theology under the heading of the "Ideal of Pure Reason". (Whereas an idea is a pure concept generated by reason, an ideal is the concept of an idea as an individual thing.) Here Kant addresses and claims to refute three traditional arguments for the existence of God: the ontological argument, the cosmological argument, and the physio-theological argument (i.e., the argument from design). The results of the transcendental dialectic so far appear to be entirely negative. In an Appendix to this section, however, Kant rejects such a conclusion. The ideas of pure reason, he argues, have an important regulatory function in directing and organizing our theoretical and practical inquiry. Kant's later works elaborate upon this function at length and in detail. Kant developed his ethics, or moral philosophy, in three works: Groundwork of the Metaphysics of Morals (1785), Critique of Practical Reason (1788), and Metaphysics of Morals (1797). With regard to morality, Kant argued that the source of the good lies not in anything outside the human subject, either in nature or given by God, but rather is only the good will itself. A good will is one that acts from duty in accordance with the universal moral law that the autonomous human being freely gives itself. This law obliges one to treat humanity—understood as rational agency, and represented through oneself as well as others—as an end in itself rather than (merely) as means to other ends the individual might hold. Kant is known for his theory that all moral obligation is grounded in what he calls the "categorical imperative", which is derived from the concept of duty. He argues that the moral law is a principle of reason itself, not based on contingent facts about the world, such as what would make us happy; to act on the moral law has no other motive than "worthiness to be happy". In the Critique of Pure Reason, Kant distinguishes between the transcendental idea of freedom, which as a psychological concept is "mainly empirical" and refers to "whether a faculty of beginning a series of successive things or states from itself is to be assumed", and the practical concept of freedom as the independence of our will from the "coercion" or "necessitation through sensuous impulses". Kant finds it a source of difficulty that the practical idea of freedom is founded on the transcendental idea of freedom, but for the sake of practical interests uses the practical meaning, taking "no account of ... its transcendental meaning", which he feels was properly "disposed of" in the Third Antinomy, and as an element in the question of the freedom of the will is for philosophy "a real stumbling block" that has embarrassed speculative reason. Kant calls practical "everything that is possible through freedom"; he calls the pure practical laws that are never given through sensuous conditions, but are held analogously with the universal law of causality, moral laws. Reason can give us only the "pragmatic laws of free action through the senses", but pure practical laws given by reason a priori dictate "what is to be done". Kant's categories of freedom function primarily as conditions for the possibility for actions (i) to be free, (ii) to be understood as free, and (iii) to be morally evaluated. For Kant, although actions as theoretical objects are constituted by means of the theoretical categories, actions as practical objects (objects of practical use of reason, and which can be good or bad) are constituted by means of the categories of freedom. Only in this way can actions, as phenomena, be a consequence of freedom, and be understood and evaluated as such. Kant makes a distinction between categorical and hypothetical imperatives. A hypothetical imperative is one that we must obey to satisfy contingent desires. A categorical imperative binds us regardless of our desires: for example, everyone has a duty to not lie, regardless of circumstances, even though it is sometimes in our narrowly selfish interest to do so. These imperatives are morally binding because they are based on reason, rather than contingent facts about an agent. Unlike hypothetical imperatives, which bind us insofar as we are part of a group or society which we owe duties to, we cannot opt out of the categorical imperative, because we cannot opt out of being rational agents. We owe a duty to rationality by virtue of being rational agents; therefore, rational moral principles apply to all rational agents at all times. Stated in other terms, with all forms of instrumental rationality excluded from morality, "the moral law itself, Kant holds, can only be the form of lawfulness itself, because nothing else is left once all content has been rejected". Kant provides three formulations for the categorical imperative. He claims that these are necessarily equivalent, as all being expressions of the pure universality of the moral law as such. Many scholars, however, are not convinced. The formulas are as follows: Kant defines maxim as a "subjective principle of volition", which is distinguished from an "objective principle or 'practical law.'" While "the latter is valid for every rational being and is a 'principle according to which they ought to act[,]' a maxim 'contains the practical rule which reason determines in accordance with the conditions of the subject (often their ignorance or inclinations) and is thus the principle according to which the subject does act.'" Maxims fail to qualify as practical laws if they produce a contradiction in conception or a contradiction in the will when universalized. A contradiction in conception happens when, if a maxim were to be universalized, it ceases to make sense, because the "maxim would necessarily destroy itself as soon as it was made a universal law". For example, if the maxim 'It is permissible to break promises' was universalized, no one would trust any promises made, so the idea of a promise would become meaningless; the maxim would be self-contradictory because, when it is universalized, promises cease to be meaningful. The maxim is not moral because it is logically impossible to universalize—that is, we could not conceive of a world where this maxim was universalized. A maxim can also be immoral if it creates a contradiction in the will when universalized. This does not mean a logical contradiction, but that universalizing the maxim leads to a state of affairs that no rational being would desire. As Kant explains in the 1785 Groundwork of the Metaphysics of Morals (and as its title directly indicates) that text is "nothing more than the search for and establishment of the supreme principle of morality". His promised Metaphysics of Morals, however, was much delayed and did not appear until its two parts, "The Doctrine of Right" and "The Doctrine of Virtue", were published separately in 1797 and 1798. The first deals with political philosophy, the second with ethics. "The Doctrine of Virtue" provides "a very different account of ordinary moral reasoning" than the one suggested by the Groundwork. It is concerned with duties of virtue or "ends that are at the same time duties". It is here, in the domain of ethics, that the greatest innovation by The Metaphysics of Morals is to be found. According to Kant's account, "ordinary moral reasoning is fundamentally teleological—it is reasoning about what ends we are constrained by morality to pursue, and the priorities among these ends we are required to observe". More specifically, There are two sorts of ends that it is our duty to have: our own perfection and the happiness of others (MS 6:385). "Perfection" includes both our natural perfection (the development of our talents, skills, and capacities of understanding) and moral perfection (our virtuous disposition) (MS 6:387). A person's "happiness" is the greatest rational whole of the ends the person set for the sake of her own satisfaction (MS 6:387–388). Kant's elaboration of this teleological doctrine offers up a moral theory very different from the one typically attributed to him on the basis of his foundational works alone. In Towards Perpetual Peace: A Philosophical Project, Kant listed several conditions that he thought necessary for ending wars and creating a lasting peace. They included a world of constitutional republics. His classical republican theory was extended in the Doctrine of Right, the first part of the Metaphysics of Morals (1797). Kant believed that universal history leads to the ultimate world of republican states at peace, but his theory was not pragmatic. The process was described in Perpetual Peace as natural rather than rational: What affords this guarantee (surety) is nothing less than the great artist nature (natura daedala rerum) from whose mechanical course purposiveness shines forth visibly, letting concord arise by means of the discord between human beings even against their will; and for this reason nature, regarded as necessitation by a cause the laws of whose operation are unknown to us, is called fate, but if we consider its purposiveness in the course of the world as the profound wisdom of a higher cause directed to the objective final end of the human race and predetermining this course of the world, it is called providence. Kant's political thought can be summarized as republican government and international organization: "In more characteristically Kantian terms, it is doctrine of the state based upon the law (Rechtsstaat) and of eternal peace. Indeed, in each of these formulations, both terms express the same idea: that of legal constitution or of 'peace through law.'" "Kant's political philosophy, being essentially a legal doctrine, rejects by definition the opposition between moral education and the play of passions as alternate foundations for social life. The state is defined as the union of men under law. The state rightly so called is constituted by laws which are necessary a priori because they flow from the very concept of law. A regime can be judged by no other criteria nor be assigned any other functions, than those proper to the lawful order as such." He opposed "democracy", which at his time meant direct democracy, believing that majority rule posed a threat to individual liberty. He stated, "democracy in the strict sense of the word is necessarily a despotism because it establishes an executive power in which all decide for and, if need be, against one (who thus does not agree), so that all, who are nevertheless not all, decide; and this is a contradiction of the general will with itself and with freedom." As with most writers at the time, he distinguished three forms of government—namely, democracy, aristocracy, and monarchy—with mixed government as the most ideal form of it. He believed in republican ideals and forms of governance, and rule of law brought on by them. Although Kant published this as a "popular piece", Mary J. Gregor points out that two years later, in The Metaphysics of Morals, Kant claims to demonstrate systematically that "establishing universal and lasting peace constitutes not merely a part of the doctrine of right, but rather the entire final end of the doctrine of right within the limits of mere reason". The Doctrine of Right, published in 1797, contains Kant's most mature and systematic contribution to political philosophy. It addresses duties according to law, which are "concerned only with protecting the external freedom of individuals" and indifferent to incentives. (Although we do have a moral duty "to limit ourselves to actions that are right, that duty is not part of [right] itself".) Its basic political idea is that "each person's entitlement to be his or her own master is only consistent with the entitlements of others if public legal institutions are in place". Starting in the twentieth century, commentators have tended to see Kant as having a strained relationship with religion, although in the nineteenth century this had not been the prevalent view. Karl Leonhard Reinhold, whose letters helped make Kant famous, wrote: "I believe that I may infer without reservation that the interest of religion, and of Christianity in particular, accords completely with the result of the Critique of Reason." According to Johann Schultz, who wrote one of the first commentaries on Kant: "And does not this system itself cohere most splendidly with the Christian religion? Do not the divinity and beneficence of the latter become all the more evident?" The reason for these views was Kant's moral theology and the widespread belief that his philosophy was the great antithesis to Spinozism, which was widely seen as a form of sophisticated pantheism or even atheism. As Kant's philosophy disregarded the possibility of arguing for God through pure reason alone, for the same reasons it also disregarded the possibility of arguing against God through pure reason alone. Kant articulates his strongest criticisms of the organization and practices of religious organizations to those that encourage what he sees as a religion of counterfeit service to God. Among the major targets of his criticism are external ritual, superstition, and a hierarchical church order. He sees these as efforts to make oneself pleasing to God in ways other than conscientious adherence to the principle of moral rightness in choosing and acting upon one's maxims. Kant's criticisms on these matters, along with his rejection of certain theoretical proofs for the existence of God that were grounded in pure reason (particularly the ontological argument) and his philosophical commentary on some Christian doctrines, have resulted in interpretations that see Kant as hostile to religion in general and to Christianity in particular. Other interpreters, nevertheless, consider that Kant was trying to mark off defensible from indefensible Christian belief. Regarding Kant's conception of religion, some critics have argued that he was sympathetic to deism. Other critics have argued that Kant's moral conception moves from deism to theism (as moral theism), for example, Allen W. Wood and Merold Westphal. As for Kant's book Religion within the Bounds of Mere Reason, it was emphasized that Kant reduced religiosity to rationality, religion to morality, and Christianity to ethics. However, many interpreters, including Allen W. Wood and Lawrence Pasternack, now agree with Stephen Palmquist's claim that a better way of reading Kant's Religion is to see him as raising morality to the status of religion. Kant discusses the subjective nature of aesthetic qualities and experiences in Observations on the Feeling of the Beautiful and Sublime (1764). Kant's contribution to aesthetic theory is developed in the Critique of the Power of Judgment (1790), where he investigates the possibility and logical status of "judgments of taste". In the "Critique of Aesthetic Judgment", the first major division of the Critique of the Power of Judgment, Kant used the term "aesthetic" in a manner that differs from its modern sense. In the Critique of Pure Reason, to note essential differences between judgments of taste, moral judgments, and scientific judgments, Kant abandoned the term "aesthetic" as "designating the critique of taste", noting that judgments of taste could never be "directed" by "laws a priori". After A. G. Baumgarten, who wrote Aesthetica (1750–58), Kant was one of the first philosophers to develop and integrate aesthetic theory into a unified and comprehensive philosophical system, utilizing ideas that played an integral role throughout his philosophy. In the chapter "Analytic of the Beautiful" in the Critique of the Power of Judgment, Kant states that beauty is not a property of an artwork or natural phenomenon, but is instead consciousness of the pleasure that attends the 'free play' of the imagination and the understanding. Even though it appears that we are using reason to decide what is beautiful, the judgment is not a cognitive judgment, "and is consequently not logical, but aesthetical". A pure judgement of taste is subjective since it refers to the emotional response of the subject and is based upon nothing but esteem for an object itself: it is a disinterested pleasure, and we feel that pure judgements of taste (i.e., judgements of beauty), lay claim to universal validity. This universal validity is not derived from a determinate concept of beauty but from common sense. Kant also believed that a judgement of taste shares characteristics engaged in a moral judgement: both are disinterested, and we hold them to be universal. In the chapter "Analytic of the Sublime" Kant identifies the sublime as an aesthetic quality that, like beauty, is subjective, but unlike beauty refers to an indeterminate relationship between the faculties of the imagination and of reason, and shares the character of moral judgments in the use of reason. The feeling of the sublime, divided into two distinct modes (the mathematical and the dynamical sublime), describes two subjective moments that concern the relationship of the faculty of the imagination to reason. Some commentators argue that Kant's critical philosophy contains a third kind of the sublime, the moral sublime, which is the aesthetic response to the moral law or a representation, and a development of the "noble" sublime in Kant's theory of 1764. The mathematical sublime results from the failure of the imagination to comprehend natural objects that appear boundless and formless, or appear "absolutely great". This imaginative failure is then recuperated through the pleasure taken in reason's assertion of the concept of infinity. In this move the faculty of reason proves itself superior to our fallible sensible self. In the dynamical sublime, there is the sense of annihilation of the sensible self as the imagination tries to comprehend a vast might. This power of nature threatens us but through the resistance of reason to such sensible annihilation, the subject feels a pleasure and a sense of the human moral vocation. This appreciation of moral feeling through exposure to the sublime helps to develop moral character. Kant developed a theory of humor that has been interpreted as an "incongruity" theory. He illustrated his theory of humor by telling three narrative jokes in the Critique of Judgment. He thought that the physiological impact of humor is akin to that of music. Kant developed a distinction between an object of art as a material value subject to the conventions of society and the transcendental condition of the judgment of taste as a "refined" value in his Idea for a Universal History with a Cosmopolitan Aim (1784). In the Fourth and Fifth Theses of that work he identified all art as the "fruits of unsociableness" due to men's "antagonism in society" and, in the Seventh Thesis, asserted that while such material property is indicative of a civilized state, only the ideal of morality and the universalization of refined value through the improvement of the mind "belongs to culture". Kant lectured on anthropology, the study of human nature, for twenty-three years. His Anthropology from a Pragmatic Point of View was published in 1798. Transcripts of Kant's lectures on anthropology were published for the first time in 1997 in German. Kant was among the first people of his time to introduce anthropology as an intellectual area of study, long before the field gained popularity, and his texts are considered to have advanced the field. His point of view was to influence the works of later philosophers such as Martin Heidegger and Paul Ricoeur. Kant was also the first to suggest using a dimensionality approach to human diversity. He analyzed the nature of the Hippocrates-Galen four temperaments and plotted in two dimensions "what belongs to a human being's faculty of desire": "his natural aptitude or natural predisposition" and "his temperament or sensibility". Cholerics were described as emotional and energetic, phlegmatics as balanced and weak, sanguines as balanced and energetic, and melancholics as emotional and weak. These two dimensions reappeared in all subsequent models of temperament and personality traits. Kant viewed anthropology in two broad categories: (1) the physiological approach, which he referred to as "what nature makes of the human being"; and (2) the pragmatic approach, which explores the things that a human "can and should make of himself". Kant's theory of race and his prejudicial beliefs are among the most contentious areas of recent Kant scholarship. While few, if any, dispute the overt racism and chauvinism present in his work, a more contested question is the degree to which it degrades or invalidates his other contributions. His most severe critics assert that Kant intentionally manipulated science to support chattel slavery and discrimination. Others acknowledge that he lived in an era of immature science, with many erroneous beliefs, some racist, all appearing decades before evolution, molecular genetics, and other sciences that today are taken for granted. Kant was one of the most notable Enlightenment thinkers to defend racism. Philosopher Charles W. Mills is unequivocal: "Kant is also seen as one of the central figures in the birth of modern 'scientific' racism. Whereas other contributors to early racial thought like Carolus Linnaeus and Johann Friedrich Blumenbach had offered only 'empirical' (scare-quotes necessary!) observation, Kant produced a full-blown theory of race." Using the four temperaments of ancient Greece, Kant proposed a hierarchy of racial categories including white Europeans, black Africans, and red Native Americans. Although he was a proponent of scientific racism for much of his career, Kant's views on race changed significantly in the last decade of his life, and he ultimately rejected racial hierarchies and European colonialism in Perpetual Peace: A Philosophical Sketch (1795). Kant was an opponent of miscegenation, believing that whites would be "degraded" and that "fusing of races" is undesirable, for "not every race adopts the morals and customs of the Europeans." He states that "instead of assimilation, which was intended by the melting together of the various races, nature has here made a law of just the opposite." Kant was also an anti-Semite, believing that Jews were incapable of transcending material forces, which a moral order required. In this way, Jews are presented as the opposite of autonomous, rational Christians, and therefore incapable of being incorporated into an ethical Christian society. In his "Anthropology", Kant called the Jews "a nation of cheaters" and portrayed them as "a group that has followed not the path of transcendental freedom but that of enslavement to the material world". Charles W. Mills wrote that Kant has been "sanitized for public consumption", his racist works conveniently ignored. Robert Bernasconi stated that Kant "supplied the first scientific definition of race". Emmanuel Chukwudi Eze is credited with bringing Kant's contributions to racism to light in the 1990s among Western philosophers, who he believed often glossed over this part of his life and works. Pauline Kleingeld argues that, while Kant "did defend a racial hierarchy until at least the end of the 1780s", his views on race changed significantly in works published in the last decade of his life. In particular, she argues that Kant rejected past views related to racial hierarchies and the diminished rights or moral status of non-whites in Perpetual Peace (1795). This work also saw him providing extended arguments against European colonialism, which he claimed was morally unjust and incompatible with the equal rights held by indigenous populations. Kleingeld argues that this shift in Kant's views later in life has often been forgotten or ignored in the literature on Kant's racist anthropology, and that the shift suggests a belated recognition of the fact that racial hierarchy was incompatible with a universalized moral framework. While Kant's racist rhetoric is indicative of the state of scholarship and science during the 18th century, German philosopher Daniel-Pascal Zorn explains the risk of taking period quotations out of context. Many of Kant's most outrageous quotations are from a series of articles from 1777–1788, a public exchange among Kant, Herder, natural scientist Georg Forster, and other scholars prominent in that period. Kant asserts that all races of humankind are of the same species, challenging the position of Forster and others that the races were distinct species. While his commentary is clearly biased at times, certain extreme statements were patterned specifically to paraphrase or counter Forster and other authors. By considering the full arc of Kant's scholarship, Zorn notes the progression in both his philosophical and his anthropological works, "with which he argues, against the zeitgeist, for the unity of humanity". Kant's influence on Western thought has been profound. Although the basic tenets of Kant's transcendental idealism (i.e., that space and time are a priori forms of human perception rather than real properties and the claim that formal logic and transcendental logic coincide) have been claimed to be falsified by modern science and logic, and no longer set the intellectual agenda of contemporary philosophers, Kant is credited with having innovated the way philosophical inquiry has been carried on at least up to the early nineteenth century. This shift consisted of several closely related innovations that, although highly contentious in themselves, have become important in subsequent philosophy and in the social sciences broadly construed: Kant's ideas have been incorporated into a variety of schools of thought. These include German idealism, Marxism, positivism, phenomenology, existentialism, critical theory, linguistic philosophy, structuralism, post-structuralism, and deconstruction. During his own life, much critical attention was paid to Kant's thought. He influenced Reinhold, Fichte, Schelling, Hegel, and Novalis during the 1780s and 1790s. Samuel Taylor Coleridge was greatly influenced by Kant and helped to spread awareness of him, and of German Idealism generally, in the UK and the US. In his Biographia Literaria (1817), he credits Kant's ideas in coming to believe that the mind is not a passive, but an active agent in the apprehension of reality. Hegel was one of Kant's first major critics. In Hegel's view the entire project of setting a "transcendental subject" (i.e., human consciousness) apart from the living individual as well as from nature, history, and society was fundamentally flawed, although parts of that very project could be put to good use in a new direction. Similar concerns motivated Hegel's criticisms of Kant's concept of moral autonomy, to which Hegel opposed an ethic focused on the "ethical life" of the community. In a sense, Hegel's notion of "ethical life" is meant to subsume, rather than replace, Kantian ethics. And Hegel can be seen as trying to defend Kant's idea of freedom as going beyond finite "desires", by means of reason. Thus, in contrast to later critics like Nietzsche or Russell, Hegel shares some of Kant's concerns. Kant's thinking on religion was used in Britain by philosophers such as Thomas Carlyle to challenge the nineteenth-century decline in religious faith. British Catholic writers, notably G. K. Chesterton and Hilaire Belloc, followed this approach. Criticisms of Kant were common in the realist views of the new positivism at that time. Arthur Schopenhauer was strongly influenced by Kant's transcendental idealism. He, like G. E. Schulze, Jacobi, and Fichte before him, was critical of Kant's theory of the thing-in-itself. Things-in-themselves, they argued, are neither the cause of what we observe, nor are they completely beyond our access. Ever since the Critique of Pure Reason, philosophers have been critical of Kant's theory of the thing-in-itself. Many have argued that, if such a thing exists beyond experience, then one cannot posit that it affects us causally, since that would entail stretching the category "causality" beyond the realm of experience. With the success and wide influence of Hegel's writings, Kant's own influence began to wane, but a re-examination of his ideas began in Germany in 1865 with the publication of Kant und die Epigonen by Otto Liebmann, whose motto was "Back to Kant". There proceeded an important revival of Kant's theoretical philosophy, known as Neo-Kantianism. Kant's notion of "critique" has been more broadly influential. The early German Romantics, especially Friedrich Schlegel in his "Athenaeum Fragments", used Kant's reflexive conception of criticism in their Romantic theory of poetry. Also in aesthetics, Clement Greenberg, in his classic essay "Modernist Painting", uses Kantian criticism, what Greenberg refers to as "immanent criticism", to justify the aims of abstract painting, a movement Greenberg saw as aware of the key limitation—flatness—that makes up the medium of painting. French philosopher Michel Foucault was also greatly influenced by Kant's notion of "critique" and wrote several pieces on Kant for a re-thinking of the Enlightenment as a form of "critical thought". He went so far as to classify his own philosophy as a "critical history of modernity, rooted in Kant". Kant believed that mathematical truths were forms of synthetic a priori knowledge, which means they are necessary and universal, yet known through the a priori intuition of space and time, as transcendental preconditions of experience. Kant's often brief remarks about mathematics influenced the mathematical school known as intuitionism, a movement in philosophy of mathematics opposed to Hilbert's formalism, and Frege and Bertrand Russell's logicism. With his Perpetual Peace, Kant is considered to have foreshadowed many of the ideas that have come to form the democratic peace theory, one of the main controversies in political science. More concretely, Constructivist theorist Alexander Wendt proposed that the anarchy of the international system could evolve from the 'brutish' Hobbesian anarchy understood by Realist theorists, through Lockean anarchy, and ultimately a Kantian anarchy in which states would see their self-interests as inextricably linked to the well being of other states, thus transforming international politics into a far more peaceful form. Prominent recent Kantians include the British philosophers P. F. Strawson, Onora O'Neill, and Quassim Cassam, and the American philosophers Wilfrid Sellars and Christine Korsgaard. Due to the influence of Strawson and Sellars, among others, there has been a renewed interest in Kant's view of the mind. Central to many debates in philosophy of psychology and cognitive science is Kant's conception of the unity of consciousness. Jürgen Habermas and John Rawls are two significant political and moral philosophers whose work is strongly influenced by Kant's moral philosophy. They have argued against relativism, supporting the Kantian view that universality is essential to any viable moral philosophy. Mou Zongsan's study of Kant has been cited as a highly crucial part in the development of Mou's personal philosophy, namely New Confucianism. Widely regarded as the most influential Kant scholar in China, Mou's rigorous critique of Kant's philosophy—having translated all three of Kant's critiques—served as an ardent attempt to reconcile Chinese and Western philosophy whilst increasing pressure to westernize in China. Kant's influence has also extended to the social, behavioral, and physical sciences—as in the sociology of Max Weber, the psychology of Jean Piaget and Carl Gustav Jung, and the linguistics of Noam Chomsky. Kant's work on mathematics and synthetic a priori knowledge is also cited by theoretical physicist Albert Einstein as an early influence on his intellectual development, though one which he later criticized and rejected. In recent years, there has also been renewed interest in Kant's theory of mind from the point of view of formal logic and computer science. Because of the thoroughness of Kant's paradigm shift, his influence extends well beyond this to thinkers who neither specifically refer to his work nor use his terminology. Unless otherwise noted, all citations are to The Cambridge Edition of the Works of Immanuel Kant in English Translation, 16 vols., ed. Guyer, Paul, and Wood, Allen W. Cambridge: Cambridge University Press, 1992. Citations in the article are to individual works per abbreviations in List of Major works below. Abbreviations used in body of article are boldface in brackets. Unless otherwise noted, pagination is to the critical Akademie edition, which can be found in the margins of the Cambridge translations. Wilhelm Dilthey inaugurated the Academy edition (the Akademie-Ausgabe abbreviated as AA or Ak) of Kant's writings (Gesammelte Schriften, Königlich-Preußische Akademie der Wissenschaften, Berlin, 1902–38) in 1895, and served as its first editor. The volumes are grouped into four sections: An electronic version is also available: Elektronische Edition der Gesammelten Werke Immanuel Kants (vols. 1–23).
[[File:Kinship (english).jpg|thumb|A chart showing the names for relationships in an extended family.]] An '''aunt''' is [[Kinship|relationship]] that describes a [[woman]] who is a [[sibling]] of a [[parent]] or is [[married]] to a sibling of a parent. Aunts are sometimes called Auntie or Aunty. A man is a [[nephew]] to their Aunt and a woman is a [[niece]]. An '''aunt''' is part of an [[extended family]]. The [[child]] of an aunt's is a [[cousin]]. Often the word aunt is used with people who are not related to show [[:wikt:endearment|closeness]] or [[respect]]. In many [[culture]]s there are different words to describe a person's mother's sister and father's sister. In [[Latin]] a mother's sister is ''matertera'' while a father's sister is ''amita''.<ref>Constance Brittain Bouchard, Those of My Blood: Creating Noble Families in Medieval Francia (Philadelphia : University of Pennsylvania Press, 2001), p. 5.</ref> ==Detail== A maternal or paternal relative is one who is related through one's mother or father, respectively. For example, a maternal aunt is the subject's mother's sister. An in-law is a relationship that is not by blood, but instead by marriage. The in-law shares the in-law relationships of the spouse's relatives. For example, an aunt in-law could be the wife of the subject's uncle. * '''aunt''': parent's sister. * '''maternal aunt''': mother's sister. * '''paternal aunt''': father's sister. * '''half-aunt''': parent's half-sister. * '''double-half-aunt''': a person who is a half-sister of both of the subjects parents.<ref>{{Cite web|title=GENETIC AND QUANTITATIVE ASPECTS OF GENEALOGY|url=http://www.genetic-genealogy.co.uk/Toc115570138.html#Toc115570463|url-status=dead|archive-url=https://web.archive.org/web/20170803155256/http://www.genetic-genealogy.co.uk/Toc115570138.html#Toc115570463|archive-date=2017-08-03|access-date=2020-07-06}}</ref> * '''grandaunt''' (or '''great-aunt'''): grandparent's sister. * '''great-grandaunt''' (or '''great-great-aunt'''): great-grandparent's sister. * '''aunt-in-law''': uncle/aunt's wife. * '''aunt-in-law''': spouse's aunt. * '''co-aunt-in-law''': spouse's uncle/aunt's wife. * '''stepaunt''': stepparent's sister. * '''stepaunt''': parent's stepsister. == References == {{reflist}} {{wikt}} [[Category:Family]] {{stub}}
An aunt is a woman who is a sibling of a parent or married to a sibling of a parent. Aunts who are related by birth are second-degree relatives. Alternate terms include auntie or aunty. Children in other cultures and families may refer to the cousins of their parents as aunt or uncle due to the age and generation gap. The word comes from Latin: amita via Old French ante and is a family relationship within an extended or immediate family. The male counterpart of an aunt is an uncle, and the reciprocal relationship is that of a nephew or niece. Aunts by birth (sister of a parent) are related to their nieces and nephews by 25%. As half-aunts are related through half-sisters, they are related by 12.5% to their nieces and nephews. Non-consanguineous aunts (female spouse of a relative) are not genetically related to their nieces and nephews. In some cultures, such as Aboriginal and Torres Strait Islander peoples of Australia, respected senior members of the community, often also referred to as Elders, are addressed as "uncle" (for men) and "aunt" for women, as a mark of seniority and respect, whether related or not, such as Aunty Kathy Mills. In several cultures, no single inclusive term describing both a person's kinship to their parental female sibling or parental female in-law exists. Instead, there are specific terms describing a person's kinship to their mother's female sibling, and a person's kinship to their father's female sibling, per the following table: Aunts in popular culture have not always been portrayed as positive roles. Childless aunts are often subjected to othering in popular culture and presented as exotic or as having a second-best role, with motherhood preferred. Fictional aunts include: Aunt Flo is a popular euphemism referring to the menstrual cycle. An agony aunt is a colloquial term for a female advice columnist.
{{About|the Ancient Greek|the TV character from [[The Simpsons]]|Homer Simpson}} [[File:British Museum (44437274571).jpg|thumb|Bust of Homer in the British Museum]] '''Homer''' is the name of the [[Ancient Greece|Greek]] poet who wrote the [[epic poem]]s the ''[[Iliad]]'' and the ''[[Odyssey]]''.<ref>{{cite web|url=http://www.wsu.edu:8080/~dee/MINOA/HOMER.HTM|publisher=wsu.edu|title=Homer|access-date=2010-05-01|archive-date=2010-05-27|archive-url=https://web.archive.org/web/20100527174651/http://www.wsu.edu:8080/~dee/MINOA/HOMER.HTM|url-status=dead}}</ref> These are the earliest works of Greek literature which have survived to the present day, and are among the greatest treasures of the ancient world. They are a product of [[Mycenaean culture]]. The Iliad tells the story of the [[Trojan war]], which took place around 1190 BC. The manuscripts of Homer were written much later, probably later than 800 BC. ==The history== What makes the history of this period so difficult is that the Mycenaean civilization was followed by several hundred years of decline, from which little or no writing has survived. This period, the [[Greek Dark Ages]], ended around 800 BC, when once again we find Greek writing, though this time in their newly developed [[alphabet]]ic script. Therefore, there is a distinct possibility that the Homeric epics had a life, perhaps a long life, as oral literature, spoken by story-tellers. We know little about Homer, and the later Greeks also knew little about him. Some have doubted he was a real person; others think he was a [[woman]].<ref>{{cite web|url=http://www.atrium-media.com/rogueclassicism/Posts/00004001.html|title=rogueclassicism: Homer a Woman?|website=www.atrium-media.com}}</ref> A [[legend]], has it that he was a [[blind]] poet who lived in [[Ionia]]. There is no definite evidence for any of these ideas. There is [[archaeology|archaeological]] evidence that Troy existed. The site of Troy has been discovered, and excavated. Therefore, it is a fair guess that the epic began as a long oral poem based on the siege of Troy, but with details added and changed as time went on.<ref>E. Bakker 1997. Poetry in speech: orality and Homeric discourse. Ithaca NY. {{ISBN|0-8014-3295-2}}</ref> Homer's poetry is different from other poetry because it had to be short in order to be memorized.<ref>Parry, Milman 1971. ''The making of Homeric verse: the collected papers of Milman Parry''. (ed Adam Parry) Oxford.</ref><ref>{{Cite journal|title=Formular Economy in Homer. The Poetics of the Breaches. Hermes Einzelschriften, 100 – Bryn Mawr Classical Review|journal=Bryn Mawr Classical Review|url=https://bmcr.brynmawr.edu/2008/2008.10.27/|access-date=2020-12-20|language=en-US}}</ref><ref>Ong, Walter. ''Orality and literacy: the technologizing of the word''. Methuen, London & N.Y. Milman Parry's discovery, p20–27.</ref> ==The date of Homer, and the transcription== The date of Homer's existence is not known. [[Herodotus]] said that Homer lived 400 years before his own time, which would place him at around 850 BC;<ref>[[Herodotus]] 2.53.</ref><ref>{{cite web|url=http://controversialhistory.blogspot.com/2008/03/date-of-homer.html|title=Controversies in History: Date of Homer|publisher=}}</ref><ref>{{cite web|url=https://www.gradesaver.com/author/homer|title=Homer Biography - List of Works, Study Guides & Essays - GradeSaver|last=GradeSaver|website=www.gradesaver.com}}</ref><ref>{{cite web|url=http://www.nndb.com/people/841/000087580/|title=Homer|website=www.nndb.com}}</ref><ref>http://webcache.googleusercontent.com/search?q=cache:NtFnt7Nod-0J:www1.pu.edu.tw/~bmon/Literature/Iliad.doc&cd=9&hl=en&ct=clnk&gl=us{{Dead link|date=December 2021 |bot=InternetArchiveBot |fix-attempted=yes }}</ref><ref>{{cite web|url=https://www.greeka.com/cyclades/ios/ios-history/Homer-ios.htm|title=Homer buried in Ios island - Greeka.com|website=Greeka}}</ref> but other ancient sources gave dates much closer to the time of the [[Trojan War]].<ref>{{cite journal|author= Graziosi, Barbara|title=The invention of Homer|location=Cambridge|date=2002|pages=98–101}}</ref> The date of the Trojan War was given as 1194–1184 BC by [[Eratosthenes]]. Today, "the date of Homer" means the date of the writing down the oral poems, rather than the life of Homer. This is called the ''transcription date''. The language suggests that the earliest possible date for the transcription is 800 BC, and the latest possible date is 600 AD.<ref>{{cite journal|author= Nagy, Gregory|title=Homeric poetry and problems of multiformity: the Panathenaic Bottleneck|publisher=Classical Philology|volume=96|date=2001|pages=109–119}}</ref> ==References== {{reflist}} {{wikiquote-en|Homer}} {{wikisource|Author:Homer|Homer}} {{Commonscat|Homer}} {{Ancient Greek Military World}} {{Ancient Greece: Arts and Culture}} {{Authority control}} [[Category:Ancient Greek poets]]
Homer (/ˈhoʊmər/; Ancient Greek: Ὅμηρος [hómɛːros], Hómēros; born c. 8th century BC) was a Greek poet who is credited as the author of the Iliad and the Odyssey, two epic poems that are foundational works of ancient Greek literature. Homer is considered one of the most revered and influential authors in history. Homer's Iliad centers on a quarrel between King Agamemnon and the warrior Achilles during the last year of the Trojan War. The Odyssey chronicles the ten-year journey of Odysseus, king of Ithaca, back to his home after the fall of Troy. The poems are in Homeric Greek, also known as Epic Greek, a literary language which shows a mixture of features of the Ionic and Aeolic dialects from different centuries; the predominant influence is Eastern Ionic. Most researchers believe that the poems were originally transmitted orally. Despite being predominantly known for its tragic and serious themes, the Homeric poems also contain instances of comedy and laughter. Homer's epic poems shaped aspects of ancient Greek culture and education, fostering ideals of heroism, glory, and honor. To Plato, Homer was simply the one who "has taught Greece" (τὴν Ἑλλάδα πεπαίδευκεν, tēn Helláda pepaídeuken). In Dante Alighieri's Divine Comedy, Virgil refers to Homer as "Poet sovereign", king of all poets; in the preface to his translation of the Iliad, Alexander Pope acknowledges that Homer has always been considered the "greatest of poets". From antiquity to the present day, Homeric epics have inspired many famous works of literature, music, art, and film. The question of by whom, when, where and under what circumstances the Iliad and Odyssey were composed continues to be debated. Scholars remain divided as to whether the two works are the product of a single author. It is thought that the poems were composed at some point around the late eighth or early seventh century BC. Many accounts of Homer's life circulated in classical antiquity; the most widespread account was that he was a blind bard from Ionia, a region of central coastal Anatolia in present-day Turkey. Modern scholars consider these accounts legendary. Today, only the Iliad and the Odyssey are associated with the name 'Homer'. In antiquity, a large number of other works were sometimes attributed to him, including the Homeric Hymns, the Contest of Homer and Hesiod, several epigrams, the Little Iliad, the Nostoi, the Thebaid, the Cypria, the Epigoni, the comic mini-epic Batrachomyomachia ("The Frog–Mouse War"), the Margites, the Capture of Oechalia, and the Phocais. These claims are not considered authentic today and were by no means universally accepted in the ancient world. As with the multitude of legends surrounding Homer's life, they indicate little more than the centrality of Homer to ancient Greek culture. Some ancient claims about Homer were established early and repeated often. They include that Homer was blind (taking as self-referential a passage describing the blind bard Demodocus), that he resided at Chios, that he was the son of the river Meles and the nymph Critheïs, that he was a wandering bard, that he composed a varying list of other works (the "Homerica"), that he died either in Ios or after failing to solve a riddle set by fishermen, and various explanations for the name "Homer" (Ὅμηρος, Hómēros). Another tradition from the days of the Roman emperor Hadrian says Epicaste (daughter of Nestor) and Telemachus (son of Odysseus) were the parents of Homer. The two best known ancient biographies of Homer are the Life of Homer by the Pseudo-Herodotus and the Contest of Homer and Hesiod. In the early fourth century BC Alcidamas composed a fictional account of a poetry contest at Chalcis with both Homer and Hesiod. Homer was expected to win, and answered all of Hesiod's questions and puzzles with ease. Then, each of the poets was invited to recite the best passage from their work. Hesiod selected the beginning of Works and Days: "When the Pleiades born of Atlas ... all in due season". Homer chose a description of Greek warriors in formation, facing the foe, taken from the Iliad. Though the crowd acclaimed Homer victor, the judge awarded Hesiod the prize; the poet who praised husbandry, he said, was greater than the one who told tales of battles and slaughter. The study of Homer is one of the oldest topics in scholarship, dating back to antiquity. Nonetheless, the aims of Homeric studies have changed over the course of the millennia. The earliest preserved comments on Homer concern his treatment of the gods, which hostile critics such as the poet Xenophanes of Colophon denounced as immoral. The allegorist Theagenes of Rhegium is said to have defended Homer by arguing that the Homeric poems are allegories. The Iliad and the Odyssey were widely used as school texts in ancient Greek and Hellenistic cultures. They were the first literary works taught to all students. The Iliad, particularly its first few books, was far more intently studied than the Odyssey during the Hellenistic and Roman periods. As a result of the poems' prominence in classical Greek education, extensive commentaries on them developed to explain parts that were culturally or linguistically difficult. During the Hellenistic and Roman periods, many interpreters, especially the Stoics, who believed that Homeric poems conveyed Stoic doctrines, regarded them as allegories, containing hidden wisdom. Perhaps partially because of the Homeric poems' extensive use in education, many authors believed that Homer's original purpose had been to educate. Homer's wisdom became so widely praised that he began to acquire the image of almost a prototypical philosopher. Byzantine scholars such as Eustathius of Thessalonica and John Tzetzes produced commentaries, extensions and scholia to Homer, especially in the twelfth century. Eustathius's commentary on the Iliad alone is massive, sprawling over nearly 4,000 oversized pages in a twenty-first century printed version and his commentary on the Odyssey an additional nearly 2,000. In 1488, the Greek scholar Demetrios Chalkokondyles published the editio princeps of the Homeric poems. The earliest modern Homeric scholars started with the same basic approaches towards the Homeric poems as scholars in antiquity. The allegorical interpretation of the Homeric poems that had been so prevalent in antiquity returned to become the prevailing view of the Renaissance. Renaissance humanists praised Homer as the archetypically wise poet, whose writings contain hidden wisdom, disguised through allegory. In western Europe during the Renaissance, Virgil was more widely read than Homer and Homer was often seen through a Virgilian lens. In 1664, contradicting the widespread praise of Homer as the epitome of wisdom, François Hédelin, abbé d'Aubignac wrote a scathing attack on the Homeric poems, declaring that they were incoherent, immoral, tasteless, and without style, that Homer never existed, and that the poems were hastily cobbled together by incompetent editors from unrelated oral songs. Fifty years later, the English scholar Richard Bentley concluded that Homer did exist, but that he was an obscure, prehistoric oral poet whose compositions bear little relation to the Iliad and the Odyssey as they have been passed down. According to Bentley, Homer "wrote a Sequel of Songs and Rhapsodies, to be sung by himself for small Earnings and good Cheer at Festivals and other Days of Merriment; the Ilias he wrote for men, and the Odysseis for the other Sex. These loose songs were not collected together in the Form of an epic Poem till Pisistratus' time, about 500 Years after." Friedrich August Wolf's Prolegomena ad Homerum, published in 1795, argued that much of the material later incorporated into the Iliad and the Odyssey was originally composed in the tenth century BC in the form of short, separate oral songs, which passed through oral tradition for roughly four hundred years before being assembled into prototypical versions of the Iliad and the Odyssey in the sixth century BC by literate authors. After being written down, Wolf maintained that the two poems were extensively edited, modernized, and eventually shaped into their present state as artistic unities. Wolf and the "Analyst" school, which led the field in the nineteenth century, sought to recover the original, authentic poems which were thought to be concealed by later excrescences. Within the Analyst school were two camps: proponents of the "lay theory", which held that the Iliad and the Odyssey were put together from a large number of short, independent songs, and proponents of the "nucleus theory", which held that Homer had originally composed shorter versions of the Iliad and the Odyssey, which later poets expanded and revised. A small group of scholars opposed to the Analysts, dubbed "Unitarians", saw the later additions as superior, the work of a single inspired poet. By around 1830, the central preoccupations of Homeric scholars, dealing with whether or not "Homer" actually existed, when and how the Homeric poems originated, how they were transmitted, when and how they were finally written down, and their overall unity, had been dubbed "the Homeric Question". Following World War I, the Analyst school began to fall out of favor among Homeric scholars. It did not die out entirely, but it came to be increasingly seen as a discredited dead end. Starting in around 1928, Milman Parry and Albert Lord, after their studies of folk bards in the Balkans, developed the "Oral-Formulaic Theory" that the Homeric poems were originally composed through improvised oral performances, which relied on traditional epithets and poetic formulas. This theory found very wide scholarly acceptance and explained many previously puzzling features of the Homeric poems, including their unusually archaic language, their extensive use of stock epithets, and their other "repetitive" features. Many scholars concluded that the "Homeric Question" had finally been answered. Meanwhile, the 'Neoanalysts' sought to bridge the gap between the 'Analysts' and 'Unitarians'. The Neoanalysts sought to trace the relationships between the Homeric poems and other epic poems, which have now been lost, but of which modern scholars do possess some patchy knowledge. Neoanalysts hold that knowledge of earlier versions of the epics can be derived from anomalies of structure and detail in the surviving versions of the Iliad and Odyssey. These anomalies point to earlier versions of the Iliad in which Ajax played a more prominent role, in which the Achaean embassy to Achilles comprised different characters, and in which Patroclus was actually mistaken for Achilles by the Trojans. They point to earlier versions of the Odyssey in which Telemachus went in search of news of his father not to Menelaus in Sparta but to Idomeneus in Crete, in which Telemachus met up with his father in Crete and conspired with him to return to Ithaca disguised as the soothsayer Theoclymenus, and in which Penelope recognized Odysseus much earlier in the narrative and conspired with him in the destruction of the suitors. Most contemporary scholars, although they disagree on other questions about the genesis of the poems, agree that the Iliad and the Odyssey were not produced by the same author, based on "the many differences of narrative manner, theology, ethics, vocabulary, and geographical perspective, and by the apparently imitative character of certain passages of the Odyssey in relation to the Iliad." Nearly all scholars agree that the Iliad and the Odyssey are unified poems, in that each poem shows a clear overall design, and that they are not merely strung together from unrelated songs. It is also generally agreed that each poem was composed mostly by a single author, who probably relied heavily on older oral traditions. Nearly all scholars agree that the Doloneia in Book X of the Iliad is not part of the original poem, but rather a later insertion by a different poet. Some ancient scholars believed Homer to have been an eyewitness to the Trojan War; others thought he had lived up to 500 years afterwards. Contemporary scholars continue to debate the date of the poems. A long history of oral transmission lies behind the composition of the poems, complicating the search for a precise date. At one extreme, Richard Janko has proposed a date for both poems to the eighth century BC based on linguistic analysis and statistics. Barry B. Powell dates the composition of the Iliad and the Odyssey to sometime between 800 and 750 BC, based on the statement from Herodotus, who lived in the late fifth century BC, that Homer lived four hundred years before his own time "and not more" (καὶ οὐ πλέοσι), and on the fact that the poems do not mention hoplite battle tactics, inhumation, or literacy. Martin Litchfield West has argued that the Iliad echoes the poetry of Hesiod, and that it must have been composed around 660–650 BC at the earliest, with the Odyssey up to a generation later. He also interprets passages in the Iliad as showing knowledge of historical events that occurred in the ancient Near East during the middle of the seventh century BC, including the destruction of Babylon by Sennacherib in 689 BC and the Sack of Thebes by Ashurbanipal in 663/4 BC. At the other extreme, a few American scholars such as Gregory Nagy see "Homer" as a continually evolving tradition, which grew much more stable as the tradition progressed, but which did not fully cease to continue changing and evolving until as late as the middle of the second century BC. "'Homer" is a name of unknown etymological origin, around which many theories were erected in antiquity. One such linkage was to the Greek ὅμηρος (hómēros 'hostage' or 'surety'). The explanations suggested by modern scholars tend to mirror their position on the overall Homeric Question. Nagy interprets it as "he who fits (the song) together". West has advanced both possible Greek and Phoenician etymologies. Scholars continue to debate questions such as whether the Trojan War actually took place – and if so when and where – and to what extent the society depicted by Homer is based on his own or one which was, even at the time of the poems' composition, known only as legends. The Homeric epics are largely set in the east and center of the Mediterranean, with some scattered references to Egypt, Ethiopia and other distant lands, in a warlike society that resembles that of the Greek world slightly before the hypothesized date of the poems' composition. In ancient Greek chronology, the sack of Troy was dated to 1184 BC. By the nineteenth century, there was widespread scholarly skepticism that the Trojan War had ever happened and that Troy had even existed, but in 1873 Heinrich Schliemann announced to the world that he had discovered the ruins of Homer's Troy at Hisarlik in modern Turkey. Some contemporary scholars think the destruction of Troy VIIa c. 1220 BC was the origin of the myth of the Trojan War, others that the poem was inspired by multiple similar sieges that took place over the centuries. Most scholars now agree that the Homeric poems depict customs and elements of the material world that are derived from different periods of Greek history. For instance, the heroes in the poems use bronze weapons, characteristic of the Bronze Age in which the poems are set, rather than the later Iron Age during which they were composed; yet the same heroes are cremated (an Iron Age practice) rather than buried (as they were in the Bronze Age). In some parts of the Homeric poems, heroes are described as carrying large shields like those used by warriors during the Mycenaean period, but, in other places, they are instead described carrying the smaller shields that were commonly used during the time when the poems were written in the early Iron Age. In the Iliad 10.260–265, Odysseus is described as wearing a helmet made of boar's tusks. Such helmets were not worn in Homer's time, but were commonly worn by aristocratic warriors between 1600 and 1150 BC. The decipherment of Linear B in the 1950s by Michael Ventris and continued archaeological investigation has increased modern scholars' understanding of Aegean civilisation, which in many ways resembles the ancient Near East more than the society described by Homer. Some aspects of the Homeric world are simply made up; for instance, the Iliad 22.145–56 describes there being two springs that run near the city of Troy, one that runs steaming hot and the other that runs icy cold. It is here that Hector takes his final stand against Achilles. Archaeologists, however, have uncovered no evidence that springs of this description ever actually existed. The Homeric epics are written in an artificial literary language or 'Kunstsprache' only used in epic hexameter poetry. Homeric Greek shows features of multiple regional Greek dialects and periods, but is fundamentally based on Ionic Greek, in keeping with the tradition that Homer was from Ionia. Linguistic analysis suggests that the Iliad was composed slightly before the Odyssey, and that Homeric formulae preserve older features than other parts of the poems. The poems were composed in unrhymed dactylic hexameter; ancient Greek metre was quantity-based rather than stress-based. Homer frequently uses set phrases such as epithets ('crafty Odysseus', 'rosy-fingered Dawn', 'owl-eyed Athena', etc.), Homeric formulae ('and then answered [him/her], Agamemnon, king of men', 'when the early-born rose-fingered Dawn came to light', 'thus he/she spoke'), simile, type scenes, ring composition and repetition. These habits aid the extemporizing bard, and are characteristic of oral poetry. For instance, the main words of a Homeric sentence are generally placed towards the beginning, whereas literate poets like Virgil or Milton use longer and more complicated syntactical structures. Homer then expands on these ideas in subsequent clauses; this technique is called parataxis. The so-called 'type scenes' (typische Szenen), were named by Walter Arend in 1933. He noted that Homer often, when describing frequently recurring activities such as eating, praying, fighting and dressing, used blocks of set phrases in sequence that were then elaborated by the poet. The 'Analyst' school had considered these repetitions as un-Homeric, whereas Arend interpreted them philosophically. Parry and Lord noted that these conventions are found in many other cultures. 'Ring composition' or chiastic structure (when a phrase or idea is repeated at both the beginning and end of a story, or a series of such ideas first appears in the order A, B, C ... before being reversed as ... C, B, A) has been observed in the Homeric epics. Opinion differs as to whether these occurrences are a conscious artistic device, a mnemonic aid or a spontaneous feature of human storytelling. Both of the Homeric poems begin with an invocation to the Muse. In the Iliad, the poet beseeches her to sing of "the anger of Achilles", and, in the Odyssey, he asks her to tell of "the man of many ways". A similar opening was later employed by Virgil in his Aeneid. The orally transmitted Homeric poems were put into written form at some point between the eighth and sixth centuries BC. Some scholars believe that they were dictated to a scribe by the poet and that our inherited versions of the Iliad and Odyssey were in origin orally-dictated texts. Albert Lord noted that the Balkan bards that he was studying revised and expanded their songs in their process of dictating. Some scholars hypothesize that a similar process of revision and expansion occurred when the Homeric poems were first written down. Other scholars hold that, after the poems were created in the eighth century, they continued to be orally transmitted with considerable revision until they were written down in the sixth century. After textualisation, the poems were each divided into 24 rhapsodes, today referred to as books, and labelled by the letters of the Greek alphabet. Most scholars attribute the book divisions to the Hellenistic scholars of Alexandria, in Egypt. Some trace the divisions back further to the Classical period. Very few credit Homer himself with the divisions. In antiquity, it was widely held that the Homeric poems were collected and organised in Athens in the late sixth century BC by Pisistratus (died 528/7 BC), in what subsequent scholars have dubbed the "Peisistratean recension". The idea that the Homeric poems were originally transmitted orally and first written down during the reign of Pisistratus is referenced by the first-century BC Roman orator Cicero and is also referenced in a number of other surviving sources, including two ancient Lives of Homer. From around 150 BC, the texts of the Homeric poems seem to have become relatively established. After the establishment of the Library of Alexandria, Homeric scholars such as Zenodotus of Ephesus, Aristophanes of Byzantium and in particular Aristarchus of Samothrace helped establish a canonical text. The first printed edition of Homer was produced in 1488 in Milan, Italy. Today scholars use medieval manuscripts, papyri and other sources; some argue for a "multi-text" view, rather than seeking a single definitive text. The nineteenth-century edition of Arthur Ludwich mainly follows Aristarchus's work, whereas van Thiel's (1991, 1996) follows the medieval vulgate. Others, such as Martin West (1998–2000) or T. W. Allen, fall somewhere between these two extremes. Him with that falchion in his hand behold, Who comes before the three, even as their lord. That one is Homer, Poet sovereign; This is a partial list of translations into English of Homer's Iliad and Odyssey.
[[Image:Euro accession.svg|thumb|300px| {{legend|#0080C0|[[European Union|EU]] Eurozone (20)}} {{legend|#66BB66|EU states obliged to join the Eurozone (6)}} {{legend|#B24400|EU states with an opt-out on Eurozone participation (1)}} {{legend|#ffff00|States outside the EU with issuing rights (4)}} {{legend|purple|Other non-EU users of euro (2)}}]] The '''Eurozone''' (also called '''Euro area''' or '''Euroland''') is the set of countries in the [[European Union]] which have adopted the [[Euro]] (€) currency. The [[European Central Bank]] is responsible for managing the supply of money within the eurozone and political decisions are taken by the "euro group", which is a meeting of the politicians from each euro country in charge of that country's [[economy]]. EU members that are not part of the Eurozone are [[Denmark]], [[Sweden]], [[Poland]], [[Czech Republic]], [[Hungary]], [[Bulgaria]], and [[Romania]]. == Official members == There are 20 members in the Eurozone Made the area on 1 January 1999<br/> 1 [[Austria]]<br/> 2 [[Belgium]]<br/> 3 [[Finland]]<br/> 4 [[France]]<br/> 5 [[Germany]]<br/> 6 [[Republic of Ireland|Ireland]]<br/> 7 [[Italy]]<br/> 8 [[Luxembourg]]<br/> 9 [[Netherlands]]<br/> 10 [[Portugal]]<br/> 11 [[Spain]] Joined on 1 January 2001<br/> 12 [[Greece]] Joined on 1 January 2006<br/> 13 [[Slovenia]] Joined on 1 January 2008<br/> 14 [[Cyprus]]<br/> 15 [[Malta]] Joined on 1 January 2009<br/> 16 [[Slovakia]] Joined on 1 January 2011<br/> 17 [[Estonia]] Joined on 1 January 2014<br/> 18 [[Latvia]] Joined on 1 January 2015<br/> 19 [[Lithuania]] Joined on 1 January 2023<br/> 20 [[Croatia]] There are other countries outside the European Union which use the euro as well, but these are not officially in the eurozone. == Other websites == * [http://en.citizendium.org/wiki/Eurozone Eurozone] -Citizendium {{europe-stub}} [[Category:European Union]]
The euro area, commonly called the eurozone (EZ), is a currency union of 20 member states of the European Union (EU) that have adopted the euro (€) as their primary currency and sole legal tender, and have thus fully implemented EMU policies. The 20 eurozone members are: The seven non-eurozone members of the EU are Bulgaria, the Czech Republic, Denmark, Hungary, Poland, Romania, and Sweden. They continue to use their own national currencies, although all but Denmark are obliged to join once they meet the euro convergence criteria. Among non-EU member states, Andorra, Monaco, San Marino, and Vatican City have formal agreements with the EU to use the euro as their official currency and issue their own coins. In addition, Kosovo and Montenegro have adopted the euro unilaterally, relying on euros already in circulation rather than minting currencies of their own. These six countries, however, have no representation in any eurozone institution. The Eurosystem is the monetary authority of the eurozone, the Eurogroup is an informal body of finance ministers that makes fiscal policy for the currency union, and the European System of Central Banks is responsible for fiscal and monetary cooperation between eurozone and non-eurozone EU members. The European Central Bank (ECB) makes monetary policy for the eurozone, sets its base interest rate, and issues euro banknotes and coins. Since the financial crisis of 2007–2008, the eurozone has established and used provisions for granting emergency loans to member states in return for enacting economic reforms. The eurozone has also enacted some limited fiscal integration; for example, in peer review of each other's national budgets. The issue is political and in a state of flux in terms of what further provisions will be agreed for eurozone change. No eurozone member state has left, and there are no provisions to do so or to be expelled. In 1998, eleven member states of the European Union had met the euro convergence criteria, and the eurozone came into existence with the official launch of the euro (alongside national currencies) on 1 January 1999 in those countries: Austria, Belgium, Finland, France, Germany, Ireland, Italy, Luxembourg, the Netherlands, Portugal, and Spain. Greece qualified in 2000 and was admitted on 1 January 2001. These twelve founding members introduced physical euro banknotes and euro coins on 1 January 2002. After a short transition period, they took out of circulation and rendered invalid their pre-euro national coins and notes. Between 2007 and 2023, eight new states have acceded: Croatia, Cyprus, Estonia, Latvia, Lithuania, Malta, Slovakia, and Slovenia. Three French dependent territories that are not part of the EU have adopted the euro, with France ensuring eurozone laws are implemented: The euro is also used in countries outside the EU. Four states (Andorra, Monaco, San Marino, and Vatican City) have signed formal agreements with the EU to use the euro and issue their own coins. Nevertheless, they are not considered part of the eurozone by the ECB and do not have a seat in the ECB or Euro Group. Akrotiri and Dhekelia (located on the island of Cyprus) belong to the United Kingdom, but there are agreements between the UK and Cyprus and between UK and EU about their partial integration with Cyprus and partial adoption of Cypriot law, including the usage of euro in Akrotiri and Dhekelia. Several currencies are pegged to the euro, some of them with a fluctuation band and others with an exact rate. The Bosnia and Herzegovina convertible mark was once pegged to the Deutsche mark at par, and continues to be pegged to the euro today at the Deutsche mark's old rate (1.95583 per euro). The Bulgarian lev was initially pegged to the Deutsche Mark at a rate of BGL 1000 to DEM 1 in 1997, and has been pegged at a rate of BGN 1.95583 to EUR 1 since the introduction of the euro and the redenomination of the lev in 1999. The West African and Central African CFA francs are pegged exactly at 655.957 CFA to 1 EUR. In 1998, in anticipation of Economic and Monetary Union of the European Union, the Council of the European Union addressed the monetary agreements France had with the CFA Zone and Comoros, and ruled that the ECB had no obligation towards the convertibility of the CFA and Comorian francs. The responsibility of the free convertibility remained in the French Treasury. Kosovo and Montenegro officially adopted the euro as their sole currency without an agreement and, therefore, have no issuing rights. These states are not considered part of the eurozone by the ECB. However, sometimes the term eurozone is applied to all territories that have adopted the euro as their sole currency. Further unilateral adoption of the euro (euroisation), by both non-euro EU and non-EU members, is opposed by the ECB and EU. The chart below provides a full summary of all applying exchange-rate regimes for EU members, since the birth, on 13 March 1979, of the European Monetary System with its Exchange Rate Mechanism and the related new common currency ECU. On 1 January 1999, the euro replaced the ECU 1:1 at the exchange rate markets. During 1979–1999, the D-Mark functioned as a de facto anchor for the ECU, meaning there was only a minor difference between pegging a currency against the ECU and pegging it against the D-Mark. Sources: EC convergence reports 1996-2014, Italian lira, Spanish peseta, Portuguese escudo, Finnish markka, Greek drachma, Sterling The eurozone was born with its first 11 member states on 1 January 1999. The first enlargement of the eurozone, to Greece, took place on 1 January 2001, one year before the euro physically entered into circulation. The next enlargements were to states which joined the EU in 2004, and then joined the eurozone on 1 January of the year noted: Slovenia in 2007, Cyprus in 2008, Malta in 2008, Slovakia in 2009, Estonia in 2011, Latvia in 2014, and Lithuania in 2015. Croatia, which acceded to the EU in 2013, adopted the euro in 2023. All new EU members joining the bloc after the signing of the Maastricht Treaty in 1992 are obliged to adopt the euro under the terms of their accession treaties. However, the last of the five economic convergence criteria which need first to be complied with in order to qualify for euro adoption, is the exchange rate stability criterion, which requires having been an ERM-member for a minimum of two years without the presence of "severe tensions" for the currency exchange rate. In September 2011, a diplomatic source close to the euro adoption preparation talks with the seven remaining new member states who had yet to adopt the euro at that time (Bulgaria, the Czech Republic, Hungary, Latvia, Lithuania, Poland, and Romania), claimed that the monetary union (eurozone) they had thought they were going to join upon their signing of the accession treaty may very well end up being a very different union, entailing a much closer fiscal, economic, and political convergence than originally anticipated. This changed legal status of the eurozone could potentially cause them to conclude that the conditions for their promise to join were no longer valid, which "could force them to stage new referendums" on euro adoption. Seven countries (Bulgaria, the Czech Republic, Denmark, Hungary, Poland, Romania, and Sweden) are EU members but do not use the euro. Before joining the eurozone, a state must spend at least two years in the European Exchange Rate Mechanism (ERM II). As of January 2023, the Danish central bank and the Bulgarian central bank participate in ERM II. Denmark obtained a special opt-out in the original Maastricht Treaty, and thus is legally exempt from joining the eurozone unless its government decides otherwise, either by parliamentary vote or referendum. The United Kingdom likewise had an opt-out prior to withdrawing from the EU in 2020. The remaining six countries are obliged to adopt the euro in future, although the EU has so far not tried to enforce any time plan. They should join as soon as they fulfill the convergence criteria, which include being part of ERM II for two years. Sweden, which joined the EU in 1995 after the Maastricht Treaty was signed, is required to join the eurozone. However, the Swedish people turned down euro adoption in a 2003 referendum and since then the country has intentionally avoided fulfilling the adoption requirements by not joining ERM II, which is voluntary. Bulgaria joined ERM II on 10 July 2020. Interest in joining the eurozone increased in Denmark, and initially in Poland, as a result of the financial crisis of 2007–2008. In Iceland, there was an increase in interest in joining the European Union, a pre-condition for adopting the euro. However, by 2010 the debt crisis in the eurozone caused interest from Poland, as well as the Czech Republic, Denmark and Sweden to cool. In the opinion of journalist Leigh Phillips and Locke Lord's Charles Proctor, there is no provision in any European Union treaty for an exit from the eurozone. In fact, they argued, the Treaties make it clear that the process of monetary union was intended to be "irreversible" and "irrevocable". However, in 2009, a European Central Bank legal study argued that, while voluntary withdrawal is legally not possible, expulsion remains "conceivable". Although an explicit provision for an exit option does not exist, many experts and politicians in Europe have suggested an option to leave the eurozone should be included in the relevant treaties. On the issue of leaving the eurozone, the European Commission has stated that "[t]he irrevocability of membership in the euro area is an integral part of the Treaty framework and the Commission, as a guardian of the EU Treaties, intends to fully respect [that irrevocability]." It added that it "does not intend to propose [any] amendment" to the relevant Treaties, the current status being "the best way going forward to increase the resilience of euro area Member States to potential economic and financial crises. The European Central Bank, responding to a question by a Member of the European Parliament, has stated that an exit is not allowed under the Treaties. Likewise there is no provision for a state to be expelled from the euro. Some, however, including the Dutch government, favour the creation of an expulsion provision for the case whereby a heavily indebted state in the eurozone refuses to comply with an EU economic reform policy. In a Texas law journal, University of Texas at Austin law professor Jens Dammann has argued that even now EU law contains an implicit right for member states to leave the eurozone if they no longer meet the criteria that they had to meet in order to join it. Furthermore, he has suggested that, under narrow circumstances, the European Union can expel member states from the eurozone. The monetary policy of all countries in the eurozone is managed by the European Central Bank (ECB) and the Eurosystem which comprises the ECB and the central banks of the EU states who have joined the eurozone. Countries outside the eurozone are not represented in these institutions. Whereas all EU member states are part of the European System of Central Banks (ESCB), non EU member states have no say in all three institutions, even those with monetary agreements such as Monaco. The ECB is entitled to authorise the design and printing of euro banknotes and the volume of euro coins minted, and its president is currently Christine Lagarde. The eurozone is represented politically by its finance ministers, known collectively as the Eurogroup, and is presided over by a president, currently Paschal Donohoe. The finance ministers of the EU member states that use the euro meet a day before a meeting of the Economic and Financial Affairs Council (Ecofin) of the Council of the European Union. The Group is not an official Council formation but when the full EcoFin council votes on matters only affecting the eurozone, only Euro Group members are permitted to vote on it. Since the global financial crisis of 2007–2008, the Euro Group has met irregularly not as finance ministers, but as heads of state and government (like the European Council). It is in this forum, the Euro summit, that many eurozone reforms have been decided upon. In 2011, former French President Nicolas Sarkozy pushed for these summits to become regular and twice a year in order for it to be a 'true economic government'. In April 2008 in Brussels, future European Commission President Jean-Claude Juncker suggested that the eurozone should be represented at the IMF as a bloc, rather than each member state separately: "It is absurd for those 15 countries not to agree to have a single representation at the IMF. It makes us look absolutely ridiculous. We are regarded as buffoons on the international scene". In 2017 Juncker stated that he aims to have this agreed by the end of his mandate in 2019. However, Finance Commissioner Joaquín Almunia stated that before there is common representation, a common political agenda should be agreed upon. Leading EU figures including the commission and national governments have proposed a variety of reforms to the eurozone's architecture; notably the creation of a Finance Minister, a larger eurozone budget, and reform of the current bailout mechanisms into either a "European Monetary Fund" or a eurozone Treasury. While many have similar themes, details vary greatly. HICP figures from the ECB, overall index: Interest rates for the eurozone, set by the ECB since 1999. Levels are in percentages per annum. Between June 2000 and October 2008, the main refinancing operations were variable rate tenders, as opposed to fixed rate tenders. The figures indicated in the table from 2000 to 2008 refer to the minimum interest rate at which counterparties may place their bids. The following table states the ratio of public debt to GDP in percent for eurozone countries given by EuroStat. The euro convergence criterion is to not exceed 60%. The primary means for fiscal coordination within the EU lies in the Broad Economic Policy Guidelines which are written for every member state, but with particular reference to the 20 current members of the eurozone. These guidelines are not binding, but are intended to represent policy coordination among the EU member states, so as to take into account the linked structures of their economies. For their mutual assurance and stability of the currency, members of the eurozone have to respect the Stability and Growth Pact, which sets agreed limits on deficits and national debt, with associated sanctions for deviation. The Pact originally set a limit of 3% of GDP for the yearly deficit of all eurozone member states; with fines for any state which exceeded this amount. In 2005, Portugal, Germany, and France had all exceeded this amount, but the Council of Ministers had not voted to fine those states. Subsequently, reforms were adopted to provide more flexibility and ensure that the deficit criteria took into account the economic conditions of the member states, and additional factors. The Fiscal Compact (formally, the Treaty on Stability, Coordination and Governance in the Economic and Monetary Union), is an intergovernmental treaty introduced as a new stricter version of the Stability and Growth Pact, signed on 2 March 2012 by all member states of the European Union (EU), except the Czech Republic, the United Kingdom, and Croatia (subsequently acceding the EU in July 2013). The treaty entered into force on 1 January 2013 for the 16 states which completed ratification prior of this date. As of 1 April 2014, it had been ratified and entered into force for all 25 signatories. Olivier Blanchard suggests that a fiscal union in the eurozone can mitigate devastating effects of the single currency on the eurozone peripheral countries. But he adds that the currency bloc will not work perfectly even if a fiscal transfer system is built, because, he argues, the fundamental issue about competitiveness adjustment is not tackled. The problem is, since the eurozone peripheral countries do not have their own currencies, they are forced to adjust their economies by decreasing their wages instead of devaluation. The financial crisis of 2007–2008 prompted a number of reforms in the eurozone. One was a U-turn on the eurozone's bailout policy that led to the creation of a specific fund to assist eurozone states in trouble. The European Financial Stability Facility (EFSF) and the European Financial Stability Mechanism (EFSM) were created in 2010 to provide, alongside the International Monetary Fund (IMF), a system and fund to bail out members. However, the EFSF and EFSM were temporary, small and lacked a basis in the EU treaties. Therefore, it was agreed in 2011 to establish a European Stability Mechanism (ESM) which would be much larger, funded only by eurozone states (not the EU as a whole as the EFSF/EFSM were) and would have a permanent treaty basis. As a result of that its creation involved agreeing an amendment to TEFU Article 136 allowing for the ESM and a new ESM treaty to detail how the ESM would operate. If both are successfully ratified according to schedule, the ESM would be operational by the time the EFSF/EFSM expire in mid-2013. In February 2016, the UK secured further confirmation that countries that do not use the Euro would not be required to contribute to bailouts for eurozone countries. In June 2010, a broad agreement was finally reached on a controversial proposal for member states to peer review each other's budgets prior to their presentation to national parliaments. Although showing the entire budget to each other was opposed by Germany, Sweden and the UK, each government would present to their peers and the Commission their estimates for growth, inflation, revenue and expenditure levels six months before they go to national parliaments. If a country was to run a deficit, they would have to justify it to the rest of the EU while countries with a debt more than 60% of GDP would face greater scrutiny. The plans would apply to all EU members, not just the eurozone, and have to be approved by EU leaders along with proposals for states to face sanctions before they reach the 3% limit in the Stability and Growth Pact. Poland has criticised the idea of withholding regional funding for those who break the deficit limits, as that would only impact the poorer states. In June 2010 France agreed to back Germany's plan for suspending the voting rights of members who breach the rules. In March 2011 was initiated a new reform of the Stability and Growth Pact aiming at straightening the rules by adopting an automatic procedure for imposing of penalties in case of breaches of either the deficit or the debt rules. In 1997, Arnulf Baring expressed concern that the European Monetary Union would make Germans the most hated people in Europe. Baring suspected the possibility that the people in Mediterranean countries would regard Germans and the currency bloc as economic policemen. In 2001, James Tobin thought that the euro project would not succeed without making drastic changes to European institutions, pointing out the difference between the US and the eurozone. Concerning monetary policies, the system of Federal Reserve banks in the US aims at both growth and reducing unemployment, while the ECB tends to give its first priority to price stability under the Bundesbank's supervision. As the price level of the currency bloc is kept low, the unemployment level of the region has become higher than that of the US since 1982. Concerning fiscal policies, 12% of the US federal budget is used for transfers to states and local governments. The US government does not impose restrictions on state budget policies, whereas the Treaty of Maastricht requires each eurozone member country to keep its budget deficit below 3% of its GDP. In 2008, a study by Alberto Alesina and Vincenzo Galasso found that the adoption of euro promoted market deregulation and market liberalization. Furthermore, the euro was also linked to wage moderation, as wage growth slowed down in countries that adopted the new currency. Oliver Hart, who received the Nobel Memorial Prize in Economic Sciences in 2016, criticized the euro, calling it a "mistake" and emphasising his opposition to monetary union since its inception. He also expressed opposition to European integration, arguing that the European Union should instead focus on decentralisation as it has “gone too far in centralising power”. In 2018, a study based on DiD methodology found that the adoption of euro produced no systematic growth effects, as no growth-enhancing effects were found when compared to European economies outside the eurozone. The eurozone has also been criticized for deepening inequality in Europe, particularly between the richest and poorest countries. According to a study by Bertelsmann Stiftung, countries such as Austria and the Netherlands benefited significantly from the common currency, while southern and eastern European members of the eurozone gained very little, and some countries are considered to have suffered adverse effects from adopting the euro. In an article for the Politico, Joseph Stiglitz argues: "The result for the eurozone has been slower growth, and especially for the weaker countries within it. The euro was supposed to usher in greater prosperity, which in turn would lead to renewed commitment to European integration. It has done just the opposite — increasing divisions within the EU, especially between creditor and debtor countries." Matthias Matthijs believes that the euro resulted in a "winner-take-all" economy, as national income differences between eurozone members have widened further. He argues that countries such as Austria and Germany have gained from the eurozone at the expense of southern countries like Italy and Spain. By adopting the euro and abandoning their national currencies, eurozone countries gave up their ability to conduct independent monetary policy; as such, monetary policies used to combat recession, such as monetary stimulus or currency devaluation, are no longer available. During the European debt crisis, several eurozone countries (Greece, Italy, Portugal, Ireland, Spain, and Cyprus) were unable to repay their debt without third-party intervention by the European Central Bank and the International Monetary Fund. In order to grant the bailout, the ECB and the IMF forced the affected countries to adopt strict austerity measures. The European bailouts were largely about shifting exposure from banks onto European taxpayers, and exacerbated issues such as high unemployment and poverty. In 2019, a study from the Centre for European Policy concluded that while some countries had gained from adopting the euro, several countries were poorer than they would have been had they not adopted it, with France and Italy being particularly affected. The publication prompted a large number of reactions, pushing its authors to put out a statement clarifying some points. In 2020, a study from the University of Bonn reached a different conclusion: the adoption of the euro made “some mild losers (France, Germany, Italy, and Portugal) and a clear winner (Ireland)”. Both studies used the synthetic control method to estimate what might have happened if the euro hadn't been adopted.
The word '''Gnome''' can mean different things: * A Gnome is a [[creature]] of [[myth]]. ** Gnomes are [[element|elemental]] [[spirit]]s of [[earth]]. ** Gnomes are tiny, shaped like humans, with long beards. A gnome is like a [[Dwarf (mythology)|dwarf]] but smaller. Gnomes like [[gems]] and [[mining]]. * The [[GNOME]] desktop is a [[desktop environment]] for [[computer]]s running [[Unix]] and Unix-like computer operating systems. {{disambig}}
A gnome (/noʊm/) is a mythological creature and diminutive spirit in Renaissance magic and alchemy, introduced by Paracelsus in the 16th century and widely adopted by authors including those of modern fantasy literature. Typically small humanoids who live underground, gnome characteristics are reinterpreted to suit various storytellers and artists. Lawn ornaments crafted as gnomes were introduced during the 19th century, growing in popularity during the 20th century as garden gnomes. The word comes from Renaissance Latin gnomus, which first appears in A Book on Nymphs, Sylphs, Pygmies, and Salamanders, and on the Other Spirits by Paracelsus, published posthumously in Nysa in 1566 (and again in the Johannes Huser edition of 1589–1591 from an autograph by Paracelsus). The term may be an original invention of Paracelsus, possibly deriving the term from Latin gēnomos (itself representing a Greek γη-νομος, approximately "gē-nomos", literally "earth-dweller"). In this case, the omission of the ē is referred to as a blunder by the Oxford English Dictionary (OED). Paracelsus uses Gnomi as a synonym of Pygmæi and classifies them as earth elementals. He describes them as two spans high, very reluctant to interact with humans and able to move through solid earth as easily as humans move through air. The chthonic or earth-dwelling spirit has precedents in numerous ancient and medieval mythologies, often guarding mines and precious underground treasures, notably in the Germanic dwarfs and the Greek Chalybes, Telchines or Dactyls. The gnomes of Swiss folklore follow this template, as they are said to have caused the landslide that destroyed the Swiss village of Plurs in 1618 - the villagers had become wealthy from a local gold mine created by the gnomes, who poured liquid gold down into a vein for the benefit of humans, and were corrupted by this newfound prosperity, which greatly offended the gnomes. The English word is attested from the early 18th century. Gnomes are used in Alexander Pope's "The Rape of the Lock". The creatures from this mock-epic are small, celestial creatures which were prudish women in their past lives, and now spend all of eternity looking out for prudish women (in parallel to the guardian angels in Catholic belief). Other uses of the term gnome remain obscure until the early 19th century, when it is taken up by authors of Romanticist collections of fairy tales and becomes mostly synonymous with the older word goblin. Pope's stated source, the 1670 French satire Comte de Gabalis by Nicolas-Pierre-Henri de Montfaucon de Villars, the abbot of Villars, describes gnomes as such: The Earth is filled almost to the center with Gnomes or Pharyes, a people of small stature, the guardians of treasures, of mines, and of precious stones. They are ingenious, friends of men, and easie (sic) to be commandded (sic). They furnish the children of the Sages with as much money, as they have need of; and never ask any other reward of their services, than the glory of being commanded. The Gnomides or wives of these Gnomes or Pharyes, are little, but very handsom (sic); and their habit marvellously (sic) curious. De Villars used the term gnomide to refer to female gnomes (often "gnomid" in English translations). Modern fiction instead uses the word "gnomess" to refer to female gnomes. In 19th-century fiction, the chthonic gnome became a sort of antithesis to the more airy or luminous fairy. Nathaniel Hawthorne in Twice-Told Tales (1837) contrasts the two in "Small enough to be king of the fairies, and ugly enough to be king of the gnomes" (cited after OED). Similarly, gnomes are contrasted to elves, as in William Cullen Bryant's Little People of the Snow (1877), which has "let us have a tale of elves that ride by night, with jingling reins, or gnomes of the mine" (cited after OED). The Russian composer Mussorgsky produced a movement in his work Pictures at an Exhibition, (1874) named "Gnomus" (Latin for "The Gnome"). It is written to sound as if a gnome is moving about. Franz Hartmann in 1895 satirized materialism in an allegorical tale entitled Unter den Gnomen im Untersberg. The English translation appeared in 1896 as Among the Gnomes: An Occult Tale of Adventure in the Untersberg. In this story, the Gnomes are still clearly subterranean creatures, guarding treasures of gold within the Untersberg mountain. As a figure of 19th-century fairy tales, the term gnome became largely synonymous with other terms for "little people" by the 20th century, such as goblin, brownie, leprechaun and other instances of the household spirit type, losing its strict association with earth or the underground world. The 1967 Walt Disney movie The Gnome-Mobile The 2011 animated movie Gnomeo & Juliet The 2018 animated movie Sherlock Gnomes featured gnomish versions of several classic Sherlock Holmes characters. After World War II (with early references, in ironic use, from the late 1930s) the diminutive figurines introduced as lawn ornaments during the 19th century came to be known as garden gnomes. The image of the gnome changed further during the 1960s to 1970s, when the first plastic garden gnomes were manufactured. These gnomes followed the style of the 1937 depiction of the seven dwarves in Snow White and the Seven Dwarfs by Disney. This "Disneyfied" image of the gnome was built upon by the illustrated children's book classic Gnomes (1976), in the original Dutch Leven en werken van de Kabouter, by author Wil Huygen and artist Rien Poortvliet, followed in 1981 by The Secret Book of Gnomes. Garden gnomes share a resemblance to the Scandinavian tomte and nisse, and the Swedish term "tomte" can be translated as "gnome" in English. Several gnome themed entertainment parks exist. Notable ones are: Gnome parades are held annually at Atlanta's Inman Park Festival. Numerous one-off gnome parades have been held, including in Savannah, Georgia (April 2012) and Cleveland, Ohio (May 2011).
{{More citations needed|date=January 2025}} A '''republic''' is a form of [[government]] that has no [[monarchy]] and no hereditary [[aristocracy]].<ref>{{Cite web|url=http://www.ushistory.org/civ/6a.asp|title=The Roman Republic [ushistory.org] |website=www.ushistory.org |access-date=2019-03-14}}</ref> It originates from Rome. In 509 BC, the Romans overthrew the [[Roman Kingdom]] and established a republic, a government in which citizens elected representatives to rule on their behalf. National [[sovereignty]] lies in the authority of the government, not in an emperor or monarch. The word ''republic'' comes from the [[Latin]] words '''res publica''', meaning a "public thing". For example, the [[United States]] and [[India]] are republics. However, the [[United Kingdom]] and [[Canada]] are not republics since they have a monarch (King Charles III in both cases). Countries with a king or other [[monarch]] and free elections is called a [[constitutional monarchy]], not republics. A constitutional monarchy resembles a republic because the constitution has been amended to remove power from the monarch and install institutions conforming to a philosophy of republicanism. That includes the [[United Kingdom]] and the other [[Commonwealth realm]]s, the [[Netherlands]], [[Thailand]], and countries in [[Scandinavia]] and elsewhere. What makes a republic different is that its laws are made and enforced without royal authority. The [[head of state]] in a republic is generally a person who has been chosen by the [[citizen]]s by [[direct election]] or by a group of elected representatives to act as the top representative of the people. In most republics, the head of state is called the [[president]]. In some countries, the president is elected and has a lot of political power. In others, the president does not hold much direct power but is important in the legal system for other reasons. Sometimes, state is called a "republic" when its head is not called a "king". For example, the [[Roman Empire]] had an "emperor" and the [[Dutch Republic]] had a "stadholder", but they worked as [[hereditary monarchy|hereditary monarchs]]. [[Licchavi]], in [[India]], was an aristocratic state in the 6th and 5th centuries BCE and is nowadays sometimes called the first republic. The earliest republics that were much imitated later were [[Classical Greece|Greek]] cities, for example [[democracy in Athens]]. The biggest difference from the other [[polis|city-states]] was that people chose their leaders by [[voting]] or by [[lottery]]. Several Italian city-states during the [[Renaissance]] were ruled by a small group of aristocrats, and were called republics. ==References== {{reflist}} {{Politics and government}} [[Category:Forms of government]]
A republic, based on the Latin phrase res publica ("public affair"), is a state in which political power rests with the public and their representatives—in contrast to a monarchy. Representation in a republic may or may not be freely elected by the general citizenry. In many historical republics, representation has been based on personal status and the role of elections has been limited. This remains true today; among the 159 states that use the word "republic" in their official names as of 2017, and other states formally constituted as republics, are states that narrowly constrain both the right of representation and the process of election. The term developed its modern meaning in reference to the constitution of the ancient Roman Republic, lasting from the overthrow of the kings in 509 BC to the establishment of the Empire in 27 BC. This constitution was characterized by a Senate composed of wealthy aristocrats wielding significant influence; several popular assemblies of all free citizens, possessing the power to elect magistrates and pass laws; and a series of magistracies with varying types of civil and political authority. Most often a republic is a single sovereign state, but there are also subnational state entities that are referred to as republics, or that have governments that are described as republican in nature. The term originates from the Latin translation of Greek word politeia. Cicero, among other Latin writers, translated politeia as res publica and it was in turn translated by Renaissance scholars as "republic" (or similar terms in various European languages). The term politeia can be translated as form of government, polity, or regime and is therefore not always a word for a specific type of regime as the modern word republic is. One of Plato's major works on political science was titled Politeia and in English it is thus known as The Republic. However, apart from the title, in modern translations of The Republic, alternative translations of politeia are also used. However, in Book III of his Politics, Aristotle was apparently the first classical writer to state that the term politeia can be used to refer more specifically to one type of politeia: "When the citizens at large govern for the public good, it is called by the name common to all governments (to koinon onoma pasōn tōn politeiōn), government (politeia)". Also amongst classical Latin, the term "republic" can be used in a general way to refer to any regime, or in a specific way to refer to governments which work for the public good. In medieval Northern Italy, a number of city states had commune or signoria based governments. In the late Middle Ages, writers such as Giovanni Villani began writing about the nature of these states and the differences from other types of regime. They used terms such as libertas populi, a free people, to describe the states. The terminology changed in the 15th century as the renewed interest in the writings of Ancient Rome caused writers to prefer using classical terminology. To describe non-monarchical states, writers (most importantly, Leonardo Bruni) adopted the Latin phrase res publica. While Bruni and Machiavelli used the term to describe the states of Northern Italy, which were not monarchies, the term res publica has a set of interrelated meanings in the original Latin. The term can quite literally be translated as "public matter". It was most often used by Roman writers to refer to the state and government, even during the period of the Roman Empire. In subsequent centuries, the English word "commonwealth" came to be used as a translation of res publica, and its use in English was comparable to how the Romans used the term res publica. Notably, during The Protectorate of Oliver Cromwell the word commonwealth was the most common term to call the new monarchless state, but the word republic was also in common use. Likewise, in Polish the term was translated as rzeczpospolita, although the translation is now only used with respect to Poland. Presently, the term "republic" commonly means a system of government which derives its power from the people rather than from another basis, such as heredity or divine right. While the philosophical terminology developed in classical Greece and Rome, as already noted by Aristotle there was already a long history of city states with a wide variety of constitutions, not only in Greece but also in the Middle East. After the classical period, during the Middle Ages, many free cities developed again, such as Venice. The modern type of "republic" itself is different from any type of state found in the classical world. Nevertheless, there are a number of states of the classical era that are today still called republics. This includes ancient Athens and the Roman Republic. While the structure and governance of these states was different from that of any modern republic, there is debate about the extent to which classical, medieval, and modern republics form a historical continuum. J. G. A. Pocock has argued that a distinct republican tradition stretches from the classical world to the present. Other scholars disagree. Paul Rahe, for instance, argues that the classical republics had a form of government with few links to those in any modern country. The political philosophy of the classical republics has influenced republican thought throughout the subsequent centuries. Philosophers and politicians advocating republics, such as Machiavelli, Montesquieu, Adams, and Madison, relied heavily on classical Greek and Roman sources which described various types of regimes. Aristotle's Politics discusses various forms of government. One form Aristotle named politeia, which consisted of a mixture of the other forms, oligarchy and democracy. He argued that this was one of the ideal forms of government. Polybius expanded on many of these ideas, again focusing on the idea of mixed government and differentiated basic forms of government between "benign" monarchy, aristocracy, and democracy, and the "malignant" tyranny, oligarchy, and ochlocracy. The most important Roman work in this tradition is Cicero's De re publica. Over time, the classical republics became empires or were conquered by empires. Most of the Greek republics were annexed to the Macedonian Empire of Alexander. The Roman Republic expanded dramatically conquering the other states of the Mediterranean that could be considered republics, such as Carthage. The Roman Republic itself then became the Roman Empire. The term "republic" is not commonly used to refer to pre-classical city-states, especially if outside Europe and the area which was under Graeco-Roman influence. However some early states outside Europe had governments that are sometimes today considered similar to republics. In the ancient Near East, a number of cities of the Eastern Mediterranean achieved collective rule. Republic city-states flourished in Phoenicia along the Levantine coast starting from the 11th century BC. In ancient Phoenicia, the concept of Shophet was very similar to a Roman consul. Under Persian rule (539–332 BC), Phoenician city-states such as Tyre abolished the king system and adopted "a system of the suffetes (judges), who remained in power for short mandates of 6 years". Arwad has been cited as one of the earliest known examples of a republic, in which the people, rather than a monarch, are described as sovereign. The Israelite confederation of the era of the Judges before the United Monarchy has also been considered a type of republic. The system of government of the Igbo people in what is now Nigeria has been described as "direct and participatory democracy." Early republican institutions come from the independent gaṇasaṅghas—gaṇa means "tribe" and saṅgha means "assembly"—which may have existed as early as the 6th century BC and persisted in some areas until the 4th century AD in India. The evidence for this is scattered, however, and no pure historical source exists for that period. Diodorus, a Greek historian who wrote two centuries after the time of Alexander the Great's invasion of India (now Pakistan and northwest India) mentions, without offering any detail, that independent and democratic states existed in India. Modern scholars note the word democracy at the time of the 3rd century BC and later suffered from degradation and could mean any autonomous state, no matter how oligarchic in nature. Key characteristics of the gaṇa seem to include a monarch, usually known by the name raja, and a deliberative assembly. The assembly met regularly. It discussed all major state decisions. At least in some states, attendance was open to all free men. This body also had full financial, administrative, and judicial authority. Other officers, who rarely receive any mention, obeyed the decisions of the assembly. Elected by the gaṇa, the monarch apparently always belonged to a family of the noble class of Kshatriya Varna. The monarch coordinated his activities with the assembly; in some states, he did so with a council of other nobles. The Licchavis had a primary governing body of 7,077 rajas, the heads of the most important families. On the other hand, the Shakyas, Koliyas, Mallakas, and Licchavis, during the period around Gautama Buddha, had the assembly open to all men, rich and poor. Early "republics" or gaṇasaṅgha, such as Mallakas, centered in the city of Kusinagara, and the Vajjika (or Vṛjika) League, centered in the city of Vaishali, existed as early as the 6th century BC and persisted in some areas until the 4th century AD. The most famous clan amongst the ruling confederate clans of the Vajji Mahajanapada were the Licchavis. The Magadha kingdom included republican communities such as the community of Rajakumara. Villages had their own assemblies under their local chiefs called Gramakas. Their administrations were divided into executive, judicial, and military functions. Scholars differ over how best to describe these governments, and the vague, sporadic quality of the evidence allows for wide disagreements. Some emphasize the central role of the assemblies and thus tout them as democracies; other scholars focus on the upper-class domination of the leadership and possible control of the assembly and see an oligarchy or an aristocracy. Despite the assembly's obvious power, it has not yet been established whether the composition and participation were truly popular. This is reflected in the Arthashastra, an ancient handbook for monarchs on how to rule efficiently. It contains a chapter on how to deal with the saṅghas, which includes injunctions on manipulating the noble leaders, yet it does not mention how to influence the mass of the citizens, indicating that the "gaṇasaṅgha" are more of an aristocratic rule, or oligarchic republic, than "democracy". The Icelandic Commonwealth was established in 930 AD by refugees from Norway who had fled the unification of that country under King Harald Fairhair. The Commonwealth consisted of a number of clans run by chieftains, and the Althing was a combination of parliament and supreme court where disputes appealed from lower courts were settled, laws were decided, and decisions of national importance were taken. One such example was the Christianisation of Iceland in 1000, where the Althing decreed that all Icelanders must be baptized into Christianity, and forbade celebration of pagan rituals. Contrary to most states, the Icelandic Commonwealth had no official leader. In the early 13th century, the Age of the Sturlungs, the Commonwealth began to suffer from long conflicts between warring clans. This, combined with pressure from the Norwegian king Haakon IV for the Icelanders to rejoin the Norwegian "family", led the Icelandic chieftains to accept Haakon IV as king by the signing of the Gamli sáttmáli ("Old Covenant") in 1262. This effectively brought the Commonwealth to an end. The Althing, however, is still Iceland's parliament, almost 800 years later. In Europe new republics appeared in the late Middle Ages when a number of small states embraced republican systems of government. These were generally small, but wealthy, trading states, like the Mediterranean maritime republics and the Hanseatic League, in which the merchant class had risen to prominence. Knud Haakonssen has noted that, by the Renaissance, Europe was divided with those states controlled by a landed elite being monarchies and those controlled by a commercial elite being republics. Italy was the most densely populated area of Europe, and also one with the weakest central government. Many of the towns thus gained considerable independence and adopted commune forms of government. Completely free of feudal control, the Italian city-states expanded, gaining control of the rural hinterland. The two most powerful were the Republic of Venice and its rival the Republic of Genoa. Each were large trading ports, and further expanded by using naval power to control large parts of the Mediterranean. It was in Italy that an ideology advocating for republics first developed. Writers such as Bartholomew of Lucca, Brunetto Latini, Marsilius of Padua, and Leonardo Bruni saw the medieval city-states as heirs to the legacy of Greece and Rome. Across Europe a wealthy merchant class developed in the important trading cities. Despite their wealth they had little power in the feudal system dominated by the rural land owners, and across Europe began to advocate for their own privileges and powers. The more centralized states, such as France and England, granted limited city charters. In the more loosely governed Holy Roman Empire, 51 of the largest towns became free imperial cities. While still under the dominion of the Holy Roman Emperor most power was held locally and many adopted republican forms of government. The same rights to imperial immediacy were secured by the major trading cities of Switzerland. The towns and villages of alpine Switzerland had, courtesy of geography, also been largely excluded from central control. Unlike Italy and Germany, much of the rural area was thus not controlled by feudal barons, but by independent farmers who also used communal forms of government. When the Habsburgs tried to reassert control over the region both rural farmers and town merchants joined the rebellion. The Swiss were victorious, and the Swiss Confederacy was proclaimed, and Switzerland has retained a republican form of government to the present. Two Russian cities with a powerful merchant class—Novgorod and Pskov—also adopted republican forms of government in 12th and 13th centuries, respectively, which ended when the republics were conquered by Muscovy/Russia at the end of 15th – beginning of 16th century. The dominant form of government for these early republics was control by a limited council of elite patricians. In those areas that held elections, property qualifications or guild membership limited both who could vote and who could run. In many states no direct elections were held and council members were hereditary or appointed by the existing council. This left the great majority of the population without political power, and riots and revolts by the lower classes were common. The late Middle Ages saw more than 200 such risings in the towns of the Holy Roman Empire. Similar revolts occurred in Italy, notably the Ciompi Revolt in Florence. Following the collapse of the Seljuk Sultanate of Rum and establishment of the Turkish Anatolian Beyliks, the Ahiler merchant fraternities established a state centered on Ankara that is sometimes compared to the Italian mercantile republics. While the classical writers had been the primary ideological source for the republics of Italy, in Northern Europe, the Protestant Reformation would be used as justification for establishing new republics. Most important was Calvinist theology, which developed in the Swiss Confederacy, one of the largest and most powerful of the medieval republics. John Calvin did not call for the abolition of monarchy, but he advanced the doctrine that the faithful had the duty to overthrow irreligious monarchs. Advocacy for republics appeared in the writings of the Huguenots during the French Wars of Religion. Calvinism played an important role in the republican revolts in England and the Netherlands. Like the city-states of Italy and the Hanseatic League, both were important trading centres, with a large merchant class prospering from the trade with the New World. Large parts of the population of both areas also embraced Calvinism. During the Dutch Revolt (beginning in 1566), the Dutch Republic emerged from rejection of Spanish Habsburg rule. However, the country did not adopt the republican form of government immediately: in the formal declaration of independence (Act of Abjuration, 1581), the throne of king Philip was only declared vacant, and the Dutch magistrates asked the Duke of Anjou, queen Elizabeth of England and prince William of Orange, one after another, to replace Philip. It took until 1588 before the Estates (the Staten, the representative assembly at the time) decided to vest the sovereignty of the country in themselves. In 1641 the English Civil War began. Spearheaded by the Puritans and funded by the merchants of London, the revolt was a success, and King Charles I was executed. In England James Harrington, Algernon Sidney, and John Milton became some of the first writers to argue for rejecting monarchy and embracing a republican form of government. The English Commonwealth was short-lived, and the monarchy was soon restored. The Dutch Republic continued in name until 1795, but by the mid-18th century the stadtholder had become a de facto monarch. Calvinists were also some of the earliest settlers of the British and Dutch colonies of North America. Along with these initial republican revolts, early modern Europe also saw a great increase in monarchical power. The era of absolute monarchy replaced the limited and decentralized monarchies that had existed in most of the Middle Ages. It also saw a reaction against the total control of the monarch as a series of writers created the ideology known as liberalism. Most of these Enlightenment thinkers were far more interested in ideas of constitutional monarchy than in republics. The Cromwell regime had discredited republicanism, and most thinkers felt that republics ended in either anarchy or tyranny. Thus philosophers like Voltaire opposed absolutism while at the same time being strongly pro-monarchy. Jean-Jacques Rousseau and Montesquieu praised republics, and looked on the city-states of Greece as a model. However, both also felt that a state like France, with 20 million people, would be impossible to govern as a republic. Rousseau admired the republican experiment in Corsica (1755–1769) and described his ideal political structure of small, self-governing communes. Montesquieu felt that a city-state should ideally be a republic, but maintained that a limited monarchy was better suited to a state with a larger territory. The American Revolution began as a rejection only of the authority of the British Parliament over the colonies, not of the monarchy. The failure of the British monarch to protect the colonies from what they considered the infringement of their rights to representative government, the monarch's branding of those requesting redress as traitors, and his support for sending combat troops to demonstrate authority resulted in widespread perception of the British monarchy as tyrannical. With the United States Declaration of Independence the leaders of the revolt firmly rejected the monarchy and embraced republicanism. The leaders of the revolution were well versed in the writings of the French liberal thinkers, and also in history of the classical republics. John Adams had notably written a book on republics throughout history. In addition, the widely distributed and popularly read-aloud tract Common Sense, by Thomas Paine, succinctly and eloquently laid out the case for republican ideals and independence to the larger public. The Constitution of the United States, went into effect in 1789, created a relatively strong federal republic to replace the relatively weak confederation under the first attempt at a national government with the Articles of Confederation and Perpetual Union ratified in 1781. The first ten amendments to the Constitution, called the United States Bill of Rights, guaranteed certain natural rights fundamental to republican ideals that justified the Revolution. The French Revolution was also not republican at its outset. Only after the Flight to Varennes removed most of the remaining sympathy for the king was a republic declared and Louis XVI sent to the guillotine. The stunning success of France in the French Revolutionary Wars saw republics spread by force of arms across much of Europe as a series of client republics were set up across the continent. The rise of Napoleon saw the end of the French First Republic and her Sister Republics, each replaced by "popular monarchies". Throughout the Napoleonic period, the victors extinguished many of the oldest republics on the continent, including the Republic of Venice, the Republic of Genoa, and the Dutch Republic. They were eventually transformed into monarchies or absorbed into neighboring monarchies. Outside Europe another group of republics was created as the Napoleonic Wars allowed the states of Latin America to gain their independence. Liberal ideology had only a limited impact on these new republics. The main impetus was the local European descended Creole population in conflict with the Peninsulares—governors sent from overseas. The majority of the population in most of Latin America was of either African or Amerindian descent, and the Creole elite had little interest in giving these groups power and broad-based popular sovereignty. Simón Bolívar, both the main instigator of the revolts and one of its most important theorists, was sympathetic to liberal ideals but felt that Latin America lacked the social cohesion for such a system to function and advocated autocracy as necessary. In Mexico this autocracy briefly took the form of a monarchy in the First Mexican Empire. Due to the Peninsular War, the Portuguese court was relocated to Brazil in 1808. Brazil gained independence as a monarchy on September 7, 1822, and the Empire of Brazil lasted until 1889. In many other Latin American states various forms of autocratic republic existed until most were liberalized at the end of the 20th century. The French Second Republic was created in 1848, but abolished by Napoleon III who proclaimed himself Emperor in 1852. The French Third Republic was established in 1870, when a civil revolutionary committee refused to accept Napoleon III's surrender during the Franco-Prussian War. Spain briefly became the First Spanish Republic in 1873–74, but the monarchy was soon restored. By the start of the 20th century France, Switzerland and San Marino remained the only republics in Europe. This changed when, after the 1908 Lisbon Regicide, the 5 October 1910 revolution established the Portuguese Republic. In East Asia, China had seen considerable anti-Qing sentiment during the 19th century, and a number of protest movements developed calling for constitutional monarchy. The most important leader of these efforts was Sun Yat-sen, whose Three Principles of the People combined American, European, and Chinese ideas. Under his leadership the Republic of China was proclaimed on January 1, 1912. Republicanism expanded significantly in the aftermath of World War I, when several of the largest European empires collapsed: the Russian Empire (1917), German Empire (1918), Austro-Hungarian Empire (1918), and Ottoman Empire (1922) were all replaced by republics. New states gained independence during this turmoil, and many of these, such as Ireland, Poland, Finland and Czechoslovakia, chose republican forms of government. Following Greece's defeat in the Greco-Turkish War (1919–22), the monarchy was briefly replaced by the Second Hellenic Republic (1924–35). In 1931, the proclamation of the Second Spanish Republic (1931–39) resulted in the Spanish Civil War that would be the prelude of World War II. Republican ideas were spreading, especially in Asia. The United States began to have considerable influence in East Asia in the later part of the 19th century, with Protestant missionaries playing a central role. The liberal and republican writers of the west also exerted influence. These combined with native Confucian inspired political philosophy that had long argued that the populace had the right to reject unjust governments that had lost the Mandate of Heaven. Two short-lived republics were proclaimed in East Asia, the Republic of Formosa and the First Philippine Republic. In the years following World War II, most of the remaining European colonies gained their independence, and most became republics. The two largest colonial powers were France and the United Kingdom. Republican France encouraged the establishment of republics in its former colonies. The United Kingdom attempted to follow the model it had for its earlier settler colonies of creating independent Commonwealth realms still linked under the same monarch. While most of the settler colonies and the smaller states of the Caribbean retained this system, it was rejected by the newly independent countries in Africa and Asia, which revised their constitutions and became republics instead. Britain followed a different model in the Middle East; it installed local monarchies in several colonies and mandates including Iraq, Jordan, Kuwait, Bahrain, Oman, Yemen and Libya. In subsequent decades revolutions and coups overthrew a number of monarchs and installed republics. Several monarchies remain, and the Middle East is the only part of the world where several large states are ruled by monarchs with almost complete political control. In the wake of the First World War, the Russian monarchy fell during the Russian Revolution. The Russian Provisional Government was established in its place on the lines of a liberal republic, but this was overthrown by the Bolsheviks who went on to establish the Union of Soviet Socialist Republics (USSR). This was the first republic established under Marxist–Leninist ideology. Communism was wholly opposed to monarchy, and became an important element of many republican movements during the 20th century. The Russian Revolution spread into Mongolia, and overthrew its theocratic monarchy in 1924. In the aftermath of the Second World War the communists gradually gained control of Romania, Bulgaria, Yugoslavia, Hungary and Albania, ensuring that the states were reestablished as socialist republics rather than monarchies. Communism also intermingled with other ideologies. It was embraced by many national liberation movements during decolonization. In Vietnam, communist republicans pushed aside the Nguyễn dynasty, and monarchies in neighbouring Laos and Cambodia were overthrown by communist movements in the 1970s. Arab socialism contributed to a series of revolts and coups that saw the monarchies of Egypt, Iraq, Libya, and Yemen ousted. In Africa, Marxism–Leninism and African socialism led to the end of monarchy and the proclamation of republics in states such as Burundi and Ethiopia. Islamic political philosophy has a long history of opposition to absolute monarchy, notably in the work of Al-Farabi. Sharia law took precedence over the will of the ruler, and electing rulers by means of the Shura was an important doctrine. While the early caliphate maintained the principles of an elected ruler, later states became hereditary or military dictatorships though many maintained some pretense of a consultative shura. None of these states are typically referred to as republics. The current usage of republic in Muslim countries is borrowed from the western meaning, adopted into the language in the late 19th century. The 20th century saw republicanism become an important idea in much of the Middle East, as monarchies were removed in many states of the region. Iraq became a secular state. Some nations, such as Indonesia and Azerbaijan, began as secular. In Iran, the 1979 revolution overthrew the monarchy and created an Islamic republic based on the ideas of Islamic democracy. A republic has not necessarily a constitution but is often constitutional in the sense of constitutionalism, meaning that it is constituted by a set of institutions which provide a separation of powers. The term constitutional republic is a way to highlight an emphasis of the separation of powers in a given republic, as with constitutional monarchy or absolute monarchy highlighting the absolute autocratic character of a monarchy. With no monarch, most modern republics use the title president for the head of state. Originally used to refer to the presiding officer of a committee or governing body in Great Britain the usage was also applied to political leaders, including the leaders of some of the Thirteen Colonies (originally Virginia in 1608); in full, the "President of the Council". The first republic to adopt the title was the United States of America. Keeping its usage as the head of a committee the President of the Continental Congress was the leader of the original congress. When the new constitution was written the title of President of the United States was conferred on the head of the new executive branch. If the head of state of a republic is also the head of government, this is called a presidential system. There are a number of forms of presidential government. A full-presidential system has a president with substantial authority and a central political role. In other states the legislature is dominant and the presidential role is almost purely ceremonial and apolitical, such as in Germany, Italy, India, and Trinidad and Tobago. These states are parliamentary republics and operate similarly to constitutional monarchies with parliamentary systems where the power of the monarch is also greatly circumscribed. In parliamentary systems the head of government, most often titled prime minister, exercises the most real political power. Semi-presidential systems have a president as an active head of state with important powers, but they also have a prime minister as a head of government with important powers. The rules for appointing the president and the leader of the government, in some republics permit the appointment of a president and a prime minister who have opposing political convictions: in France, when the members of the ruling cabinet and the president come from opposing political factions, this situation is called cohabitation. In some countries, like Bosnia and Herzegovina, San Marino, and Switzerland, the head of state is not a single person but a committee (council) of several persons holding that office. The Roman Republic had two consuls, elected for a one-year term by the comitia centuriata, consisting of all adult, freeborn males who could prove citizenship. In liberal democracies, presidents are elected, either directly by the people or indirectly by a parliament or council. Typically in presidential and semi-presidential systems the president is directly elected by the people, or is indirectly elected as done in the United States. In that country the president is officially elected by an electoral college, chosen by the States. All U.S. States have chosen electors by popular election since 1832. The indirect election of the president through the electoral college conforms to the concept of republic as one with a system of indirect election. In the opinion of some, direct election confers legitimacy upon the president and gives the office much of its political power. However, this concept of legitimacy differs from that expressed in the United States Constitution which established the legitimacy of the United States president as resulting from the signing of the Constitution by nine states. The idea that direct election is required for legitimacy also contradicts the spirit of the Great Compromise, whose actual result was manifest in the clause that provides voters in smaller states with more representation in presidential selection than those in large states; for example citizens of Wyoming in 2016 had 3.6 times as much electoral vote representation as citizens of California. In states with a parliamentary system the president is usually elected by the parliament. This indirect election subordinates the president to the parliament, and also gives the president limited legitimacy and turns most presidential powers into reserve powers that can only be exercised under rare circumstance. There are exceptions where elected presidents have only ceremonial powers, such as in Ireland. The distinction between a republic and a monarchy is not always clear. The constitutional monarchies of the former British Empire and Western Europe today have almost all real political power vested in the elected representatives, with the monarchs only holding either theoretical powers, no powers or rarely used reserve powers. Real legitimacy for political decisions comes from the elected representatives and is derived from the will of the people. While hereditary monarchies remain in place, political power is derived from the people as in a republic. These states are thus sometimes referred to as crowned republics. Terms such as "liberal republic" are also used to describe all of the modern liberal democracies. There are also self-proclaimed republics that act similarly to absolute monarchies with absolute power vested in the leader and passed down from father to son. North Korea and Syria are two notable examples where a son has inherited political control. Neither of these states are officially monarchies. There is no constitutional requirement that power be passed down within one family, but it has occurred in practice. There are also elective monarchies where ultimate power is vested in a monarch, but the monarch is chosen by some manner of election. A current example of such a state is Malaysia where the Yang di-Pertuan Agong is elected every five years by the Conference of Rulers composed of the nine hereditary rulers of the Malay states, and the Vatican City-State, where the pope is selected by cardinal-electors, currently all cardinals under the age of 80. While rare today, elective monarchs were common in the past. The Holy Roman Empire is an important example, where each new emperor was chosen by a group of electors. Islamic states also rarely employed primogeniture, instead relying on various forms of election to choose a monarch's successor. The Polish–Lithuanian Commonwealth had an elective monarchy, with a wide suffrage of some 500,000 nobles. The system, known as the Golden Liberty, had developed as a method for powerful landowners to control the crown. The proponents of this system looked to classical examples, and the writings of the Italian Renaissance, and called their elective monarchy a rzeczpospolita, based on res publica. In general being a republic also implies sovereignty as for the state to be ruled by the people it cannot be controlled by a foreign power. There are important exceptions to this, for example, republics in the Soviet Union were member states which had to meet three criteria to be named republics: It is sometimes argued that the former Soviet Union was also a supra-national republic, based on the claim that the member states were different nation states. The Socialist Federal Republic of Yugoslavia was a federal entity composed of six republics (Socialist Republic of Bosnia and Herzegovina, Croatia, Macedonia, Montenegro, Serbia, and Slovenia). Each republic had its parliament, government, institute of citizenship, constitution, etc., but certain functions were delegated to the federation (army, monetary matters). Each republic also had a right of self-determination according to the conclusions of the second session of the AVNOJ and according to the federal constitution. In Switzerland, all cantons can be considered to have a republican form of government, with constitutions, legislatures, executives and courts; many of them being originally sovereign states. As a consequence, several Romance-speaking cantons are still officially referred to as republics, reflecting their history and will of independence within the Swiss Confederation. Notable examples are the Republic and Canton of Geneva and the Republic and Canton of Ticino. States of the United States are required, like the federal government, to be republican in form, with final authority resting with the people. This was required because the states were intended to create and enforce most domestic laws, with the exception of areas delegated to the federal government and prohibited to the states. The founders of the country intended most domestic laws to be handled by the states. Requiring the states to be a republic in form was seen as protecting the citizens' rights and preventing a state from becoming a dictatorship or monarchy, and reflected unwillingness on the part of the original 13 states (all independent republics) to unite with other states that were not republics. Additionally, this requirement ensured that only other republics could join the union. In the example of the United States, the original 13 British colonies became independent states after the American Revolution, each having a republican form of government. These independent states initially formed a loose confederation called the United States and then later formed the current United States by ratifying the current U.S. Constitution, creating a union that was a republic. Any state joining the union later was also required to be a republic. Before the 17th Century, the term 'republic' could be used to refer to states of any form of government as long as it was not a tyrannical regime. French philosopher Jean Bodin's definition of the republic was "the rightly ordered government of a number of families, and of those things which are their common concern, by a sovereign power." Oligarchies and monarchies could also be included as they were also organised toward 'public' shared interests. In medieval texts, 'republic' was used to refer to the body of shared interest with the king at its head. For instance, the Holy Roman Empire was also known as the Sancta Respublica Romana, the Holy Roman Republic. The Byzantine Empire also continued calling itself the Roman Republic as the Byzantines did not regard monarchy as a contradiction to republicanism. Instead, republics were defined as any state based on popular sovereignty and whose institutions were based on shared values. While the term democracy has been used interchangeably with the term republic by some, others have made sharp distinctions between the two for millennia. "Montesquieu, founder of the modern constitutional state, repeated in his The Spirit of the Laws of 1748 the insight that Aristotle had expressed two millennia earlier, ‘Voting by lot is in the nature of democracy; voting by choice is in the nature of aristocracy.’" Additional critics of elections include Rousseau, Robespierre, and Marat, who said of the new French Republic, "What use is it to us, that we have broken the aristocracy of the nobles, if that is replaced by the aristocracy of the rich?" The term republic originated from the writers of the Renaissance as a descriptive term for states that were not monarchies. These writers, such as Machiavelli, also wrote important prescriptive works describing how such governments should function. These ideas of how a government and society should be structured is the basis for an ideology known as classical republicanism or civic humanism. This ideology is based on the Roman Republic and the city states of Ancient Greece and focuses on ideals such as civic virtue, rule of law and mixed government. This understanding of a republic as a form of government distinct from a liberal democracy is one of the main theses of the Cambridge School of historical analysis. This grew out of the work of J. G. A. Pocock who in 1975 argued that a series of scholars had expressed a consistent set of republican ideals. These writers included Machiavelli, Milton, Montesquieu and the founders of the United States of America. Pocock argued that this was an ideology with a history and principles distinct from liberalism. These ideas were embraced by a number of different writers, including Quentin Skinner, Philip Pettit and Cass Sunstein. These subsequent writers have further explored the history of the idea, and also outlined how a modern republic should function. A distinct set of definitions of the term "republic" evolved in the United States, where the term is often equated with "representative democracy." This narrower understanding of the term was originally developed by James Madison and notably employed in Federalist Paper No. 10. This meaning was widely adopted early in the history of the United States, including in Noah Webster's dictionary of 1828. It was a novel meaning to the term; representative democracy was not an idea mentioned by Machiavelli and did not exist in the classical republics. There is also evidence that contemporaries of Madison considered the meaning of "republic" to reflect the broader definition found elsewhere, as is the case with a quotation of Benjamin Franklin taken from the notes of James McHenry where the question is put forth, "a Republic or a Monarchy?". The term republic does not appear in the Declaration of Independence, but it does appear in Article IV of the Constitution, which "guarantee[s] to every State in this Union a Republican form of Government." What exactly the writers of the constitution felt this should mean is uncertain. The Supreme Court, in Luther v. Borden (1849), declared that the definition of republic was a "political question" in which it would not intervene. In two later cases, it did establish a basic definition. In United States v. Cruikshank (1875), the court ruled that the "equal rights of citizens" were inherent to the idea of a republic. However, the term republic is not synonymous with the republican form. The republican form is defined as one in which the powers of sovereignty are vested in the people and are exercised by the people, either directly, or through representatives chosen by the people, to whom those powers are specially delegated. Beyond these basic definitions, the word republic has a number of other connotations. W. Paul Adams observes that republic is most often used in the United States as a synonym for "state" or "government," but with more positive connotations than either of those terms. Republicanism is often referred to as the founding ideology of the United States. Traditionally scholars believed this American republicanism was a derivation of the classical liberal ideologies of John Locke and others developed in Europe. In the 1960s and 1970s, Bernard Bailyn began to argue that republicanism was just as, or even more important than liberalism in the creation of the United States. This issue is still much disputed and scholars like Isaac Kramnick completely reject this view.
{{redlinks|date=May 2020}} A '''manga''' ({{lang-ja|漫画}}) is a [[Japan]]ese [[comic book]]. Manga is drawn by a ''mangaka'' ({{lang-ja|漫画家}}) (Japanese for [[cartoonist]]: an [[artist]] of [[comics]]). Manga is usually read from right to left. The word manga can be both [[singular]] and [[plural]], and mean both the medium of comics or a single comic. It is a form of art that is used to draw comics and develop [[Anime]] ([[animated]] [[cartoon]]s of manga art). [[Colors]] and [[symbols]] are important. [[File:Wikipe-tan manga page1.jpg|thumb|Example of a manga]] == Types == * '''''Yōji''''' (幼児向け漫画) - manga for people aged 1–4 * '''''Kodomo''''' (or Jidō) manga (児童漫画、子ども漫画) - manga that appeals to many small children ** ''[[Doraemon]]'' ** ''[[Crayon Shin-chan|Kureyon Shinchan]] * '''''[[Shōnen manga|Shōnen]]''''' (少年漫画) - A boy is usually the main character in these types of manga. The storyline is mostly about adventure/fighting. **''Weekly Shōnen Jump'' (published by [[Shueisha]]) ***''[[Bleach (manga)|Bleach]]'' ***''[[Naruto (manga)|Naruto]]'' ***''[[black clover|Black Clover]]'' ***''[[One Piece]]'' ***''[[Dragon Ball]]'' ***''[[The Prince of Tennis]]'' ***''[[Gintama (manga)|Gintama]]'' ***''[[Death Note]]'' ** ''Weekly Shōnen Sunday'' (published by [[Shōgakukan]]) ***''[[InuYasha]]'' ***''[[Detective Conan]]'' ***''[[Major(manga)|Major]]'' ***''[[Magi(manga)|Magi]]'' ** ''Weekly Shōnen Magazine'' (published by Kōdansha) ***''[[Fairy Tail]]'' ***''[[Fighting Spirit (manga)]]'' ** ''[[Monthly GFantasy]]'' (published by [[Square Enix]]) ***''[[Black Butler]]'' * '''''[[Shōjo]]''''' ({{lang|ja|少女漫画)}} - Manga that features human emotions and relationships. Mostly for girls aged 13–17. ** ''[[Ciao (magazine)|Ciao]]'' (published by [[Shogakukan]]) ** ''[[Chuchu]]'' (published by Shogakukan) ** ''[[Sho-Comi]]'' (published by Shogakukan) ** ''[[Ribon]]'' (published by Shueisha) ** ''[[Nakayoshi]]'' (published by [[Kodansha]]) *** ''[[Sailor Moon]]'' - an example of ''shōjo'' manga published by ''Nakayoshi'' ** ''[[LaLa]]'' (published by [[Hakusensha]]) ** ''[[Hana to yume]]'' (published by Hakusensha) * '''''Seinen''''' - manga that is written for [[college]]-aged young men **''[[Young animal]]'' (published by Hakusensha) *** ''[[Berserk (manga)|Berserk]]'' *** ''[[Nana to Kaoru]]'' *** ''[[Futari Ecchi]]'' *** ''[[Ai Yori Aoshi]]'' ** ''[[Young King OURs]]'' (published by [[Shōnen Gahōsha]]) ***''[[Hellsing]]'' *** ** ''[[Young Ace]]'' (published by [[Kadokawa Shoten]]) ***''[[Another (novel)|Another]]'' * '''''Seijin''''' (成人漫画) - manga that might appeal to men more than women * '''''Josei/Redikomi''''' (女性漫画 レディースコミック、レディコミ) - ([[abbreviation]] of English "LADIes' COMIc") manga that might appeal to women more than men * '''''Shōnen-ai''''' (少年愛) - manga that features a romantic relationship between two male characters. Also known as Yaoi * '''''Shōjo-ai''''' (少女愛) - Manga that features a romantic relationship between two female characters. Another name Yuri(百合). ==Related pages== * [[Anime]] * [[Cosplay]] * [[Mangaka]] * [[Light novel]] * [[Otaku]] * [[Hentai]] * [[Ecchi]] * [https://mangapk.com/genres/manhwa Manhwa] {{Webarchive|url=https://web.archive.org/web/20210929032522/https://mangapk.com/genres/manhwa |date=2021-09-29 }} – [[South Korea|Korean]] comic books [[Category:Manga| ]]
Manga (漫画, IPA: [maŋga] ) are comics or graphic novels originating from Japan. Most manga conform to a style developed in Japan in the late 19th century, and the form has a long history in earlier Japanese art. The term manga is used in Japan to refer to both comics and cartooning. Outside of Japan, the word is typically used to refer to comics originally published in the country. In Japan, people of all ages and walks of life read manga. The medium includes works in a broad range of genres: action, adventure, business and commerce, comedy, detective, drama, historical, horror, mystery, romance, science fiction and fantasy, erotica (hentai and ecchi), sports and games, and suspense, among others. Many manga are translated into other languages. Since the 1950s, manga has become an increasingly major part of the Japanese publishing industry. By 1995, the manga market in Japan was valued at ¥586.4 billion ($6–7 billion), with annual sales of 1.9 billion manga books and manga magazines (also known as manga anthologies) in Japan (equivalent to 15 issues per person). In 2020 Japan's manga market value hit a new record of ¥612.6 billion due to the fast growth of digital manga sales as well as increase of print sales. In 2022 Japan's manga market hit yet another record value of ¥675.9 billion. Manga have also gained a significant worldwide readership. Beginning with the late 2010s manga started massively outselling American comics. As of 2021, the top four comics publishers in the world are manga publishers Shueisha, Kodansha, Kadokawa, and Shogakukan. In 2020 the North American manga market was valued at almost $250 million. According to NPD BookScan manga made up 76% of overall comics and graphic novel sales in the US in 2021. The fast growth of the North American manga market is attributed to manga's wide availability on digital reading apps, book retailer chains such as Barnes & Noble and online retailers such as Amazon as well as the increased streaming of anime. According to Jean-Marie Bouissou, manga represented 38% of the French comics market in 2005. This is equivalent to approximately three times that of the United States and was valued at about €460 million ($640 million). In Europe and the Middle East, the market was valued at $250 million in 2012. In April 2023, the Japan Business Federation laid out a proposal aiming to spur the economic growth of Japan by further promoting the contents industry abroad, primarily anime, manga and video games, for measures to invite industry experts from abroad to come to Japan to work, and to link with the tourism sector to help foreign fans of manga and anime visit sites across the country associated with particular manga stories. The federation seeks on quadrupling the sales of Japanese content in overseas markets within the upcoming 10 years. Manga stories are typically printed in black-and-white—due to time constraints, artistic reasons (as coloring could lessen the impact of the artwork) and to keep printing costs low—although some full-color manga exist (e.g., Colorful). In Japan, manga are usually serialized in large manga magazines, often containing many stories, each presented in a single episode to be continued in the next issue. A single manga story is almost always longer than a single issue from a Western comic. Collected chapters are usually republished in tankōbon volumes, frequently but not exclusively paperback books. A manga artist (mangaka in Japanese) typically works with a few assistants in a small studio and is associated with a creative editor from a commercial publishing company. If a manga series is popular enough, it may be animated after or during its run. Sometimes, manga are based on previous live-action or animated films. Manga-influenced comics, among original works, exist in other parts of the world, particularly in those places that speak Chinese ("manhua"), Korean ("manhwa"), English ("OEL manga"), and French ("manfra"), as well as in the nation of Algeria ("DZ-manga"). The word "manga" comes from the Japanese word 漫画 (katakana: マンガ; hiragana: まんが), composed of the two kanji 漫 (man) meaning "whimsical or impromptu" and 画 (ga) meaning "pictures". The same term is the root of the Korean word for comics, manhwa, and the Chinese word manhua. The word first came into common usage in the late 18th century with the publication of such works as Santō Kyōden's picturebook Shiji no yukikai (1798), and in the early 19th century with such works as Aikawa Minwa's Manga hyakujo (1814) and the celebrated Hokusai Manga books (1814–1834) containing assorted drawings from the sketchbooks of the famous ukiyo-e artist Hokusai. Rakuten Kitazawa (1876–1955) first used the word "manga" in the modern sense. In Japanese, "manga" refers to all kinds of cartooning, comics, and animation. Among English speakers, "manga" has the stricter meaning of "Japanese comics", in parallel to the usage of "anime" in and outside Japan. The term "ani-manga" is used to describe comics produced from animation cels. According to art resource Widewalls manga originated from emakimono (scrolls), Chōjū-jinbutsu-giga, dating back to the 12th century. During the Edo period (1603–1867), a book of drawings titled Toba Ehon further developed what would later be called manga. The word itself first came into common usage in 1798, with the publication of works such as Santō Kyōden's picturebook Shiji no yukikai (1798), and in the early 19th century with such works as Aikawa Minwa's Manga hyakujo (1814) and the Hokusai Manga books (1814–1834). Adam L. Kern has suggested that kibyoshi, picture books from the late 18th century, may have been the world's first comic books. These graphical narratives share with modern manga humorous, satirical, and romantic themes. Some works were mass-produced as serials using woodblock printing. however Eastern comics are generally held separate from the evolution of Western comics and Western comic art probably originated in 17th Italy, Writers on manga history have described two broad and complementary processes shaping modern manga. One view represented by other writers such as Frederik L. Schodt, Kinko Ito, and Adam L. Kern, stress continuity of Japanese cultural and aesthetic traditions, including pre-war, Meiji, and pre-Meiji culture and art. The other view, emphasizes events occurring during and after the Allied occupation of Japan (1945–1952), and stresses U.S. cultural influences, including U.S. comics (brought to Japan by the GIs) and images and themes from U.S. television, film, and cartoons (especially Disney). Regardless of its source, an explosion of artistic creativity occurred in the post-war period, involving manga artists such as Osamu Tezuka (Astro Boy) and Machiko Hasegawa (Sazae-san). Astro Boy quickly became (and remains) immensely popular in Japan and elsewhere, and the anime adaptation of Sazae-san drew more viewers than any other anime on Japanese television in 2011. Tezuka and Hasegawa both made stylistic innovations. In Tezuka's "cinematographic" technique, the panels are like a motion picture that reveals details of action bordering on slow motion as well as rapid zooms from distance to close-up shots. This kind of visual dynamism was widely adopted by later manga artists. Hasegawa's focus on daily life and on women's experience also came to characterize later shōjo manga. Between 1950 and 1969, an increasingly large readership for manga emerged in Japan with the solidification of its two main marketing genres, shōnen manga aimed at boys and shōjo manga aimed at girls. In 1969, a group of female manga artists (later called the Year 24 Group, also known as Magnificent 24s) made their shōjo manga debut ("year 24" comes from the Japanese name for the year 1949, the birth-year of many of these artists). The group included Moto Hagio, Riyoko Ikeda, Yumiko Ōshima, Keiko Takemiya, and Ryoko Yamagishi. Thereafter, primarily female manga artists would draw shōjo for a readership of girls and young women. In the following decades (1975–present), shōjo manga continued to develop stylistically while simultaneously evolving different but overlapping subgenres. Major subgenres include romance, superheroines, and "Ladies Comics" (in Japanese, redisu レディース, redikomi レディコミ, and josei 女性). Modern shōjo manga romance features love as a major theme set into emotionally intense narratives of self-realization. With the superheroines, shōjo manga saw releases such as Pink Hanamori's Mermaid Melody Pichi Pichi Pitch, Reiko Yoshida's Tokyo Mew Mew, and Naoko Takeuchi's Pretty Soldier Sailor Moon, which became internationally popular in both manga and anime formats. Groups (or sentais) of girls working together have also been popular within this genre. Like Lucia, Hanon, and Rina singing together, and Sailor Moon, Sailor Mercury, Sailor Mars, Sailor Jupiter, and Sailor Venus working together. Manga for male readers sub-divides according to the age of its intended readership: boys up to 18 years old (shōnen manga) and young men 18 to 30 years old (seinen manga); as well as by content, including action-adventure often involving male heroes, slapstick humor, themes of honor, and sometimes explicit sex. The Japanese use different kanji for two closely allied meanings of "seinen"—青年 for "youth, young man" and 成年 for "adult, majority"—the second referring to pornographic manga aimed at grown men and also called seijin ("adult" 成人) manga. Shōnen, seinen, and seijin manga share a number of features in common. Boys and young men became some of the earliest readers of manga after World War II. From the 1950s on, shōnen manga focused on topics thought to interest the archetypal boy, including subjects like robots, space-travel, and heroic action-adventure. Popular themes include science fiction, technology, sports, and supernatural settings. Manga with solitary costumed superheroes like Superman, Batman, and Spider-Man generally did not become as popular. The role of girls and women in manga produced for male readers has evolved considerably over time to include those featuring single pretty girls (bishōjo) such as Belldandy from Oh My Goddess!, stories where such girls and women surround the hero, as in Negima and Hanaukyo Maid Team, or groups of heavily armed female warriors (sentō bishōjo) By the turn of the 21st century, manga "achieved worldwide popularity". With the relaxation of censorship in Japan in the 1990s, an assortment of explicit sexual material appeared in manga intended for male readers, and correspondingly continued into the English translations. In 2010, the Tokyo Metropolitan Government considered a bill to restrict minors' access to such content. The gekiga style of storytelling—thematically somber, adult-oriented, and sometimes deeply violent—focuses on the day-in, day-out grim realities of life, often drawn in a gritty and unvarnished fashion. Gekiga such as Sampei Shirato's 1959–1962 Chronicles of a Ninja's Military Accomplishments (Ninja Bugeichō) arose in the late 1950s and 1960s, partly from left-wing student and working-class political activism, and partly from the aesthetic dissatisfaction of young manga artists like Yoshihiro Tatsumi with existing manga. In Japan, manga constituted an annual 40.6 billion yen (approximately US$395 million) publication-industry by 2007. In 2006 sales of manga books made up for about 27% of total book-sales, and sale of manga magazines, for 20% of total magazine-sales. The manga industry has expanded worldwide, where distribution companies license and reprint manga into their native languages. Marketeers primarily classify manga by the age and gender of the target readership. In particular, books and magazines sold to boys (shōnen) and girls (shōjo) have distinctive cover-art, and most bookstores place them on different shelves. Due to cross-readership, consumer response is not limited by demographics. For example, male readers may subscribe to a series intended for female readers, and so on. Japan has manga cafés, or manga kissa (kissa is an abbreviation of kissaten). At a manga kissa, people drink coffee, read manga and sometimes stay overnight. The Kyoto International Manga Museum maintains a very large website listing manga published in Japanese. Manga magazines or anthologies (漫画雑誌, manga zasshi) usually have many series running concurrently with approximately 20–40 pages allocated to each series per issue. Other magazines such as the anime fandom magazine Newtype featured single chapters within their monthly periodicals. Other magazines like Nakayoshi feature many stories written by many different artists; these magazines, or "anthology magazines", as they are also known (colloquially "phone books"), are usually printed on low-quality newsprint and can be anywhere from 200 to more than 850 pages thick. Manga magazines also contain one-shot comics and various four-panel yonkoma (equivalent to comic strips). Manga series can run for many years if they are successful. Popular shonen magazines include Weekly Shōnen Jump, Weekly Shōnen Magazine and Weekly Shōnen Sunday - Popular shoujo manga include Ciao, Nakayoshi and Ribon. Manga artists sometimes start out with a few "one-shot" manga projects just to try to get their name out. If these are successful and receive good reviews, they are continued. Magazines often have a short life. After a series has run for a while, publishers often collect the chapters and print them in dedicated book-sized volumes, called tankōbon. These can be hardcover, or more usually softcover books, and are the equivalent of U.S. trade paperbacks or graphic novels. These volumes often use higher-quality paper, and are useful to those who want to "catch up" with a series so they can follow it in the magazines or if they find the cost of the weeklies or monthlies to be prohibitive. "Deluxe" versions have also been printed as readers have gotten older and the need for something special grew. Old manga have also been reprinted using somewhat lesser quality paper and sold for 100 yen (about $1 U.S. dollar) each to compete with the used book market. Kanagaki Robun and Kawanabe Kyōsai created the first manga magazine in 1874: Eshinbun Nipponchi. The magazine was heavily influenced by Japan Punch, founded in 1862 by Charles Wirgman, a British cartoonist. Eshinbun Nipponchi had a very simple style of drawings and did not become popular with many people. Eshinbun Nipponchi ended after three issues. The magazine Kisho Shimbun in 1875 was inspired by Eshinbun Nipponchi, which was followed by Marumaru Chinbun in 1877, and then Garakuta Chinpo in 1879. Shōnen Sekai was the first shōnen magazine created in 1895 by Iwaya Sazanami, a famous writer of Japanese children's literature back then. Shōnen Sekai had a strong focus on the First Sino-Japanese War. In 1905, the manga-magazine publishing boom started with the Russo-Japanese War, Tokyo Pakku was created and became a huge hit. After Tokyo Pakku in 1905, a female version of Shōnen Sekai was created and named Shōjo Sekai, considered the first shōjo magazine. Shōnen Pakku was made and is considered the first children's manga magazine. The children's demographic was in an early stage of development in the Meiji period. Shōnen Pakku was influenced from foreign children's magazines such as Puck which an employee of Jitsugyō no Nihon (publisher of the magazine) saw and decided to emulate. In 1924, Kodomo Pakku was launched as another children's manga magazine after Shōnen Pakku. During the boom, Poten (derived from the French "potin") was published in 1908. All the pages were in full color with influences from Tokyo Pakku and Osaka Puck. It is unknown if there were any more issues besides the first one. Kodomo Pakku was launched May 1924 by Tokyosha and featured high-quality art by many members of the manga artistry like Takei Takeo, Takehisa Yumeji and Aso Yutaka. Some of the manga featured speech balloons, where other manga from the previous eras did not use speech balloons and were silent. Published from May 1935 to January 1941, Manga no Kuni coincided with the period of the Second Sino-Japanese War (1937–1945). Manga no Kuni featured information on becoming a mangaka and on other comics industries around the world. Manga no Kuni handed its title to Sashie Manga Kenkyū in August 1940. Dōjinshi, produced by small publishers outside of the mainstream commercial market, resemble in their publishing small-press independently published comic books in the United States. Comiket, the largest comic book convention in the world with around 500,000 visitors gathering over three days, is devoted to dōjinshi. While they most often contain original stories, many are parodies of or include characters from popular manga and anime series. Some dōjinshi continue with a series' story or write an entirely new one using its characters, much like fan fiction. In 2007, dōjinshi sales amounted to 27.73 billion yen (US$245 million). In 2006 they represented about a tenth of manga books and magazines sales. Thanks to the advent of the internet, there have been new ways for aspiring mangaka to upload and sell their manga online. Before, there were two main ways in which a mangaka's work could be published: taking their manga drawn on paper to a publisher themselves, or submitting their work to competitions run by magazines. In recent years, there has been a rise in manga released digitally. Web manga, as it is known in Japan, has seen an increase thanks in part to image hosting websites where anyone can upload pages from their works for free. Although released digitally, almost all web manga sticks to the conventional black-and-white format despite some never getting physical publication. Pixiv is the most popular site where amateur and professional work gets published on the site. It has grown to be the most visited site for artwork in Japan. Twitter has also become a popular place for web manga with many artists releasing pages weekly on their accounts in the hope of their work getting picked up or published professionally. One of the best examples of an amateur work becoming professional is One-Punch Man which was released online and later received a professional remake released digitally and an anime adaptation soon thereafter. Many of the big print publishers have also released digital only magazines and websites where web manga get published alongside their serialized magazines. Shogakukan for instance has two websites, Sunday Webry and Ura Sunday, that release weekly chapters for web manga and even offer contests for mangaka to submit their work. Both Sunday Webry and Ura Sunday have become one of the top web manga sites in Japan. Some have even released apps that teach how to draw professional manga and learn how to create them. Weekly Shōnen Jump released Jump Paint, an app that guides users on how to make their own manga from making storyboards to digitally inking lines. It also offers more than 120 types of pen tips and more than 1,000 screentones for artists to practice. Kodansha has also used the popularity of web manga to launch more series and also offer better distribution of their officially translated works under Kodansha Comics thanks in part to the titles being released digitally first before being published physically. The rise web manga has also been credited to smartphones and computers as more and more readers read manga on their phones rather than from a print publication. While paper manga has seen a decrease over time, digital manga have been growing in sales each year. The Research Institute for Publications reports that sales of digital manga books excluding magazines jumped 27.1 percent to ¥146 billion in 2016 from the year before while sales of paper manga saw a record year-on-year decline of 7.4 percent to ¥194.7 billion. They have also said that if the digital and paper keep the same growth and drop rates, web manga would exceed their paper counterparts. In 2020 manga sales topped the ¥600 billion mark for the first time in history, beating the 1995 peak due to a fast growth of the digital manga market which rose by ¥82.7 billion from a previous year, surpassing print manga sales which have also increased. While webtoons have caught on in popularity as a new medium for comics in Asia, Japan has been slow to adopt webtoons as the traditional format and print publication still dominate the way manga is created and consumed(although this is beginning to change). Despite this, one of the biggest webtoon publishers in the world, Comico, has had success in the traditional Japanese manga market. Comico was launched by NHN Japan, the Japanese subsidiary of Korean company, NHN Entertainment. As of now, there are only two webtoon publishers that publish Japanese webtoons: Comico and Naver Webtoon (under the name XOY in Japan). Kakao has also had success by offering licensed manga and translated Korean webtoons with their service Piccoma. All three companies credit their success to the webtoon pay model where users can purchase each chapter individually instead of having to buy the whole book while also offering some chapters for free for a period of time allowing anyone to read a whole series for free if they wait long enough. The added benefit of having all of their titles in color and some with special animations and effects have also helped them succeed. Some popular Japanese webtoons have also gotten anime adaptations and print releases, the most notable being ReLIFE and Recovery of an MMO Junkie. By 2007, the influence of manga on international comics had grown considerably over the past two decades. "Influence" is used here to refer to effects on the comics markets outside Japan and to aesthetic effects on comics artists internationally. Traditionally, manga stories flow from top to bottom and from right to left. Some publishers of translated manga keep to this original format. Other publishers mirror the pages horizontally before printing the translation, changing the reading direction to a more "Western" left to right, so as not to confuse foreign readers or traditional comics-consumers. This practice is known as "flipping". For the most part, criticism suggests that flipping goes against the original intentions of the creator (for example, if a person wears a shirt that reads "MAY" on it, and gets flipped, then the word is altered to "YAM"), who may be ignorant of how awkward it is to read comics when the eyes must flow through the pages and text in opposite directions, resulting in an experience that's quite distinct from reading something that flows homogeneously. If the translation is not adapted to the flipped artwork carefully enough it is also possible for the text to go against the picture, such as a person referring to something on their left in the text while pointing to their right in the graphic. Characters shown writing with their right hands, the majority of them, would become left-handed when a series is flipped. Flipping may also cause oddities with familiar asymmetrical objects or layouts, such as a car being depicted with the gas pedal on the left and the brake on the right, or a shirt with the buttons on the wrong side, however these issues are minor when compared to the unnatural reading flow, and some of them could be solved with an adaptation work that goes beyond just translation and blind flipping. Manga has highly influenced the art styles of manhwa and manhua. Manga in Indonesia is published by Elex Media Komputindo, Level Comic, M&C and Gramedia. Manga has influenced Indonesia's original comic industry. Manga in the Philippines were imported from the US and were sold only in specialty stores and in limited copies. The first manga in Filipino language is Doraemon which was published by J-Line Comics and was then followed by Case Closed. In 2015, Boys' Love manga became popular through the introduction of BL manga by printing company BLACKink. Among the first BL titles to be printed were Poster Boy, Tagila, and Sprinters, all were written in Filipino. BL manga have become bestsellers in the top three bookstore companies in the Philippines since their introduction in 2015. During the same year, Boys' Love manga have become a popular mainstream with Thai consumers, leading to television series adapted from BL manga stories since 2016. Manga has influenced European cartooning in a way that is somewhat different from in the U.S. Broadcast anime in France and Italy opened the European market to manga during the 1970s. French art has borrowed from Japan since the 19th century (Japonism) and has its own highly developed tradition of bande dessinée cartooning. Manga was introduced to France in the late 1990s, where Japanese pop culture became massively popular: in 2021, 55% of comics sold in the country were manga and France is the biggest manga importer. By mid-2021, 75 percent of the €300 value of Culture Pass [fr] accounts given to French 18 year-olds was spent on manga. According to the Japan External Trade Organization, sales of manga reached $212.6 million within France and Germany alone in 2006. France represents about 50% of the European market and is the second worldwide market, behind Japan. In 2013, there were 41 publishers of manga in France and, together with other Asian comics, manga represented around 40% of new comics releases in the country, surpassing Franco-Belgian comics for the first time. European publishers marketing manga translated into French include Asuka, Casterman, Glénat, Kana, and Pika Édition, among others. European publishers also translate manga into Dutch, German, Italian, and other languages. In 2007, about 70% of all comics sold in Germany were manga. Since 2010 the country celebrates Manga Day on every August 27. In 2021 manga sales in Germany rose by 75% from its original record of 70 million in 2005. As of 2022 Germany is the third largest manga market in Europe after Italy and France. In 2021, the Spanish manga market hit a record of 1033 new title publications. In 2022 the 28th edition of the Barcelona Manga Festival opened its doors to more than 163,000 fans, compared to a pre-pandemic 120,000 in 2019. Manga publishers based in the United Kingdom include Gollancz and Titan Books. Manga publishers from the United States have a strong marketing presence in the United Kingdom: for example, the Tanoshimi line from Random House. In 2019 The British Museum held a mass exhibition dedicated to manga. Manga made their way only gradually into U.S. markets, first in association with anime and then independently. Some U.S. fans became aware of manga in the 1970s and early 1980s. However, anime was initially more accessible than manga to U.S. fans, many of whom were college-age young people who found it easier to obtain, subtitle, and exhibit video tapes of anime than translate, reproduce, and distribute tankōbon-style manga books. One of the first manga translated into English and marketed in the U.S. was Keiji Nakazawa's Barefoot Gen, an autobiographical story of the atomic bombing of Hiroshima issued by Leonard Rifas and Educomics (1980–1982). More manga were translated between the mid-1980s and 1990s, including Golgo 13 in 1986, Lone Wolf and Cub from First Comics in 1987, and Kamui, Area 88, and Mai the Psychic Girl, also in 1987 and all from Viz Media-Eclipse Comics. Others soon followed, including Akira from Marvel Comics' Epic Comics imprint, Nausicaä of the Valley of the Wind from Viz Media, and Appleseed from Eclipse Comics in 1988, and later Iczer-1 (Antarctic Press, 1994) and Ippongi Bang's F-111 Bandit (Antarctic Press, 1995). During the 1980s and 1990s, Japanese animation, such as Akira, Dragon Ball, Neon Genesis Evangelion, and Pokémon, made a larger impact on the fan experience and in the market than manga. Matters changed when translator-entrepreneur Toren Smith founded Studio Proteus in 1986. Smith and Studio Proteus acted as an agent and translator of many Japanese manga, including Masamune Shirow's Appleseed and Kōsuke Fujishima's Oh My Goddess!, for Dark Horse and Eros Comix, eliminating the need for these publishers to seek their own contacts in Japan. Simultaneously, the Japanese publisher Shogakukan opened a U.S. market initiative with their U.S. subsidiary Viz, enabling Viz to draw directly on Shogakukan's catalogue and translation skills. Japanese publishers began pursuing a U.S. market in the mid-1990s, due to a stagnation in the domestic market for manga. The U.S. manga market took an upturn with mid-1990s anime and manga versions of Masamune Shirow's Ghost in the Shell (translated by Frederik L. Schodt and Toren Smith) becoming very popular among fans. An extremely successful manga and anime translated and dubbed in English in the mid-1990s was Sailor Moon. By 1995–1998, the Sailor Moon manga had been exported to over 23 countries, including China, Brazil, Mexico, Australia, North America and most of Europe. In 1997, Mixx Entertainment began publishing Sailor Moon, along with CLAMP's Magic Knight Rayearth, Hitoshi Iwaaki's Parasyte and Tsutomu Takahashi's Ice Blade in the monthly manga magazine MixxZine. Mixx Entertainment, later renamed Tokyopop, also published manga in trade paperbacks and, like Viz, began aggressive marketing of manga to both young male and young female demographics. During this period, Dark Horse Manga was a major publisher of translated manga. In addition to Oh My Goddess!, the company published Akira, Astro Boy, Berserk, Blade of the Immortal, Ghost in the Shell, Lone Wolf and Cub, Yasuhiro Nightow's Trigun and Blood Blockade Battlefront, Gantz, Kouta Hirano's Hellsing and Drifters, Blood+, Multiple Personality Detective Psycho, FLCL, Mob Psycho 100, and Oreimo. The company received 13 Eisner Award nominations for its manga titles, and three of the four manga creators admitted to The Will Eisner Award Hall of Fame — Osamu Tezuka, Kazuo Koike, and Goseki Kojima — were published in Dark Horse translations. In the following years, manga became increasingly popular, and new publishers entered the field while the established publishers greatly expanded their catalogues. The Pokémon manga Electric Tale of Pikachu issue #1 sold over 1 million copies in the United States, making it the best-selling single comic book in the United States since 1993. By 2008, the U.S. and Canadian manga market generated $175 million in annual sales. Simultaneously, mainstream U.S. media began to discuss manga, with articles in The New York Times, Time magazine, The Wall Street Journal, and Wired magazine. As of 2017, manga distributor Viz Media is the largest publisher of graphic novels and comic books in the United States, with a 23% share of the market. BookScan sales show that manga is one of the fastest-growing areas of the comic book and narrative fiction markets. From January 2019 to May 2019, the manga market grew 16%, compared to the overall comic book market's 5% growth. The NPD Group noted that, compared to other comic book readers, manga readers are younger (76% under 30) and more diverse, including a higher female readership (16% higher than other comic books). As of January 2020, manga is the second largest category in the US comic book and graphic novel market, accounting for 27% of the entire market share. During the COVID-19 pandemic some stores of the American bookseller Barnes & Noble saw up to a 500% increase in sales from graphic novel and manga sales due to the younger generations showing a high interest in the medium. Sales of print manga titles in the U.S. increased by 3.6 million units in the first quarter of 2021 compared to the same period in 2020. In 2021, 24.4 million units of manga were sold in the United States. This is an increase of about 15 million (160%) more sales than in 2020. In 2022, most of the top-selling comic creators in the United States were mangaka. The same year manga sales saw an increase of 9%. A number of artists in the United States have drawn comics and cartoons influenced by manga. As an early example, Vernon Grant drew manga-influenced comics while living in Japan in the late 1960s and early 1970s. Others include Frank Miller's mid-1980s Ronin, Adam Warren and Toren Smith's 1988 The Dirty Pair, Ben Dunn's 1987 Ninja High School and Manga Shi 2000 from Crusade Comics (1997). By the beginning of the 21st century, several U.S. manga publishers had begun to produce work by U.S. artists under the broad marketing-label of manga. In 2002, I.C. Entertainment, formerly Studio Ironcat and now out of business, launched a series of manga by U.S. artists called Amerimanga. In 2004, eigoMANGA launched the Rumble Pak and Sakura Pakk anthology series. Seven Seas Entertainment followed suit with World Manga. Simultaneously, TokyoPop introduced original English-language manga (OEL manga) later renamed Global Manga. Francophone artists have also developed their own versions of manga (manfra), like Frédéric Boilet's la nouvelle manga. Boilet has worked in France and in Japan, sometimes collaborating with Japanese artists. The Japanese manga industry grants a large number of awards, mostly sponsored by publishers, with the winning prize usually including publication of the winning stories in magazines released by the sponsoring publisher. Examples of these awards include: The Japanese Ministry of Foreign Affairs has awarded the International Manga Award annually since May 2007. Kyoto Seika University in Japan has offered a highly competitive course in manga since 2000. Then, several established universities and vocational schools (専門学校: Semmon gakkou) established a training curriculum. Shuho Sato, who wrote Umizaru and Say Hello to Black Jack, has created some controversy on Twitter. Sato says, "Manga school is meaningless because those schools have very low success rates. Then, I could teach novices required skills on the job in three months. Meanwhile, those school students spend several million yen, and four years, yet they are good for nothing." and that, "For instance, Keiko Takemiya, the then professor of Seika Univ., remarked in the Government Council that 'A complete novice will be able to understand where is "Tachikiri" (i.e., margin section) during four years.' On the other hand, I would imagine that, It takes about thirty minutes to completely understand that at work."
'''Zeno's paradoxes''' are a famous set of [[thought experiment|thought-provoking]] stories or puzzles created by [[Zeno of Elea]] in the mid-5th century BC. [[Philosopher]]s, [[physicist]]s, and [[mathematician]]s have argued for 25 centuries over how to answer the questions raised by Zeno's paradoxes. Nine paradoxes have been attributed to him. Zeno constructed them to answer those who thought that [[Parmenides]]'s idea that "all is one and unchanging" was absurd. Three of Zeno's paradoxes are the most famous: two are presented below. They all deal with problems of the apparently continuous nature of space and time. ==Achilles and the tortoise== In the [[paradox]] of [[Achilles]] and the [[Tortoise]], Achilles is in a footrace with the tortoise. Achilles allows the tortoise a [[Head start (positioning)| head start]] of 100 [[Metre|metres]], for example. Suppose that each racer starts running at a constant speed, one very fast and one very slow. After some [[finite]] time, Achilles will have run 100 metres, bringing him to the tortoise's starting point. During this time, the slower tortoise has run a much shorter distance. It will then take Achilles some further time to run that distance, by which time the tortoise will have advanced farther. It will then take still more time for Achilles to reach this third point, while the tortoise again moves ahead. Thus, whenever Achilles reaches somewhere the tortoise has been, he still has farther to go. Therefore, because there are an [[infinite]] number of points Achilles must reach where the tortoise has already been, he can never overtake the tortoise.<ref>{{cite web |url=http://mathforum.org/isaac/problems/zeno1.html |title=Math Forum |accessdate=2012-09-12}}, matchforum.org </ref><ref> {{Cite encyclopedia |last=Huggett |first=Nick|url=http://plato.stanford.edu/entries/paradox-zeno/#AchTor |title=Zeno's Paradoxes: 3.2 Achilles and the Tortoise|year=2010 |encyclopedia=Stanford Encyclopedia of Philosophy|accessdate=2012-09-12 }}</ref> == The dichotomy paradox == Suppose someone wishes to get from point ''A'' to point ''B''. First, they must move halfway. Then, they must go half of the remaining way. Continuing in this manner, there will always be some small distance remaining, and the goal would never actually be reached. There will always be another number to add in a series such as 1 + 1/2 + 1/4 + 1/8 + 1/16 + .... So, motion from any point ''A'' to any different point ''B'' seems an impossibility. Going half the distance to a destination means you will not ever reach it, since you will have to go halfway from wherever you are at. == Commentary == This then is where Zeno's paradox lies: both pictures of reality cannot be true at the same time. Hence, either: 1. There is something wrong with the way we perceive the continuous nature of time, 2. In reality there is no such thing as a discrete, or incremental, amounts of time, distance, or perhaps anything else for that matter, or 3. There is a third picture of reality that unifies the two pictures--the mathematical one and the common sense or philosophical one--that we do not yet have the tools to fully understand. ==Proposed solutions== Few people would bet that the tortoise would win the race against an athlete. But, what is wrong with the argument? As one begins adding the terms in the series 1/2 + 1/4 + 1/8 + 1/16 + 1/32 + 1/64 + ...., one may notice that the sum gets closer and closer to 1, and will never exceed 1. [[Aristotle]] (who is the source for much of what we know about Zeno) noted that as the distance (in the dichotomy paradox) decreases, the time to travel each distance gets exceedingly smaller and smaller. Before 212 BC, [[Archimedes]] had developed a method to get a finite answer for the sum of infinitely many terms which get progressively smaller (such as 1/2 + 1/4 + 1/8 + 1/16 + 1/32 + ...). Modern calculus achieves the same result, using more rigorous methods.<ref name=boyer /><ref>Thomas, George B. 1951. ''Calculus and analytic geometry'' Addison Wesley.</ref> Some mathematicians, such as [[Carl Boyer]], think that Zeno's paradoxes are just mathematical problems, for which modern [[calculus]] provides a mathematical solution.<ref name=boyer>{{cite book |last=Boyer |first=Carl |title=The history of the calculus and its conceptual development |url=https://books.google.com/books?id=w3xKLt_da2UC&q=zeno |year=1959 |publisher=Dover Publications |page=295 | quote=If the paradoxes are thus stated in the precise mathematical terminology of continuous variables (...) the seeming contradictions resolve themselves. |isbn=978-0-486-60509-8}}</ref> However, Zeno's questions remain problematic if one approaches an infinite series of steps, one step at a time. This is known as a 'supertask'. Calculus does not actually involve adding numbers one at a time. Instead, it determines the value (called a ''limit'') that the addition is approaching. == References == {{Reflist}} {{paradoxes}} [[Category:Mathematical paradoxes]] [[Category:Philosophical problems]] [[Category:Logical paradoxes]]
Zeno's paradoxes are a set of philosophical problems devised by the Eleatic Greek philosopher Zeno of Elea (c. 490–430 BC). The origins of the paradoxes are somewhat unclear, but they are generally thought to have been developed to support Parmenides' doctrine of monism, that all of reality is one, and that all change is impossible. Diogenes Laërtius, citing Favorinus, says that Zeno's teacher Parmenides was the first to introduce the paradox of Achilles and the tortoise. But in a later passage, Laërtius attributes the origin of the paradox to Zeno, explaining that Favorinus disagrees. Many of these paradoxes argue that contrary to the evidence of one's senses, motion is nothing but an illusion. In Plato's Parmenides (128a–d), Zeno is characterized as taking on the project of creating these paradoxes because other philosophers claimed paradoxes arise when considering Parmenides' view. Zeno's arguments may then be early examples of a method of proof called reductio ad absurdum, also known as proof by contradiction. Thus Plato has Zeno say the purpose of the paradoxes "is to show that their hypothesis that existences are many, if properly followed up, leads to still more absurd results than the hypothesis that they are one." Plato has Socrates claim that Zeno and Parmenides were essentially arguing exactly the same point. They are also credited as a source of the dialectic method used by Socrates. Some of Zeno's nine surviving paradoxes (preserved in Aristotle's Physics and Simplicius's commentary thereon) are essentially equivalent to one another. Aristotle offered a response to some of them. Popular literature often misrepresents Zeno's arguments. For example, Zeno is often said to have argued that the sum of an infinite number of terms must itself be infinite–with the result that not only the time, but also the distance to be travelled, become infinite. However, none of the original ancient sources has Zeno discussing the sum of any infinite series. Simplicius has Zeno saying "it is impossible to traverse an infinite number of things in a finite time". This presents Zeno's problem not with finding the sum, but rather with finishing a task with an infinite number of steps: how can one ever get from A to B, if an infinite number of (non-instantaneous) events can be identified that need to precede the arrival at B, and one cannot reach even the beginning of a "last event"? Three of the strongest and most famous—that of Achilles and the tortoise, the Dichotomy argument, and that of an arrow in flight—are presented in detail below. That which is in locomotion must arrive at the half-way stage before it arrives at the goal. Suppose Atalanta wishes to walk to the end of a path. Before she can get there, she must get halfway there. Before she can get halfway there, she must get a quarter of the way there. Before traveling a quarter, she must travel one-eighth; before an eighth, one-sixteenth; and so on. The resulting sequence can be represented as: This description requires one to complete an infinite number of tasks, which Zeno maintains is an impossibility. This sequence also presents a second problem in that it contains no first distance to run, for any possible (finite) first distance could be divided in half, and hence would not be first after all. Hence, the trip cannot even begin. The paradoxical conclusion then would be that travel over any finite distance can be neither completed nor begun, and so all motion must be an illusion. This argument is called the "Dichotomy" because it involves repeatedly splitting a distance into two parts. An example with the original sense can be found in an asymptote. It is also known as the Race Course paradox. In a race, the quickest runner can never overtake the slowest, since the pursuer must first reach the point whence the pursued started, so that the slower must always hold a lead. In the paradox of Achilles and the tortoise, Achilles is in a footrace with the tortoise. Achilles allows the tortoise a head start of 100 meters, for example. Suppose that each racer starts running at some constant speed, one faster than the other. After some finite time, Achilles will have run 100 meters, bringing him to the tortoise's starting point. During this time, the tortoise has run a much shorter distance, say 2 meters. It will then take Achilles some further time to run that distance, by which time the tortoise will have advanced farther; and then more time still to reach this third point, while the tortoise moves ahead. Thus, whenever Achilles arrives somewhere the tortoise has been, he still has some distance to go before he can even reach the tortoise. As Aristotle noted, this argument is similar to the Dichotomy. It lacks, however, the apparent conclusion of motionlessness. If everything when it occupies an equal space is at rest at that instant of time, and if that which is in locomotion is always occupying such a space at any moment, the flying arrow is therefore motionless at that instant of time and at the next instant of time but if both instants of time are taken as the same instant or continuous instant of time then it is in motion. In the arrow paradox, Zeno states that for motion to occur, an object must change the position which it occupies. He gives an example of an arrow in flight. He states that at any one (durationless) instant of time, the arrow is neither moving to where it is, nor to where it is not. It cannot move to where it is not, because no time elapses for it to move there; it cannot move to where it is, because it is already there. In other words, at every instant of time there is no motion occurring. If everything is motionless at every instant, and time is entirely composed of instants, then motion is impossible. Whereas the first two paradoxes divide space, this paradox starts by dividing time—and not into segments, but into points. Aristotle gives three other paradoxes. From Aristotle: If everything that exists has a place, place too will have a place, and so on ad infinitum. Description of the paradox from the Routledge Dictionary of Philosophy: The argument is that a single grain of millet makes no sound upon falling, but a thousand grains make a sound. Hence a thousand nothings become something, an absurd conclusion. Aristotle's response: Zeno's reasoning is false when he argues that there is no part of the millet that does not make a sound: for there is no reason why any such part should not in any length of time fail to move the air that the whole bushel moves in falling. In fact it does not of itself move even such a quantity of the air as it would move if this part were by itself: for no part even exists otherwise than potentially. Description from Nick Huggett: This is a Parmenidean argument that one cannot trust one's sense of hearing. Aristotle's response seems to be that even inaudible sounds can add to an audible sound. From Aristotle: ... concerning the two rows of bodies, each row being composed of an equal number of bodies of equal size, passing each other on a race-course as they proceed with equal velocity in opposite directions, the one row originally occupying the space between the goal and the middle point of the course and the other that between the middle point and the starting-post. This...involves the conclusion that half a given time is equal to double that time. For an expanded account of Zeno's arguments as presented by Aristotle, see Simplicius's commentary On Aristotle's Physics. According to Simplicius, Diogenes the Cynic said nothing upon hearing Zeno's arguments, but stood up and walked, in order to demonstrate the falsity of Zeno's conclusions (see solvitur ambulando). To fully solve any of the paradoxes, however, one needs to show what is wrong with the argument, not just the conclusions. Through history, several solutions have been proposed, among the earliest recorded being those of Aristotle and Archimedes. Aristotle (384 BC–322 BC) remarked that as the distance decreases, the time needed to cover those distances also decreases, so that the time needed also becomes increasingly small. Aristotle also distinguished "things infinite in respect of divisibility" (such as a unit of space that can be mentally divided into ever smaller units while remaining spatially the same) from things (or distances) that are infinite in extension ("with respect to their extremities"). Aristotle's objection to the arrow paradox was that "Time is not composed of indivisible nows any more than any other magnitude is composed of indivisibles." Thomas Aquinas, commenting on Aristotle's objection, wrote "Instants are not parts of time, for time is not made up of instants any more than a magnitude is made of points, as we have already proved. Hence it does not follow that a thing is not in motion in a given time, just because it is not in motion in any instant of that time." Some mathematicians and historians, such as Carl Boyer, hold that Zeno's paradoxes are simply mathematical problems, for which modern calculus provides a mathematical solution. Infinite processes remained theoretically troublesome in mathematics until the late 19th century. With the epsilon-delta definition of limit, Weierstrass and Cauchy developed a rigorous formulation of the logic and calculus involved. These works resolved the mathematics involving infinite processes. Some philosophers, however, say that Zeno's paradoxes and their variations (see Thomson's lamp) remain relevant metaphysical problems. While mathematics can calculate where and when the moving Achilles will overtake the Tortoise of Zeno's paradox, philosophers such as Kevin Brown and Francis Moorcroft hold that mathematics does not address the central point in Zeno's argument, and that solving the mathematical issues does not solve every issue the paradoxes raise. Brown concludes "Given the history of 'final resolutions', from Aristotle onwards, it's probably foolhardy to think we've reached the end. It may be that Zeno's arguments on motion, because of their simplicity and universality, will always serve as a kind of 'Rorschach image' onto which people can project their most fundamental phenomenological concerns (if they have any)." An alternative conclusion, proposed by Henri Bergson in his 1896 book Matter and Memory, is that, while the path is divisible, the motion is not. In 2003, Peter Lynds argued that all of Zeno's motion paradoxes are resolved by the conclusion that instants in time and instantaneous magnitudes do not physically exist. Lynds argues that an object in relative motion cannot have an instantaneous or determined relative position (for if it did, it could not be in motion), and so cannot have its motion fractionally dissected as if it does, as is assumed by the paradoxes. Nick Huggett argues that Zeno is assuming the conclusion when he says that objects that occupy the same space as they do at rest must be at rest. Based on the work of Georg Cantor, Bertrand Russell offered a solution to the paradoxes, what is known as the "at-at theory of motion". It agrees that there can be no motion "during" a durationless instant, and contends that all that is required for motion is that the arrow be at one point at one time, at another point another time, and at appropriate points between those two points for intervening times. In this view motion is just change in position over time. Another proposed solution is to question one of the assumptions Zeno used in his paradoxes (particularly the Dichotomy), which is that between any two different points in space (or time), there is always another point. Without this assumption there are only a finite number of distances between two points, hence there is no infinite sequence of movements, and the paradox is resolved. According to Hermann Weyl, the assumption that space is made of finite and discrete units is subject to a further problem, given by the "tile argument" or "distance function problem". According to this, the length of the hypotenuse of a right angled triangle in discretized space is always equal to the length of one of the two sides, in contradiction to geometry. Jean Paul Van Bendegem has argued that the Tile Argument can be resolved, and that discretization can therefore remove the paradox. In 1977, physicists E. C. George Sudarshan and B. Misra discovered that the dynamical evolution (motion) of a quantum system can be hindered (or even inhibited) through observation of the system. This effect is usually called the "quantum Zeno effect" as it is strongly reminiscent of Zeno's arrow paradox. This effect was first theorized in 1958. In the field of verification and design of timed and hybrid systems, the system behaviour is called Zeno if it includes an infinite number of discrete steps in a finite amount of time. Some formal verification techniques exclude these behaviours from analysis, if they are not equivalent to non-Zeno behaviour. In systems design these behaviours will also often be excluded from system models, since they cannot be implemented with a digital controller. A humorous take is offered by Tom Stoppard in his 1972 play Jumpers, in which the principal protagonist, the philosophy professor George Moore, suggests that according to Zeno's paradox, Saint Sebastian, a 3rd Century Christian saint martyred by being shot with arrows, died of fright. In 1969, The Firesign Theatre used a version of Zeno's dichotomy paradox when speaking road signs for an approaching exit continued to divide in half, with the destination never being reached by the driver, on their second LP, "How Can You Be In Two Places at Once, When You're Not Anywhere At All." Folk musicians Lou and Peter Berryman provide a humorous variation in their song "An Hour Away", in which the driver of a car keeps encountering construction that requires a slower speed, so that even as they keep driving, their destination remains an hour away and they can't reach their romantic interest. Roughly contemporaneously during the Warring States period (475–221 BCE), ancient Chinese philosophers from the School of Names, a school of thought similarly concerned with logic and dialectics, developed paradoxes similar to those of Zeno. The works of the School of Names have largely been lost, with the exception of portions of the Gongsun Longzi. The second of the Ten Theses of Hui Shi suggests knowledge of infinitesimals:That which has no thickness cannot be piled up; yet it is a thousand li in dimension. Among the many puzzles of his recorded in the Zhuangzi is one very similar to Zeno's Dichotomy: "If from a stick a foot long you every day take the half of it, in a myriad ages it will not be exhausted." The Mohist canon appears to propose a solution to this paradox by arguing that in moving across a measured length, the distance is not covered in successive fractions of the length, but in one stage. Due to the lack of surviving works from the School of Names, most of the other paradoxes listed are difficult to interpret. "What the Tortoise Said to Achilles", written in 1895 by Lewis Carroll, was an attempt to reveal an analogous paradox in the realm of pure logic. If Carroll's argument is valid, the implication is that Zeno's paradoxes of motion are not essentially problems of space and time, but go right to the heart of reasoning itself. Douglas Hofstadter made Carroll's article a centrepiece of his book Gödel, Escher, Bach: An Eternal Golden Braid, writing many more dialogues between Achilles and the Tortoise to elucidate his arguments. Hofstadter connects Zeno's paradoxes to Gödel's incompleteness theorem in an attempt to demonstrate that the problems raised by Zeno are pervasive and manifest in formal systems theory, computing and the philosophy of mind.
{{Infobox OS | name = Debian | logo = Debian-OpenLogo.svg | logo_size = 100px | logo_alt = Debian OpenLogo | screenshot = Debian 12 Bookworm GNOME Desktop English.png | screenshot_alt = Screenshot of Debian 11 (Bullseye) with the [[GNOME]] desktop environment 43.4 | caption = Debian 11 (Bullseye) running its default [[desktop environment]], [[GNOME]] version 43.4 | developer = The Debian Project | family = [[Unix-like]] <!-- per [[template:Infobox OS/doc]] as of December 18, 2019 --> | working_state = Current | source_model = [[Open-source software|Open source]] | released = {{Start date and age|1993|08}} | latest release version = {{wikidata|property|edit|reference|P348}} | latest release date = {{start date and age|{{wikidata|qualifier|P348|P577}}}} | latest preview version = 12 (Bookworm)<ref>{{cite web |url = https://www.debian.org/releases/testing/releasenotes |title = Debian Release Notes |publisher = debian.org |access-date = 14 August 2021 |archive-date = 14 August 2021 |archive-url = https://web.archive.org/web/20210814214331/https://www.debian.org/releases/testing/releasenotes |url-status = live }}</ref> | language = 75 languages | updatemodel = [[Long-term support]] in stable edition, [[rolling release]] in unstable and testing editions | package_manager = [[APT (Debian)|APT]] (front-end), [[dpkg]] | supported_platforms = [[x86-64]], [[arm64]], [https://wiki.debian.org/ArmEabiPort armel]{{Efn | Support ended at Buster(10).}}, [[armhf]], [[i386]], [[MIPS architecture|mips]], [[mipsel]], [[MIPS architecture|mips64el]], [[ppc64el]], [[s390x]],<ref>{{cite web |url = https://www.debian.org/ports/ |title = Debian -- Ports |access-date = May 26, 2014 |archive-date = November 22, 2016 |archive-url = https://web.archive.org/web/20161122100500/http://www.debian.org/ports/ |url-status = live }}</ref> [[RISC-V|riscv64]] (in progress)<ref>{{cite web | url=https://wiki.debian.org/RISC-V | title=RISC-V - Debian Wiki | access-date=2018-01-24 | archive-date=March 20, 2018 | archive-url=https://web.archive.org/web/20180320153651/https://wiki.debian.org/RISC-V | url-status=live }}</ref> | kernel_type = [[Linux kernel]] | userland = [[GNU Core Utilities|GNU]] | ui = * [[GNOME]] on DVD * [[XFCE]] on CD and non-Linux ports * [[MATE (software)|MATE]] available on Debian's website * [[KDE Plasma 5|KDE Plasma]] available on Debian's website * [[LXQt]] available on Debian's website * [[LXDE]] available on Debian's website * [[Cinnamon (desktop environment)|Cinnamon]] available on Debian's website | license = [[Debian Free Software Guidelines|DFSG]]-compatible licenses | website = {{Official URL}} }} '''Debian''' is a free [[operating system]]. It is a distribution of an operating system known as the [[GNU|GNU operating system]], which can be used with various kernels, including [[Linux kernel|Linux]], [[FreeBSD|kFreeBSD]], and [[GNU Hurd|Hurd]]. In combination with these kernels, the operating system can be referred to as Debian GNU/Linux, Debian GNU/kFreeBSD, and Debian GNU/Hurd, respectively. Debian GNU/Linux is one of the most complete and popular [[Linux|GNU/Linux distributions]], on which many others, like [[Ubuntu]], are based. == Brief history == The Debian Project officially started on August 16th, 1993, led by [[Ian Murdock]]. He was a [[computer programmer]]. Today, in this project, Debian is developed by more than 1,000 computer specialists all over the world. The name "Debian" was taken after Ian Murdock and his wife Debra. Some people say or pronounce 'deb-ee-n' but others also say 'de-bi-an' or 'de-bai-an' and in Japan 'de-bi-a-n' and so on. == Development steps == Software packages in development are either uploaded to the project distribution named ''unstable'' (also known as ''sid''), or to the ''experimental'' repository. Software packages uploaded to ''unstable'' are normally versions stable enough to be released by the original [[upstream (software development)|upstream]] developer, but with the added Debian-specific packaging and other modifications introduced by Debian developers. These additions may be new and untested. Software not ready yet for the ''unstable'' distribution is typically placed in the ''experimental'' repository.<ref name="ftparchive">{{cite web | url = http://www.us.debian.org/doc/FAQ/ch-ftparchives | accessdate = 2007-05-24 | title = The Debian GNU/Linux FAQ Chapter 6 - The Debian FTP archives | publisher = Debian | archive-date = 2008-05-15 | archive-url = https://web.archive.org/web/20080515044357/http://www.us.debian.org/doc/FAQ/ch-ftparchives | url-status = dead }}</ref> After a version of a software package has remained in ''unstable'' for a certain length of time (depending on how urgent the changes are), that package is automatically moved to the ''testing'' distribution. The package's move to testing happens only if no serious (''release-critical'') bugs in the package are reported and if other software needed for package functionality qualifies for inclusion in ''testing''.<ref name="ftparchive" /> Since updates to Debian software packages between official releases do not contain new features, some choose to use the ''testing'' and ''unstable'' distributions for their newer packages. However, these distributions are less tested than ''stable'', and ''unstable'' does not receive timely security updates. In particular, incautious upgrades to working ''unstable'' packages can sometimes seriously break software functionality.<ref name="securityfaq"> {{cite web | url = http://www.debian.org/security/faq | title = Debian security FAQ | date = 2007-02-28 | accessdate = 2008-10-21 | publisher = Debian }} </ref> Since September 9, 2005<ref> {{cite mailing list | url = http://lists.debian.org/debian-devel-announce/2005/09/msg00006.html | title = announcing the beginning of security support for testing | date = 2005-09-05 | accessdate = 2007-04-20 | mailing-list = debian-devel-announce | last = Hess | first = Joey }} </ref> the ''testing'' distributions security updates have been provided by the ''testing'' security team.<ref>{{cite web | url = http://testing-security.debian.net | title = Debian testing security team | publisher = Debian | accessdate = 2008-10-31 | archive-date = 2014-12-29 | archive-url = https://web.archive.org/web/20141229131743/http://testing-security.debian.net/ | url-status = dead }}</ref> After the packages in ''testing'' have matured and the goals for the next release are met, the ''testing'' distribution becomes the next stable release. The latest stable release of Debian (''Buster'') is 10.0, released on July 6, 2019. The next release is codenamed "''Bullseye''".<ref name="ftparchive" /> === Release history === <!-- NOTE: Use this article in English Wikipedia to help update https://en.wikipedia.org/wiki/Debian_version_history --> {| class="wikitable" style="float: center;" |- ! Legend |- | style="background-color:#F06C47;" | Release no longer supported |- | style="background-color:#A0E75A;" | Release still supported |- | style="background-color:#87CEEB;" | Future release |} {| class="wikitable" ! Version ! [[Code name|Code&nbsp;name]] ! Release date ! [[Debian#Ports|Ports]] ! Packages ! Supported until ! Notes |- | 1.1 | ''[[Buzz Lightyear|buzz]]'' | align="right" | 17 June 1996 | align="right" | 1 | align="right" | 474 | style="background-color:#F06C47;" | 1996{{Citation needed|date=December 2009}} | <tt>dpkg</tt>, [[Executable and Linkable Format|ELF]] transition, Linux 2.0 |- | 1.2 | ''rex'' | align="right" | 12 December 1996 | align="right" | 1 | align="right" | 848 | style="background-color:#F06C47;" | 1996{{Citation needed|date=December 2009}} | - |- | 1.3 | ''bo'' | align="right" | 5 June 1997 | align="right" | 1 | align="right" | 974 | style="background-color:#F06C47;" | 1997{{Citation needed|date=December 2009}} | - |- | 2.0 | ''hamm'' | align="right" | 24 July 1998 | align="right" | 2 | align="right" |≈&nbsp;1,500 | style="background-color:#F06C47;" | 1998 | [[GNU C Library|glibc]] transition, new architecture: <tt>m68k</tt> |- | 2.1 | ''slink'' | align="right" | 9 March 1999 | align="right" | 4 | align="right" | ≈&nbsp;2,250 | style="background-color:#F06C47;" | 2000-12 | <tt>APT</tt>, new architectures: <tt>alpha</tt>, <tt>sparc</tt> |- | 2.2 | ''potato'' | align="right" | 15 August 2000 | align="right" | 6 | align="right" | ≈&nbsp;3,900 | style="background-color:#F06C47;" | 2003-04 | New architectures: <tt>arm</tt>, <tt>powerpc</tt><ref name="22rel">{{cite mailing list | url = http://lists.debian.org/debian-announce/debian-announce-2000/msg00009.html | title = Debian GNU/Linux 2.2, the "Joel 'Espy' Klecker" release | mailing-list = debian-announce | date = 2000-08-15 | last = Schulze | first = Martin }}</ref> |- | 3.0 | ''[[Sheriff Woody|woody]]'' | align="right" | 19 July 2002 | align="right" | 11 | align="right" | ≈&nbsp;8,500 | style="background-color:#F06C47;" | 2006-08 | New architectures: <tt>hppa</tt>, <tt>ia64</tt>, <tt>mips</tt>, <tt>mipsel</tt>, <tt>s390</tt> |- | 3.1 | ''sarge'' | align="right" | 6 June 2005 | align="right" | 11 | align="right" | ≈&nbsp;15,400 | style="background-color:#F06C47;" | 2008-04<ref name="securityfaq" /> | Modular installer, semi-official <tt>amd64</tt> support. |- | 4.0 | ''etch'' | align="right" | 8 April 2007 | align="right" | 11 | align="right" | ≈&nbsp;18,000 | style="background-color:#F06C47;" | 2010-02-15<ref name="wiki.debian.org-DebianEtch">{{cite web | url = http://wiki.debian.org/DebianEtch | title = Debian Wiki: Debian Releases > Debian Etch | accessdate = 2010-07-16 | publisher = Debian }}</ref> | New architecture: <tt>amd64</tt>, dropped architecture: <tt>m68k</tt>.<ref name="40rel">{{cite mailing list | mailing-list = debian-announce | last = Schmehl | first = Alexander | url = http://lists.debian.org/debian-announce/debian-announce-2007/msg00002.html | title = Debian GNU/Linux 4.0 released | date = 2007-04-08 | accessdate = 2008-11-01 }}</ref> Graphical installer, [[udev]] transition, modular [[X.Org Server|X.Org]] transition. Latest update 4.0r9 was released 2010-05-22<ref>{{cite web | title = Debian GNU/Linux 4.0 updated | publisher = Debian | url = http://www.debian.org/News/2010/20100522 | accessdate = 2010-07-16 | archive-date = 2015-09-05 | archive-url = https://web.archive.org/web/20150905074801/http://www.debian.org/News/2010/20100522 | url-status = dead }}</ref> |- | 5.0<ref>{{cite mailing list | url = http://lists.debian.org/debian-devel-announce/2008/03/msg00001.html | title = Release Update: Release numbering, goals, armel architecture, BSPs | mailing-list = debian-announce | date = 2008-03-02 | accessdate = 2008-11-01 | first = Marc | last = Brockschmidt }}</ref> | ''lenny''<ref>{{cite mailing list | url = http://lists.debian.org/debian-devel-announce/2006/11/msg00004.html | title = testing d-i Release Candidate 1 and more release adjustments | mailing-list = debian-devel-announce | date = 2006-11-16 | first = Steve | last = Langasek | accessdate = 2008-11-01 }}</ref> | align="right" | 14&nbsp;February&nbsp;2009 | align="right" | 11+1{{Ref label|lenny|A|A}} | align="right" | ≈&nbsp;23,000<ref name="lenny released" /> | style="background-color:#F06C47;" | 2012-02-06 | New architecture/binary ABI: <tt>armel</tt>.<ref>{{cite mailing list | url = http://lists.debian.org/debian-devel-announce/2008/06/msg00000.html | title = Release Update: arch status, major transitions finished, freeze coming up | first = Marc | last = Brockschmidt | date = 2008-06-02 | accessdate = 2008-11-01 | mailing-list = debian-devel-announce }}</ref> [[SPARC]] 32-bit hardware support dropped.<ref>{{cite mailing list | url = http://lists.debian.org/debian-devel-announce/2007/07/msg00006.html | title = Retiring the sparc32 port | last = Smakov | first = Jurij | mailing-list = debian-devel-announce | date = 2007-07-18 | accessdate = 2008-10-31 }}</ref> Full [[ASUS Eee PC|Eee PC]] support.<ref>{{cite mailing list | url = http://lists.debian.org/debian-devel-announce/2008/08/msg00002.html | title = Bits from the Debian Eee PC team, summer 2008 | mailing-list = debian-devel-announce | first = Ben | last = Armstrong | date = 2008-08-03 | accessdate = 2008-10-31}}</ref> Latest update 5.0.8 was released 2011-01-22.<ref>{{cite web | title = Debian -- News -- Updated Debian GNU/Linux: 5.0.8 released | publisher = Debian | url = http://www.debian.org/News/2011/20110122 | accessdate = 2011-01-22 | archive-date = 2011-01-24 | archive-url = https://web.archive.org/web/20110124193529/http://www.debian.org/News/2011/20110122 | url-status = dead }}</ref> |- | 6.0<ref name="squeezenotes">{{cite web | url = http://www.debian.org/releases/squeeze/releasenotes | title = Debian GNU/Linux 6.0&nbsp;– Release Notes | accessdate = 2009-02-15 | publisher = Debian }}</ref> | ''squeeze''<ref>{{cite mailing list | url = http://lists.debian.org/debian-devel-announce/2008/09/msg00000.html | title = Release Update: freeze guidelines, testing, BSP, rc bug fixes | mailing-list = debian-devel-announce | date = 2008-09-01 | first = Luk | last = Claes | accessdate = 2008-10-31 }}</ref> | align="right" | 6 February 2011<ref>{{cite web | url = http://www.debian.org/News/2011/20110205a | accessdate = 2011-02-06 | title = Bits from the release team: Planning, request for help | first = Luk | last = Claes | archive-date = 2011-09-02 | archive-url = https://web.archive.org/web/20110902020150/http://www.debian.org/News/2011/20110205a | url-status = dead }}</ref> | align="right" | 9+2{{Ref label|squeeze|B|B}} | align="right" | ≈&nbsp;29,000 | style="background-color:#F06C47;" | 2016-02-29 | New architectures/kernels: <tt>kfreebsd-i386</tt>, <tt>kfreebsd-amd64</tt>, dropped architectures: <tt>alpha</tt>, <tt>arm</tt>.<ref name="squeeze release architectures">{{cite web | url = http://lists.debian.org/debian-devel-announce/2009/10/msg00000.html | accessdate = 2009-10-01 | title = Release architectures | first = Andreas | last = Barth }}</ref> [[Embedded GLIBC|eglibc]] in favour of [[GNU C Library|glibc]].<ref>{{cite web | url = http://blog.aurel32.net/?p=47 | accessdate = 2009-05-21 | title = Aurelien's weblog: Debian is switching to EGLIBC | publisher = Aurélien Jarno }}</ref> |- |7 |style="text-align: left;"|wheezy | align="right" | 4 May 2013 | align="right" | 13 | align="right" | ≈ 36,000 | style="background-color:#F06C47;" | 2018-05 | |- |8 |style="text-align: left;"|jessie | align="right" | 25–26 April 2015 | align="right" | 10 | align="right" | ≈ 43,000 | style="background-color:#F06C47;" | 2020-04 | |- |9 |style="text-align: left;"|stretch | align="right" | 17 June 2017 | align="right" | 10 | align="right" | ≈ 52,000 | style="background-color:#F06C47;" | 2022-06 | |- |10 |style="text-align: left;"|buster | align="right" | 6 July 2019 | align="right" | 10 | align="right" | ≈ 58,000 | style="background-color:#A0E75A;" | 2024-06 |- |11 |style="text-align: left;"|bullseye | align="right" | 14 August 2021 | align="right" | 9 | align="right" | 59,551 | style="background-color:#A0E75A;" | 2026-06 | |- |12 |bookworm |10 June 2023 | |64,419 |2028 | |} :{{note label|lenny|A|A}} 11 architectures + 1 additional ARM binary ABI (<tt>armel</tt>)<ref name="lenny released">{{cite web |url=http://www.debian.org/News/2009/20090214.en.html |title=Debian GNU/Linux 5.0 released |accessdate=2009-02-15 |publisher=Debian |date=2009-02-14 |archive-date=2009-02-17 |archive-url=https://web.archive.org/web/20090217184902/http://www.debian.org/News/2009/20090214.en.html |url-status=dead }}</ref> :{{note label|squeeze|B|B}} 9 architectures with Linux kernel + 2 architectures with FreeBSD kernel<ref name="squeeze release architectures" /> {| style="float: right" | [[File:Debian-cd-cover1.png|150px|thumb|A Debian 4.0 Box Cover<ref name="dvdcover">{{cite web | url = http://www.debian.org/CD/artwork/ | title = Artwork for Debian CDs | accessdate = 2009-01-11 | publisher = Debian }} </ref>]] |} Due to an incident involving a CD vendor who made an unofficial and broken release labeled ''1.0'', an official 1.0 release was never made.<ref name="releases"> {{cite web | url = http://wiki.debian.org/DebianReleases | title = Debian Wiki: Debian Releases | publisher = Debian | accessdate = 2010-07-16 }} </ref> == For other platforms == Debian has been ported to different architectures or platforms. One version, which is based on the developer release (sid) that has been ported to the [[Xbox (console)|Xbox]] is called Xebian. == Package == Debian's official software package repository<ref>ArchLinux.org, [https://wiki.archlinux.org/index.php/Official_Repositories "Official repositories,"] excerpt, "A software repository is a storage location from which software packages may be retrieved and installed on a computer"; retrieved 2012-6-7.</ref> includes, for example, [[UNetbootin]].<ref>Debian, [http://packages.debian.org/search?keywords=unetbootin Package unetbootin]; retrieved 2012-6-7.</ref> ==Related pages== * [[List of Linux distributions]] == References == {{reflist}} ;Notes <references group="lower-alpha"/> == Other websites == * http://www.debian.org/ - Debian Project Homepage * {{DistroWatch|debian|NAME=Debian}} {{authority control}} [[Category:Linux distributions]] [[Category:Debian]] [[Category:X86-64 Linux distributions]]
Debian (/ˈdɛbiən/), also known as Debian GNU/Linux, is a Linux distribution composed of free and open-source software and proprietary software developed by the community-supported Debian Project, which was established by Ian Murdock on August 16, 1993. The first version of Debian (0.01) was released on September 15, 1993, and its first stable version (1.1) was released on June 17, 1996. The Debian Stable branch is the most popular edition for personal computers and servers. Debian is also the basis for many other distributions, like PureOS, Ubuntu, Pardus, and Linux Mint. Debian is one of the oldest operating systems based on the Linux kernel and, as of September 2023, the second oldest Linux distribution still in active development, only behind Slackware. The project is coordinated over the Internet by a team of volunteers guided by the Debian Project Leader and three foundational documents: the Debian Social Contract, the Debian Constitution, and the Debian Free Software Guidelines. New distributions are updated continually, and the next candidate is released after a time-based freeze. Since its founding until 2022, Debian has been developed openly and distributed freely according to some of the principles of the GNU Project. Because of this, the Free Software Foundation sponsored the project from November 1994 to November 1995. On June 16, 1997, the Debian Project founded the nonprofit organization Software in the Public Interest to continue financially supporting development. Debian distribution codenames are based on the names of characters from the Toy Story films. Debian's unstable trunk is named after Sid, a character who regularly destroyed his toys. First announced on August 16, 1993, Debian was founded by Ian Murdock, who initially called the system "the Debian Linux Release". The word "Debian" was formed as a portmanteau of the first name of his then-girlfriend (later ex-wife) Debra Lynn and his own first name. Before Debian's release, the Softlanding Linux System (SLS) had been a popular Linux distribution and the basis for Slackware. The perceived poor maintenance and prevalence of bugs in SLS motivated Murdock to launch a new distribution. Debian 0.01, released on September 15, 1993, was the first of several internal releases. Version 0.90 was the first public release, providing support through mailing lists hosted at Pixar. The release included the Debian Linux Manifesto, outlining Murdock's view for the new operating system. In it he called for the creation of a distribution to be maintained "openly in the spirit of Linux and GNU." The Debian project released the 0.9x versions in 1994 and 1995. During this time it was sponsored by the Free Software Foundation for one year. Ian Murdock delegated the base system, the core packages of Debian, to Bruce Perens and Murdock focused on the management of the growing project. The first ports to non-IA-32 architectures began in 1995, and Debian 1.1 was released in 1996. By that time and thanks to Ian Jackson, the dpkg package manager was already an essential part of Debian. In 1996, Bruce Perens assumed the project leadership. Perens was a controversial leader, regarded as authoritarian and strongly attached to Debian. He drafted a social contract and edited suggestions from a month-long discussion into the Debian Social Contract and the Debian Free Software Guidelines. After the FSF withdrew their sponsorship in the midst of the free software vs. open source debate, Perens initiated the creation of the legal umbrella organization Software in the Public Interest instead of seeking renewed involvement with the FSF. He led the conversion of the project from a.out to ELF. He created the BusyBox program to make it possible to run a Debian installer on a single floppy disk, and wrote a new installer. By the time Debian 1.2 was released, the project had grown to nearly two hundred volunteers. Perens left the project in 1998. Ian Jackson became the leader in 1998. Debian 2.0 introduced the second official port, m68k. During this time the first port to a non-Linux kernel, Debian GNU/Hurd, was started. On December 2, the first Debian Constitution was ratified. From 1999, the project leader was elected yearly. The Advanced Packaging Tool was deployed with Debian 2.1. The number of applicants was overwhelming and the project established the new member process. The first Debian derivatives, namely Libranet, Corel Linux and Stormix's Storm Linux, were started in 1999. The 2.2 release in 2000 was dedicated to Joel Klecker, a developer who died of Duchenne muscular dystrophy. In late 2000, the project reorganized the archive with new package "pools" and created the Testing distribution, made up of packages considered stable, to reduce the freeze for the next release. In the same year, developers began holding an annual conference called DebConf with talks and workshops for developers and technical users. In May 2001, Hewlett-Packard announced plans to base its Linux development on Debian. In July 2002, the project released version 3.0, code-named Woody, the first release to include cryptographic software, a free licensed KDE and internationalization. During these last release cycles, the Debian project drew considerable criticism from the free software community because of the long time between stable releases. Some events disturbed the project while working on Sarge, as Debian servers were attacked by fire and hackers. One of the most memorable was the Vancouver prospectus. After a meeting held in Vancouver, release manager Steve Langasek announced a plan to reduce the number of supported ports to four in order to shorten future release cycles. There was a large reaction because the proposal looked more like a decision and because such a drop would damage Debian's aim to be "the universal operating system". The first version of the Debian-based Ubuntu, named "4.10 Warty Warthog", was released on October 20, 2004. Because it was distributed as a free download, it became one of the most popular and successful operating systems with more than "40 million users" according to Canonical Ltd. However, Murdock was critical of the differences between Ubuntu packages and Debian, stating that it leads to incompatibilities. The 3.1 Sarge release was made in June 2005. This release updated 73% of the software and included over 9,000 new packages. A new installer with a modular design, Debian-Installer, allowed installations with RAID, XFS and LVM support, improved hardware detection, made installations easier for novice users, and was translated into almost forty languages. An installation manual and release notes were in ten and fifteen languages respectively. The efforts of Skolelinux, Debian-Med and Debian-Accessibility raised the number of packages that were educational, had a medical affiliation, and ones made for people with disabilities. In 2006, as a result of a much-publicized dispute, Mozilla software was rebranded in Debian, with Firefox forked as Iceweasel and Thunderbird as Icedove. The Mozilla Corporation stated that software with unapproved modifications could not be distributed under the Firefox trademark. Two reasons that Debian modifies the Firefox software are to change the non-free artwork and to provide security patches. In February 2016, it was announced that Mozilla and Debian had reached an agreement and Iceweasel would revert to the name Firefox; similar agreement was anticipated for Icedove/Thunderbird. A fund-raising experiment, Dunc-Tank, was created to solve the release cycle problem and release managers were paid to work full-time; in response, unpaid developers slowed down their work and the release was delayed. Debian 4.0 (Etch) was released in April 2007, featuring the x86-64 port and a graphical installer. Debian 5.0 (Lenny) was released in February 2009, supporting Marvell's Orion platform and netbooks such as the Asus Eee PC. The release was dedicated to Thiemo Seufer, a developer who died in a car crash. In July 2009, the policy of time-based development freezes on a two-year cycle was announced. Time-based freezes are intended to blend the predictability of time based releases with Debian's policy of feature based releases, and to reduce overall freeze time. The Squeeze cycle was going to be especially short; however, this initial schedule was abandoned. In September 2010, the backports service became official, providing more recent versions of some software for the stable release. Debian 6.0 (Squeeze) was released in February 2011, featuring Debian GNU/kFreeBSD as a technology preview, along with adding a dependency-based boot system, and moving problematic firmware to the non-free section. Debian 7.0 (Wheezy) was released in May 2013, featuring multiarch support. Debian 8.0 (Jessie) was released in April 2015, using systemd as the new init system. Debian 9.0 (Stretch) was released in June 2017, with nftables as a replacement for iptables, support for Flatpak apps, and MariaDB as the replacement for MySQL. Debian 10.0 (Buster) was released in July 2019, adding support for Secure Boot and enabling AppArmor by default. Debian 11.0 (Bullseye) was released in August 2021, enabling persistency in the system journal, adding support for driverless scanning, and containing kernel-level support for exFAT filesystems. Debian 12.0 (Bookworm) was released on June 10, 2023, including various improvements and features, increasing the supported Linux Kernel to version 6.1, and leveraging new "Emerald" artwork. Debian is still in development and new packages are uploaded to unstable every day. Debian used to be released as a very large set of CDs for each architecture, but with the release of Debian 9 (Stretch) in 2017, many of the images have been dropped from the archive but remain buildable via jigdo. Throughout Debian's lifetime, both the Debian distribution and its website have won various awards from different organizations, including Server Distribution of the Year 2011, The best Linux distro of 2011, and a Best of the Net award for October 1998. On December 2, 2015, Microsoft announced that they would offer Debian GNU/Linux as an endorsed distribution on the Azure cloud platform. Microsoft has also added a user environment to their Windows 10 desktop operating system called Windows Subsystem for Linux that offers a Debian subset. Debian has access to online repositories that contain over 51,000 packages. Debian officially contains only free software, but non-free software can be downloaded and installed from the Debian repositories. Debian includes popular free programs such as LibreOffice, Firefox web browser, Evolution mail, K3b disc burner, VLC media player, GIMP image editor, and Evince document viewer. Debian is a popular choice for servers, for example as the operating system component of a LAMP stack. Several flavors of the Linux kernel exist for each port. For example, the i386 port has flavors for IA-32 PCs supporting Physical Address Extension and real-time computing, for older PCs, and for x86-64 PCs. The Linux kernel does not officially contain firmware lacking source code, although such firmware is available in non-free packages and alternative installation media. Debian offers CD and DVD images specifically built for Xfce, GNOME, KDE, MATE, Cinnamon, LXDE, and LXQt. MATE support was added in 2014, and Cinnamon support was added with Debian 8.0 Jessie. Less common window managers such as Enlightenment, Openbox, Fluxbox, IceWM, Window Maker and others are available. The default desktop environment of version 7.0 Wheezy was temporarily switched to Xfce, because GNOME 3 did not fit on the first CD of the set. The default for the version 8.0 Jessie was changed again to Xfce in November 2013, and back to GNOME in September 2014. Several parts of Debian are translated into languages other than American English, including package descriptions, configuration messages, documentation and the website. The level of software localization depends on the language, ranging from the highly supported German and French to the barely translated Creek and Samoan. The Debian 10 installer is available in 76 languages. Multimedia support has been problematic in Debian regarding codecs threatened by possible patent infringements, lacking source code, or under too restrictive licenses. Even though packages with problems related to their distribution could go into the non-free area, software such as libdvdcss is not hosted at Debian . A notable third party repository exists, formerly named Debian-multimedia.org, providing software not present in Debian such as Windows codecs, libdvdcss and the Adobe Flash Player. Even though this repository is maintained by Christian Marillat, a Debian developer, it is not part of the project and is not hosted on a Debian server. The repository provides packages already included in Debian, interfering with the official maintenance. Eventually, project leader Stefano Zacchiroli asked Marillat to either settle an agreement about the packaging or to stop using the "Debian" name. Marillat chose the latter and renamed the repository to deb-multimedia.org. The repository was so popular that the switchover was announced by the official blog of the Debian project. Debian offers DVD and CD images for installation that can be downloaded using BitTorrent or jigdo. Physical discs can also be bought from retailers. The full sets are made up of several discs (the amd64 port consists of 13 DVDs or 84 CDs), but only the first disc is required for installation, as the installer can retrieve software not contained in the first disc image from online repositories. Debian offers different network installation methods. A minimal install of Debian is available via the netinst CD, whereby Debian is installed with just a base and later added software can be downloaded from the Internet. Another option is to boot the installer from the network. The default bootstrap loader is GNU GRUB version 2, though the package name is simply grub, while version 1 was renamed to grub-legacy. This conflicts with distros (e.g., Fedora Linux), where grub version 2 is named grub2. The default desktop may be chosen from the DVD boot menu among GNOME, KDE Plasma, Xfce and LXDE, and from special disc 1 CDs. Debian releases live install images for CDs, DVDs and USB thumb drives, for IA-32 and x86-64 architectures, and with a choice of desktop environments. These Debian Live images allow users to boot from removable media and run Debian without affecting the contents of their computer. A full install of Debian to the computer's hard drive can be initiated from the live image environment. Personalized images can be built with the live-build tool for discs, USB drives and for network booting purposes. Installation images are hybrid on some architectures and can be used to create a bootable USB drive (Live USB). Package management operations can be performed with different tools available on Debian, from the lowest level command dpkg to graphical front-ends like Synaptic. The recommended standard for administering packages on a Debian system is the apt toolset. dpkg provides the low-level infrastructure for package management. The dpkg database contains the list of installed software on the current system. The dpkg command tool does not know about repositories. The command can work with local .deb package files, and information from the dpkg database. An Advanced Packaging Tool (APT) allows administering an installed Debian system to retrieve and resolve package dependencies from repositories. APT tools share dependency information and cached packages. GDebi is an APT tool which can be used in command-line and on the GUI. GDebi can install a local .deb file via the command line like the dpkg command, but with access to repositories to resolve dependencies. Other graphical front-ends for APT include Software Center, Synaptic and Apper. GNOME Software is a graphical front-end for PackageKit, which itself can work on top of various software packaging systems. The Debian Free Software Guidelines (DFSG) define the distinctive meaning of the word "free" as in "free and open-source software". Packages that comply with these guidelines, usually under the GNU General Public License, Modified BSD License or Artistic License, are included inside the main area; otherwise, they are included inside the non-free and contrib areas. These last two areas are not distributed within the official installation media, but they can be adopted manually. Non-free includes packages that do not comply with the DFSG, such as documentation with invariant sections and proprietary software, and legally questionable packages. Contrib includes packages which do comply with the DFSG but fail other requirements. For example, they may depend on packages which are in non-free or requires such for building them. Richard Stallman and the Free Software Foundation have criticized the Debian project for hosting the non-free repository and because the contrib and non-free areas are easily accessible, an opinion echoed by some in Debian including the former project leader Wichert Akkerman. The internal dissent in the Debian project regarding the non-free section has persisted, but the last time it came to a vote in 2004, the majority decided to keep it. The most popular optional Linux cross-distribution package manager are graphical (front-ends) package managers. They are available within the official Debian Repository but are not installed by default. They are widely popular with both Debian users and Debian software developers who are interested in installing the most recent versions of application or using the cross-distribution package manager built-in sandbox environment. While at the same time remaining in control of the security. Four most popular cross-distribution package managers, sorted in alphabetical order: Three branches of Debian (also called releases, distributions or suites) are regularly maintained: Other branches in Debian: The snapshot archive provides older versions of the branches. They may be used to install a specific older version of some software. Stable and oldstable get minor updates, called point releases; as of August 2021, the stable release is version 11.7, released on April 29, 2023; 7 months ago (2023-04-29), and the oldstable release is version 10.10. The numbering scheme for the point releases up to Debian 4.0 was to include the letter r (for revision) after the main version number and then the number of the point release; for example, the latest point release of version 4.0 is 4.0r9. This scheme was chosen because a new dotted version would make the old one look obsolete and vendors would have trouble selling their CDs. From Debian 5.0, the numbering scheme of point releases was changed, conforming to the GNU version numbering standard; the first point release of Debian 5.0 was 5.0.1 instead of 5.0r1. The numbering scheme was once again changed for the first Debian 7 update, which was version 7.1. The r scheme is no longer in use, but point release announcements include a note about not throwing away old CDs. The Debian "swirl" logo was designed by Raul Silva in 1999 as part of a contest to replace the semi-official logo that had been used. The winner of the contest received an @Debian.org email address, and a set of Debian 2.1 install CDs for the architecture of their choice. Initially, the swirl was magic smoke arising from an also included bottle of an Arabian-style genie presented in black profile, but shortly after was reduced to the red smoke swirl for situations where space or multiple colours were not an option, and before long the bottle version effectively was superseded. There has been no official statement from the Debian project on the logo's meaning, but at the time of the logo's selection, it was suggested that the logo represented the magic smoke ( or the genie ) that made computers work. One theory about the origin of the Debian logo is that Buzz Lightyear, the chosen character for the first named Debian release, has a swirl in his chin. Stefano Zacchiroli also suggested that this swirl is the Debian one. Buzz Lightyear's swirl is a more likely candidate as the codenames for Debian are names of Toy Story characters. The former Debian project leader Bruce Perens used to work for Pixar and is credited as a studio tools engineer on Toy Story 2 (1999). Hardware requirements are at least those of the kernel and the GNU toolsets. Debian's recommended system requirements depend on the level of installation, which corresponds to increased numbers of installed components: The real minimum memory requirements depend on the architecture and may be much less than the numbers listed in this table. It is possible to install Debian with 170 MB of RAM for x86-64; the installer will run in low memory mode and it is recommended to create a swap partition. The installer for z/Architecture requires about 20 MB of RAM, but relies on network hardware. Similarly, disk space requirements, which depend on the packages to be installed, can be reduced by manually selecting the packages needed. As of May 2019, no Pure Blend exists that would lower the hardware requirements easily. It is possible to run graphical user interfaces on older or low-end systems, but the installation of window managers instead of desktop environments is recommended, as desktop environments are more resource intensive. Requirements for individual software vary widely and must be considered, with those of the base operating environment. As of the upcoming Trixie release, the official ports are: Unofficial ports are available as part of the unstable distribution: Debian supports a variety of ARM-based NAS devices. The NSLU2 was supported by the installer in Debian 4.0 and 5.0, and Martin Michlmayr is providing installation tarballs since version 6.0. Other supported NAS devices are the Buffalo Kurobox Pro, GLAN Tank, Thecus N2100 and QNAP Turbo Stations. Devices based on the Kirkwood system on a chip (SoC) are supported too, such as the SheevaPlug plug computer and OpenRD products. There are efforts to run Debian on mobile devices, but this is not a project goal yet since the Debian Linux kernel maintainers would not apply the needed patches. Nevertheless, there are packages for resource-limited systems. There are efforts to support Debian on wireless access points. Debian is known to run on set-top boxes. Work is ongoing to support the AM335x processor, which is used in electronic point of service solutions. Debian may be customized to run on cash machines. BeagleBoard, a low-power open-source hardware single-board computer (made by Texas Instruments) has switched to Debian Linux preloaded on its Beaglebone Black board's flash. Roqos Core, manufactured by Roqos, is a x86-64 based IPS firewall router running Debian Linux. Debian's policies and team efforts focus on collaborative software development and testing processes. As a result, a new major release tends to occur every two years with revision releases that fix security issues and important problems. The Debian project is a volunteer organization with three foundational documents: Debian developers are organized in a web of trust. There are at present about one thousand active Debian developers, but it is possible to contribute to the project without being an official developer. The project maintains official mailing lists and conferences for communication and coordination between developers. For issues with single packages and other tasks, a public bug tracking system is used by developers and end users. Internet Relay Chat is also used for communication among developers and to provide real time help. Debian is supported by donations made to organizations authorized by the leader. The largest supporter is Software in the Public Interest, the owner of the Debian trademark, manager of the monetary donations and umbrella organization for various other community free software projects. A Project Leader is elected once per year by the developers. The leader has special powers, but they are not absolute, and appoints delegates to perform specialized tasks. Delegates make decisions as they think is best, taking into account technical criteria and consensus. By way of a General Resolution, the developers may recall the leader, reverse a decision made by the leader or a delegate, amend foundational documents and make other binding decisions. The voting method is based on the Schulze method (Cloneproof Schwartz Sequential Dropping). Project leadership is distributed occasionally. Branden Robinson was helped by the Project Scud, a team of developers that assisted the leader, but there were concerns that such leadership would split Debian into two developer classes. Anthony Towns created a supplemental position, Second In Charge (2IC), that shared some powers of the leader. Steve McIntyre was 2IC and had a 2IC himself. One important role in Debian's leadership is that of a release manager. The release team sets goals for the next release, supervises the processes and decides when to release. The team is led by the next release managers and stable release managers. Release assistants were introduced in 2003. The Debian Project has an influx of applicants wishing to become developers. These applicants must undergo a vetting process which establishes their identity, motivation, understanding of the project's principles, and technical competence. This process has become much harder throughout the years. Debian developers join the project for many reasons. Some that have been cited include: Debian developers may resign their positions at any time or, when deemed necessary, they can be expelled. Those who follow the retiring protocol are granted the "emeritus" status and they may regain their membership through a shortened new member process. Each software package has a maintainer that may be either one person or a team of Debian developers and non-developer maintainers. The maintainer keeps track of upstream releases, and ensures that the package coheres with the rest of the distribution and meets the standards of quality of Debian. Packages may include modifications introduced by Debian to achieve compliance with Debian Policy, even to fix non-Debian specific bugs, although coordination with upstream developers is advised. The maintainer releases a new version by uploading the package to the "incoming" system, which verifies the integrity of the packages and their digital signatures. If the package is found to be valid, it is installed in the package archive into an area called the "pool" and distributed every day to hundreds of mirrors worldwide. The upload must be signed using OpenPGP-compatible software. All Debian developers have individual cryptographic key pairs. Developers are responsible for any package they upload even if the packaging was prepared by another contributor. Initially, an accepted package is only available in the unstable branch. For a package to become a candidate for the next release, it must migrate to the Testing branch by meeting the following: Thus, a release-critical bug in a new version of a shared library on which many packages depend may prevent those packages from entering Testing, because the updated library must meet the requirements too. From the branch viewpoint, the migration process happens twice per day, rendering Testing in perpetual beta. Periodically, the release team publishes guidelines to the developers in order to ready the release. A new release occurs after a freeze, when all important software is reasonably up-to-date in the Testing branch and any other significant issues are solved. At that time, all packages in the testing branch become the new stable branch. Although freeze dates are time-based, release dates are not, which are announced by the release managers a couple of weeks beforehand. A version of a package can belong to more than one branch, usually testing and unstable. It is possible for a package to keep the same version between stable releases and be part of oldstable, stable, testing and unstable at the same time. Each branch can be seen as a collection of pointers into the package "pool" mentioned above. One way to resolve the challenge of a release-critical bug in a new application version is the use of optional package managers. They allow software developers to use sandbox environments, while at the same time remaining in control of security. Another benefit of a cross-distribution package manager is that they allow application developers to directly provide updates to users without going through distributions, and without having to package and test the application separately for each distribution. A new stable branch of Debian gets released approximately every 2 years. It will receive official support for about 3 years with update for major security or usability fixes. Point releases will be available every several months as determined by Stable Release Managers (SRM). Debian also launched its Long Term Support (LTS) project since Debian 6 (Debian Squeeze). For each Debian release, it will receive two years of extra security updates provided by LTS Team after its End Of Life (EOL). However, no point releases will be made. Now each Debian release can receive 5 years of security support in total. The Debian project handles security through public disclosure. Debian security advisories are compatible with the Common Vulnerabilities and Exposures dictionary, are usually coordinated with other free software vendors and are published the same day a vulnerability is made public. There used to be a security audit project that focused on packages in the stable release looking for security bugs; Steve Kemp, who started the project, retired in 2011 but resumed his activities and applied to rejoin in 2014. The stable branch is supported by the Debian security team; oldstable is supported for one year. Although Squeeze is not officially supported, Debian is coordinating an effort to provide long-term support (LTS) until February 2016, five years after the initial release, but only for the IA-32 and x86-64 platforms. Testing is supported by the testing security team, but does not receive updates in as timely a manner as stable. Unstable's security is left for the package maintainers. The Debian project offers documentation and tools to harden a Debian installation both manually and automatically. AppArmor support is available and enabled by default since Buster. Debian provides an optional hardening wrapper, and does not harden all of its software by default using gcc features such as PIE and buffer overflow protection, unlike operating systems such as OpenBSD, but tries to build as many packages as possible with hardening flags. In May 2008, a Debian developer discovered that the OpenSSL package distributed with Debian and derivatives such as Ubuntu made a variety of security keys vulnerable to a random number generator attack, since only 32,767 different keys were generated. The security weakness was caused by changes made in 2006 by another Debian developer in response to memory debugger warnings. The complete resolution procedure was cumbersome because patching the security hole was not enough; it involved regenerating all affected keys and certificates. The cost of developing all of the packages included in Debian 5.0 Lenny (323 million lines of code) has been estimated to be about US$8 billion, using one method based on the COCOMO model. As of 2016, Black Duck Open Hub estimates that the current codebase (74 million lines of code) would cost about US$1.4 billion to develop, using a different method based on the same model. A large number of forks and derivatives have been built upon Debian over the years. Among the more notable are Ubuntu, developed by Canonical Ltd. and first released in 2004, which has surpassed Debian in popularity with desktop users; Knoppix, first released in the year 2000 and one of the first distributions optimized to boot from external storage; and Devuan, which gained attention in 2014 when it forked in disagreement over Debian's adoption of the systemd software suite, and has been mirroring Debian releases since 2017. The Linux Mint Debian Edition (LMDE) uses Debian Stable as the software source base since 2014. Debian is one of the most popular Linux distributions, and many other distributions have been created from the Debian codebase. As of 2021, DistroWatch lists 121 active Debian derivatives. The Debian project provides its derivatives with guidelines for best practices and encourages derivatives to merge their work back into Debian. Debian Pure Blends are subsets of a Debian release configured out-of-the-box for users with particular skills and interests. For example, Debian Jr. is made for children, while Debian Science is for researchers and scientists. The complete Debian distribution includes all available Debian Pure Blends. "Debian Blend" (without "Pure") is a term for a Debian-based distribution that strives to become part of mainstream Debian, and have its extra features included in future releases. Debian GNU/kFreeBSD is a discontinued Debian flavor. It used the FreeBSD kernel and GNU userland. The majority of software in Debian GNU/kFreeBSD was built from the same sources as Debian, with some kernel packages from FreeBSD. The k in kFreeBSD is an abbreviation for kernel, which refers to the FreeBSD kernel. Before discontinuing the project, Debian maintained i386 and amd64 ports. The last version of Debian kFreeBSD was Debian 8 (Jessie) RC3. Debian GNU/kFreeBSD was created in 2002. It was included in Debian 6.0 (Squeeze) as a technology preview, and in Debian 7.0 (Wheezy) as an official port. Debian GNU/kFreeBSD was discontinued as an officially supported platform as of Debian 8.0. Debian developers cited OSS, pf, jails, NDIS, and ZFS as reasons for being interested in the FreeBSD kernel. It has not been updated since Debian 8. From July 2019, the operating system continues to be maintained unofficially.As of July 2023, the development of Debian GNU/kFreeBSD has officially terminated due to the lack of interest and developers. Debian GNU/Hurd is a flavor based on the Hurd microkernel, instead of Linux. Debian GNU/Hurd has been in development since 1998, and made a formal release in May 2013, with 78% of the software packaged for Debian GNU/Linux ported to the GNU Hurd. Hurd is not yet an official Debian release, and is maintained and developed as an unofficial port. Debian GNU/Hurd is distributed as an installer CD (running the official Debian installer) or ready-to-run virtual disk image (Live CD, Live USB). The CD uses the IA-32 architecture, making it compatible with IA-32 and x86-64 PCs. The current version of Debian GNU/Hurd is 2023, published in June 2023.
'''GNU''' is the name of a [[computer]] [[operating system]]. The name is short for '''''G'''NU's '''N'''ot '''U'''nix''. [[Richard Stallman]] leads the GNU Project. The popular [[Linux]] operating systems made using the [[Linux kernel]] have many GNU tools too. So, many projects and [[Software developer|developers]] call the Linux-based operating systems ''[[GNU/Linux]].'' The GNU project was started by Richard Stallman in [[1983]]. He wanted to create a computer system that was all [[free and open-source software]]. Users could change, share and publish new work based on GNU. He and a group of developers started by creating copies of each piece of [[UNIX]] software. The rest of GNU was the [[Kernel (computer science)|kernel]], called the [[GNU Hurd]], which is not yet finished. The more popular [[Linux kernel]] is often used instead. == Related pages == *[[GNU Compiler Collection]] == Other websites == * {{Official website}} * [http://www.gnu.org/gnu/gnu-user-groups.html GNU User Groups] {{Tech-stub}} [[Category:Operating systems]] [[Category:GNU project]]
GNU (/ɡnuː/ ) is an extensive collection of free software (385 packages as of September 2023), which can be used as an operating system or can be used in parts with other operating systems. The use of the completed GNU tools led to the family of operating systems popularly known as Linux. Most of GNU is licensed under the GNU Project's own General Public License (GPL). GNU is also the project within which the free software concept originated. Richard Stallman, the founder of the project, views GNU as a "technical means to a social end". Relatedly, Lawrence Lessig states in his introduction to the second edition of Stallman's book Free Software, Free Society that in it Stallman has written about "the social aspects of software and how Free Software can create community and social justice". GNU is a recursive acronym for "GNU's Not Unix!", chosen because GNU's design is Unix-like, but differs from Unix by being free software and containing no Unix code. Stallman chose the name by using various plays on words, including the song The Gnu. Development of the GNU operating system was initiated by Richard Stallman while he worked at MIT Artificial Intelligence Laboratory. It was called the GNU Project, and was publicly announced on September 27, 1983, on the net.unix-wizards and net.usoft newsgroups by Stallman. Software development began on January 5, 1984, when Stallman quit his job at the Lab so that they could not claim ownership or interfere with distributing GNU components as free software. The goal was to bring a completely free software operating system into existence. Stallman wanted computer users to be free to study the source code of the software they use, share software with other people, modify the behavior of software, and publish their modified versions of the software. This philosophy was published as the GNU Manifesto in March 1985. Richard Stallman's experience with the Incompatible Timesharing System (ITS), an early operating system written in assembly language that became obsolete due to discontinuation of PDP-10, the computer architecture for which ITS was written, led to a decision that a portable system was necessary. It was thus decided that the development would be started using C and Lisp as system programming languages, and that GNU would be compatible with Unix. At the time, Unix was already a popular proprietary operating system. The design of Unix was modular, so it could be reimplemented piece by piece. Much of the needed software had to be written from scratch, but existing compatible third-party free software components were also used such as the TeX typesetting system, the X Window System, and the Mach microkernel that forms the basis of the GNU Mach core of GNU Hurd (the official kernel of GNU). With the exception of the aforementioned third-party components, most of GNU has been written by volunteers; some in their spare time, some paid by companies, educational institutions, and other non-profit organizations. In October 1985, Stallman set up the Free Software Foundation (FSF). In the late 1980s and 1990s, the FSF hired software developers to write the software needed for GNU. As GNU gained prominence, interested businesses began contributing to development or selling GNU software and technical support. The most prominent and successful of these was Cygnus Solutions, now part of Red Hat. The system's basic components include the GNU Compiler Collection (GCC), the GNU C library (glibc), and GNU Core Utilities (coreutils), but also the GNU Debugger (GDB), GNU Binary Utilities (binutils), and the GNU Bash shell. GNU developers have contributed to Linux ports of GNU applications and utilities, which are now also widely used on other operating systems such as BSD variants, Solaris and macOS. Many GNU programs have been ported to other operating systems, including proprietary platforms such as Microsoft Windows and macOS. GNU programs have been shown to be more reliable than their proprietary Unix counterparts. As of January 2022, there are a total of 459 GNU packages (including decommissioned, 383 excluding) hosted on the official GNU development site. In its original meaning, and one still common in hardware engineering, the operating system is a basic set of functions to control the hardware and manage things like task scheduling and system calls. In modern terminology used by software developers, the collection of these functions is usually referred to as a kernel, while an 'operating system' is expected to have a more extensive set of programmes. The GNU project maintains two kernels itself, allowing the creation of pure GNU operating systems, but the GNU toolchain is also used with non-GNU kernels. Due to the two different definitions of the term 'operating system', there is an ongoing debate concerning the naming of distributions of GNU packages with a non-GNU kernel. (See below.) The original kernel of GNU Project is the GNU Hurd microkernel, which was the original focus of the Free Software Foundation (FSF). With the April 30, 2015 release of the Debian GNU/Hurd 2015 distro, GNU now provides all required components to assemble an operating system that users can install and use on a computer. However, the Hurd kernel is not yet considered production-ready but rather a base for further development and non-critical application usage. As of 2012, a fork of the Linux kernel became officially part of the GNU Project in the form of Linux-libre, a variant of Linux with all proprietary components removed. The GNU Project has endorsed Linux-libre distributions, such as Trisquel, Parabola GNU/Linux-libre, PureOS and GNU Guix System. Because of the development status of Hurd, GNU is usually paired with other kernels such as Linux or FreeBSD. Whether the combination of GNU libraries with external kernels is a GNU operating system with a kernel (e.g. GNU with Linux), because the GNU collection renders the kernel into a usable operating system as understood in modern software development, or whether the kernel is an operating system unto itself with a GNU layer on top (i.e. Linux with GNU), because the kernel can operate a machine without GNU, is a matter of ongoing debate. The FSF maintains that an operating system built using the Linux kernel and GNU tools and utilities should be considered a variant of GNU, and promotes the term GNU/Linux for such systems (leading to the GNU/Linux naming controversy). This view is not exclusive to the FSF. Notably, Debian, one of the biggest and oldest Linux distributions, refers to itself as Debian GNU/Linux. The GNU Project recommends that contributors assign the copyright for GNU packages to the Free Software Foundation, though the Free Software Foundation considers it acceptable to release small changes to an existing project to the public domain. However, this is not required; package maintainers may retain copyright to the GNU packages they maintain, though since only the copyright holder may enforce the license used (such as the GNU GPL), the copyright holder in this case enforces it rather than the Free Software Foundation. For the development of needed software, Stallman wrote a license called the GNU General Public License (first called Emacs General Public License), with the goal to guarantee users freedom to share and change free software. Stallman wrote this license after his experience with James Gosling and a program called UniPress, over a controversy around software code use in the GNU Emacs program. For most of the 80s, each GNU package had its own license: the Emacs General Public License, the GCC General Public License, etc. In 1989, FSF published a single license they could use for all their software, and which could be used by non-GNU projects: the GNU General Public License (GPL). This license is now used by most of GNU software, as well as a large number of free software programs that are not part of the GNU Project; it also historically has been the most commonly used free software license (though recently challenged by the MIT license). It gives all recipients of a program the right to run, copy, modify and distribute it, while forbidding them from imposing further restrictions on any copies they distribute. This idea is often referred to as copyleft. In 1991, the GNU Lesser General Public License (LGPL), then known as the Library General Public License, was written for the GNU C Library to allow it to be linked with proprietary software. 1991 also saw the release of version 2 of the GNU GPL. The GNU Free Documentation License (FDL), for documentation, followed in 2000. The GPL and LGPL were revised to version 3 in 2007, adding clauses to protect users against hardware restrictions that prevent users from running modified software on their own devices. Besides GNU's packages, the GNU Project's licenses can and are used by many unrelated projects, such as the Linux kernel, often used with GNU software. A majority of free software such as the X Window System, is licensed under permissive free software licenses. The logo for GNU is a gnu head. Originally drawn by Etienne Suvasa, a bolder and simpler version designed by Aurelio Heckert is now preferred. It appears in GNU software and in printed and electronic documentation for the GNU Project, and is also used in Free Software Foundation materials. There was also a modified version of the official logo. It was created by the Free Software Foundation in September 2013 in order to commemorate the 30th anniversary of the GNU Project.
{{hatnote|"Republika Srbija" redirects here, it’s not to be confused with the Serbian nation of [[Republika Srpska]] in [[Bosnia and Herzegovina]].}} {{Use dmy dates|date=February 2018}} {{Coord|44|N|21|E|display=title}} {{Infobox country | conventional_long_name = Republic of Serbia | common_name = Serbia | native_name = Република Србија ([[Serbian Cyrillic]]) Republika Srbija ([[Serbian Latin]]) | image_flag = Flag of Serbia.svg | image_coat = Coat of arms of Serbia.svg | national_motto = | national_anthem = <br/>"[[Bože pravde|Боже правде]]" / "''Bože pravde''"<br />({{Lang-en|"God of Justice"}})<br/><center>[[File:National anthem of Serbia, performed by the United States Navy Band.wav]]</center> | image_map = File:Locator map of Serbia in Europe and the world.png | map_caption = Location of Serbia in (green) and Europe in (dark grey). | image_map2 = | capital = [[Belgrade]] | coordinates = {{Coord|44|48|N|20|28|E|type:city}} | largest_city = capital | official_languages = [[Serbian language|Serbian]] | languages_type = Official minority languages | languages = * [[Montenegrin language|Montenegrin]] * [[Bosnian language|Bosnian]] * [[Croatian language|Croatian]] * [[Albanian language|Albanian]] * [[Romani language|Romani]] * [[Hungarian language|Hungarian]] | ethnic_groups = {{unbulleted list | 83.3% [[Serbs]] | 3.5% [[Hungarians in Serbia|Hungarians]] | 2.1% [[Roma in Serbia|Roma]] | 2% [[Bosniaks of Serbia|Bosniaks]] | {{nowrap|9% others}} | {{small|(excluding [[Kosovo]])}} }} | ethnic_groups_year = 2011 | demonym = [[Serbians|Serbian]] | government_type = {{nowrap|[[Unitary state|Unitary]] [[Parliamentary system|parliamentary]]<br />constitutional republic}} | leader_title1 = [[President of Serbia|President]] | leader_name1 = [[Aleksandar Vučić]] | leader_title2 = [[Prime Minister of Serbia|Prime Minister]] | leader_name2 = [[Miloš Vučević]] | legislature = [[National Assembly of Serbia|National Assembly]] | sovereignty_type = [[History of Serbia|Formation]] | established_event1 = {{nowrap|[[Principality of Serbia (medieval)|Medieval principality]]}} | established_date1 = late 8th century | established_event2 = {{nowrap|[[Kingdom of Serbia (medieval)|Medieval kingdom]]}}/{{nowrap|[[Serbian Empire|empire]]}} | established_date2 = 1217/1346 | established_event3 = {{nowrap|[[Serbian Despotate|Ottoman conquest]]}}<sup>a</sup> | established_date3 = 1459–1556 | established_event4 = {{nowrap|[[Principality of Serbia]]}} | established_date4 = 1815 | established_event5 = {{nowrap|[[Treaty of Berlin (1878)|Internationally recognized]]}} | established_date5 = 1878 | established_event6 = {{nowrap|[[Banat, Backa and Baranja|National unification]]}} | established_date6 = 1912–1918 | established_event7 = {{nowrap|Independent republic}} | established_date7 = 5 June 2006 | area_km2 = 88361 | area_label = Including Kosovo | area_rank = 111th | area_sq_mi = <!-- 34,116--> | area_label2 = Excluding Kosovo | area_data2 = {{convert|77474|km2|abbr=on}}<ref name=cia_profile>{{cite web |url=https://www.cia.gov/library/publications/the-world-factbook/geos/ri.html |title=The World Factbook: Serbia |publisher=[[Central Intelligence Agency]] |date=20 June 2014 |accessdate=18 December 2014 |archive-date=24 December 2018 |archive-url=https://web.archive.org/web/20181224211224/https://www.cia.gov/library/publications/the-world-factbook/geos/ri.html |url-status=dead }}</ref> <!-- | percent_water = 0.13 {{small|(including Kosovo)}} -->| population_estimate = 7,040,272 {{small|(excluding Kosovo)}} {{decrease}}<ref name="stat.gov.rs">{{cite web|url=http://www.stat.gov.rs/|title=PBC stats|date=2018|website=stat.gov.rs|access-date=2018-08-17|archive-date=2019-09-04|archive-url=https://web.archive.org/web/20190904082621/http://www.stat.gov.rs/|url-status=live}}</ref> | population_estimate_year = 2017 | population_estimate_rank = 104th | population_density_km2 = 91.1 | population_density_sq_mi = 211 <!--Do not remove per [[WP:MOSNUM]]--> | population_density_rank = 121th | GDP_PPP = $112.475&nbsp;billion<ref name="imf2"/> | GDP_PPP_year = 2018 | GDP_PPP_rank = 78th | GDP_PPP_per_capita = $16,063 {{small|(excluding Kosovo)}}{{lower|0.2em|<ref name="imf2">{{cite web|url=http://www.imf.org/external/pubs/ft/weo/2017/02/weodata/weorept.aspx?sy=2017&ey=2019&scsm=1&ssd=1&sort=country&ds=.&br=1&pr1.x=91&pr1.y=16&c=942&s=NGDPD%2CPPPGDP%2CNGDPDPC%2CPPPPC&grp=0&a=|title=Serbia|publisher=[[International Monetary Fund]]|accessdate=23 January 2018|archive-date=14 February 2018|archive-url=https://web.archive.org/web/20180214070415/http://www.imf.org/external/pubs/ft/weo/2017/02/weodata/weorept.aspx?sy=2017&ey=2019&scsm=1&ssd=1&sort=country&ds=.&br=1&pr1.x=91&pr1.y=16&c=942&s=NGDPD%2CPPPGDP%2CNGDPDPC%2CPPPPC&grp=0&a=|url-status=live}}</ref>}} | GDP_PPP_per_capita_rank = 83rd | GDP_nominal = $42.378&nbsp;billion<ref name=imf2/> | GDP_nominal_year = 2018 | GDP_nominal_rank = 86th | GDP_nominal_per_capita = $6,052 {{small|(excluding Kosovo)}}{{lower|0.2em|<ref name=imf2/>}} | GDP_nominal_per_capita_rank = 88th | Gini = 29.6 <!--number only--> | Gini_year = 2013 | Gini_change = <!--increase/decrease/steady--> | Gini_ref = <ref>{{cite web |title=Human Development Reports: Gini coefficient |url=http://hdr.undp.org/en/content/income-gini-coefficient |website=hdr.undp.org |publisher=United Nations Development Programme |accessdate=20 January 2018 |archive-date=10 June 2010 |archive-url=https://web.archive.org/web/20100610232357/https://www.cia.gov/library/publications/the-world-factbook/fields/2172.html |url-status=dead }}</ref> | HDI = 0.776 <!--number only--> | HDI_year = 2015<!-- Please use the year to which the data refers, not the publication year--> | HDI_change = increase<!--increase/decrease/steady--> | HDI_ref = <ref name="HDI">{{cite web |url=http://hdr.undp.org/sites/default/files/2016_human_development_report.pdf |title=2016 Human Development Report |year=2016 |accessdate=25 March 2017 |publisher=United Nations Development Programme |archive-date=22 March 2017 |archive-url=https://web.archive.org/web/20170322153238/http://hdr.undp.org/sites/default/files/2016_human_development_report.pdf |url-status=live }}</ref> | HDI_rank = 66th | currency = [[Serbian dinar]] | currency_code = RSD | time_zone = [[Central European Time|CET]] | utc_offset = [[UTC+1|+1]] | utc_offset_DST = [[UTC+2|+2]] | time_zone_DST = [[Central European Summer Time|CEST]] | drives_on = [[Right- and left-hand traffic|right]] | calling_code = [[Telephone numbers in Serbia|+381]] | cctld = {{unbulleted list |[[.rs]] |[[.срб]]}} | footnote_a = From the [[Serbian Despotate|fall of Smederevo]] until conquest of [[Belgrade]], [[Macva|Mačva]] and [[Vojvodina]] }} '''Serbia''' (Serbian: ''Република Србија'' / Republika Srbija), officially the '''Republic of Serbia''' (Serbian: ''Србија'' / Srbija), is a country in [[Southeast Europe|Southeastern Europe]], and a part of the [[Balkans]]. The [[capital city]] is [[Belgrade]]. To the north of Serbia is the country [[Hungary]]. To the east of Serbia are the countries [[Bulgaria]] and [[Romania]]. To the south of Serbia is the country [[Republic of Macedonia|North Macedonia]] and [[Kosovo]]. To the west of Serbia are the countries [[Montenegro]] (the country that was once united with Serbia), [[Bosnia and Herzegovina]] and [[Croatia]]. == Politics == Serbia is trying to join the [[European Union]]. According to the [[European Commission]] report in 2023 it is doing too little to settle its differences with [[Kosovo]]. Its [[foreign policy]] is too close to [[Russia]]. There is too much [[corruption]] and [[Organized crime|organised crime]]. Media independence is weak. [[Aleksandar Vulin]], the head of the [[state security]] agency, has been [[Sanction|sanctioned]] by the USA for his links to crime, Russia and “promoting ethno-nationalist narratives that fuel instability in Serbia and the region”.<ref>{{Cite news|last=Barber|first=Tony|date=2023-11-12|title=Serbia is a poor fit for EU enlargement plans|work=Financial Times|url=https://www.ft.com/content/548c6093-2853-4e9a-b39d-1580a6190403|access-date=2023-11-12}}</ref> ==History== An ancient civilization which inhabited Serbia is called the Vinča culture and used symbols that are called the Vinča script<ref>{{cite web|url=http://www.korenine.si/zborniki/zbornik01/htm/pesic_vinca.htm|title=THE VINCHA SCRIPT|website=www.korenine.si|access-date=2016-11-04|archive-date=2015-01-10|archive-url=https://web.archive.org/web/20150110210152/http://www.korenine.si/zborniki/zbornik01/htm/pesic_vinca.htm|url-status=live}}</ref> Some people think they are the first [[writing system]]. Serbia is also home to the earliest known copper smelting sites and the birth of the [[Copper Age]] when people left the Stone Age and went into regular use of [[metallurgy]].<ref>http://m.megalithic.co.uk/article.php?sid=2146413558</ref> The same culture contains the earliest evidence of tin alloy [[bronze]] which replaced the much weaker arsenic bronze.<ref>{{cite web|url=http://haemus.org.mk/6500-year-old-tin-bronze-from-serbia/|title=6,500-year old tin-bronze from Serbia - HAEMUS - Center for scientific research and promotion of culture|publisher=}}</ref> === Medieval history === White Serbs, an early [[Slavs|Slavic]] tribe from Northern Europe, came to Northern Greece in the 6th century. By the 8th century they had created the Serbian Principality, a Serbian country, in the Balkans.<ref>Sima M. Ćirković, SRBI MEĐU EUROPSKIM NARODIMA,(Serbs) 2008. http://www.mo-vrebac-pavlovac.hr/attachments/article/451/Sima%20%C4%86irkovi%C4%87%20SRBI%20ME%C4%90U%20EVROPSKIM%20NARODIMA.pdf {{Webarchive|url=https://web.archive.org/web/20200108093618/http://www.mo-vrebac-pavlovac.hr/attachments/article/451/Sima%20%C4%86irkovi%C4%87%20SRBI%20ME%C4%90U%20EVROPSKIM%20NARODIMA.pdf |date=2020-01-08 }} #page=26-27</ref> The Serbs became Christian around the 10th century. For 200 years, the Nemanjić dynasty ruled. They made Serbia a kingdom, built new towns, monasteries, and forts, and made Serbia bigger. In 1371 the Nemanjic Dynasty died out. Serbia became unsafe and local leaders fought each other for control. After the [[Battle of Kosovo]] 1389 the [[Ottoman Empire]] fought the Serbs for 70 years until in 1459 the Ottoman Turks conquered Serbia<ref>{{cite book|url=https://books.google.com/books?id=LvVbRrH1QBgC|title=The Late Medieval Balkans: A Critical Survey from the Late Twelfth Century to the Ottoman Conquest|last1=Fine|first1=John|date=1994|publisher=University of Michigan Press|isbn=9780472082605|page=575|access-date=2022-02-24|archive-url=https://web.archive.org/web/20230115194436/https://books.google.com/books?id=LvVbRrH1QBgC|archive-date=2023-01-15|url-status=live}}</ref> and other Balkan countries. === Ottoman period === Once the Ottomans conquered Serbia, they got rid of the Serbian upper class. Most Serbs worked as farmers on land owned by Turks. They had to pay high taxes to the Turks. Some Serbs became Muslims. Serbs had their own laws in the Ottoman Empire through the [[millet system]].{{sfn|Runciman|1968|p=204}}{{sfn|Kia|2011|p=115}} The Christian Austrian Empire fought the Turks many times, including in the [[Great Turkish War]] of the late 17th Century. In 1718 Austria took northern Serbia and some other places. The last Austro-Turkish war was in 1781-1791. During the 19th century Serbia gradually became independent. During World War I Austria conquered Serbia but lost the bigger war. Serbia joined other Balkan countries to form [[Yugoslavia]]. ==Geography== Serbia is found in the [[Balkans|Balkan peninsula]] and the [[Pannonian Plain]]. The [[Danube]] passes through Serbia. Over 31% of Serbia is covered by forest.<ref>{{cite web|url=http://www.balkans.com/sr/open-news.php?uniquenumber=53766|title=Dragin obrazlozio predloge zakona u oblasti poljoprivrede|year=2005|publisher=Vlada Srbije|access-date=2012-09-30|archive-date=2012-10-26|archive-url=https://web.archive.org/web/20121026061749/http://www.balkans.com/sr/open-news.php?uniquenumber=53766|url-status=dead}}</ref> [[National park]]s take up 10% of the country's territory.<ref>{{cite web|url=http://www.poslovnimagazin.biz/magazin/privreda/u-srbiji-do-2010-godine-10-teritorije-nacionalni-parkovi-30-377|title=U Srbiji do 2010. godine 10% teritorije nacionalni parkovi|publisher=Poslovni Magazin – Business Surfer|accessdate=28 April 2010|archive-date=22 June 2008|archive-url=https://web.archive.org/web/20080622122642/http://www.poslovnimagazin.biz/magazin/privreda/u-srbiji-do-2010-godine-10-teritorije-nacionalni-parkovi-30-377|url-status=dead}}</ref> Serbia has 5 [[national parks]] and 22 [[nature reserves]]. == Birds == Serbia is on the Vardar-Morava Flyway of bird migration, so many birds fly across Serbia when going to Africa and back. Golden Eagle, Nutcracker and Coal Tit live high in the mountains. Nightingale, Chaffinch and Greater Spotted Woodpecker live in the forests in the hills. Saker Falcon, Whinchat and Quail live in steppe and fields in the northern lowland part of Serbia which is called Vojvodina. Syrian Woodpecker, Common Redstart and Collared Dove live in the orchards and farmyards. Black Redstart, Kestrel and Common Swift live in the city center. The Hooded Crow lives everywhere. The town of Kikinda is known for Long-eared Owls that form flocks in trees at the city squares. Tourists come from other countries to see them. In winter you can see Pygmy Cormorants in Belgrade, at the rivers Sava and Danube. == Economy == Serbia is classed as an upper-middle income economy.<ref>{{cite web|url=http://web.worldbank.org/WBSITE/EXTERNAL/DATASTATISTICS/0,,contentMDK:20421402~pagePK:64133150~piPK:64133175~theSitePK:239419,00.html#Upper_middle_income|title=Upper-middle-income economies|publisher=The World Bank|access-date=2012-09-30|archive-date=2008-05-24|archive-url=https://web.archive.org/web/20080524215837/http://web.worldbank.org/WBSITE/EXTERNAL/DATASTATISTICS/0%2C%2CcontentMDK%3A20421402~pagePK%3A64133150~piPK%3A64133175~theSitePK%3A239419%2C00.html#Upper_middle_income|url-status=live}}</ref> The major processed vegetable crops in Serbia are [[potato]]es, [[tomato]]es and [[Black pepper|pepper]].<ref>{{citation|author=Marat Terterov|title=Doing Business with Serbia|year=2006|url=https://books.google.com/books?id=dduTMrBQxBsC&pg=PA169|page=169|isbn=9781905050680}}</ref> Serbia is one of the biggest world producers and exporters of [[raspberry|raspberries]].<ref>{{cite web|url=https://www.balcanicaucaso.org/eng/Areas/Serbia/Raspberries-Serbia-s-Red-Gold-83572|title=Raspberries, Serbia's Red Gold|first=Osservatorio Balcani e|last=Caucaso|website=Osservatorio Balcani e Caucaso|access-date=2019-02-11|archive-date=2020-11-12|archive-url=https://web.archive.org/web/20201112013431/https://www.balcanicaucaso.org/eng/Areas/Serbia/Raspberries-Serbia-s-Red-Gold-83572|url-status=live}}</ref> They are a leading exporter of frozen fruit.<ref>{{cite web|url=http://www.invest-in-serbia.com/modules.php?name=Content&pa=showpage&pid=66|title=Rebranding Serbia|author=Borka Tomic|date=13 April 2006|publisher=Invest in Serbia|access-date=30 September 2012|archive-date=13 May 2011|archive-url=https://web.archive.org/web/20110513161540/http://www.invest-in-serbia.com/modules.php?name=Content&pa=showpage&pid=66|url-status=dead}}</ref> There are four international airports in Serbia: [[Belgrade Nikola Tesla Airport]], [[Niš Constantine the Great Airport]], [[Morava International Airport]], and [[Vršac International Airport]]. 89% of households in Serbia have fixed telephone lines. There are over 9.60&nbsp;million cell-phones users. This is larger than the number of the total population of Serbia itself by 30%. ==Culture== Serbia has a total of eight sites on the [[World Heritage Site|UNESCO World Heritage list]]: The Early Medieval capital [[Stari Ras]] and the 13th-century monastery [[Sopoćani]], and the 12th-century monastery [[Studenica monastery|Studenica]], and the endangered [[Medieval Monuments in Kosovo]] group, comprising the monasteries of [[Visoki Dečani]], [[Our Lady of Ljeviš]], [[Gračanica Monastery|Gračanica]] and [[Patriarchate of Peć]] (former seat of the Serbian Church, mausoleum of Serbian royalty) and finally the Roman estate of [[Gamzigrad|Gamzigrad–Felix Romuliana]]. There are two literary memorials on UNESCO's [[Memory of the World Programme]]: The 12th-century [[Miroslav Gospel]], and scientist [[Nikola Tesla]]'s valuable archive. The most prominent museum in Serbia is the [[National Museum of Serbia]]. It was founded in 1844. It houses a collection of more than 400,000 exhibits, over 5,600 paintings and 8,400 drawings and prints, and includes many foreign masterpiece collections, including Miroslav Gospel. The official language, [[Serbian language|Serbian]], is written in both the [[Cyrillic]] and [[Latin alphabet]]s. Composer and [[musicology|musicologist]] [[Stevan Stojanović Mokranjac]] is said to be one of the most important founders of modern Serbian music.<ref>{{cite web |url=http://www.rastko.rs/isk/isk_17.html |title=Projekat Rastko: Istorija srpske kulture |publisher=Rastko.rs |accessdate=24 May 2012 |archive-date=3 June 2012 |archive-url=https://web.archive.org/web/20120603124009/http://www.rastko.rs/isk/isk_17.html |url-status=live }}</ref><ref>{{cite web |url=http://www.riznicasrpska.net/muzika/index.php?topic=60.0 |title=Stevan Stojanović Mokranjac (1856—1914) |publisher=Riznicasrpska.net |date=28 September 1914 |accessdate=24 May 2012 |archive-date=26 September 2013 |archive-url=https://web.archive.org/web/20130926231446/http://www.riznicasrpska.net/muzika/index.php?topic=60.0 |url-status=dead }}</ref> In the 1990s and the 2000s, many [[pop music]] performers rose to fame. Željko Joksimović won second place at the 2004 [[Eurovision Song Contest]]. Marija Šerifović won the 2007 Eurovision Song Contest. Serbia was the host of the [[Eurovision Song Contest 2008]]. ===Sports=== The most popular sports in Serbia are [[association football|football]], basketball, volleyball, [[team handball|handball]], [[water polo]] and [[tennis]]. The three main football clubs in Serbia are [[Red Star Belgrade|Red Star]] and [[FK Partizan|Partizan]], both from the capital city of [[Belgrade]], and [[FK Vojvodina|Vojvodina]] from [[Novi Sad]]. [[Novak Djokovic]], a multiple Grand Slam-winning [[tennis]] player and current number one, is from Serbia. Other tennis players from Serbia include [[Ana Ivanovic]] and [[Jelena Jankovic]]. ==Related pages== *[[List of rivers of Serbia]] *[[Serbia at the Olympics]] *[[Serbia national football team]] == References == {{commonscat}} {{Wikivoyage}} {{Reflist}} * {{Cite book|last=Runciman|first=Steven|authorlink=Steven Runciman|title=The Great Church in Captivity: A Study of the Patriarchate of Constantinople from the Eve of the Turkish Conquest to the Greek War of Independence|location=Cambridge|publisher=Cambridge University Press|year=1968|isbn=9780521071888|edition=1.|url=https://books.google.com/books?id=WxsrAAAAIAAJ}} * {{Cite book|last=Kia|first=Mehrdad|title=Daily Life in the Ottoman Empire|year=2011|location=Santa Barbara, California|publisher=Greenwood Press|isbn=9780313064029|url=https://books.google.com/books?id=aQKRtpZmX0EC}} {{Europe}} {{European Union}} {{authority control}} [[Category:Serbia| ]] [[Category:European Union candidate states]] [[Category:2006 establishments in Europe]]
Serbia, officially the Republic of Serbia, is a landlocked country at the crossroads of Southeast and Central Europe, located in the Balkans and the Pannonian Plain. It shares land borders with Hungary to the north, Romania to the northeast, Bulgaria to the southeast, North Macedonia to the south, Croatia and Bosnia and Herzegovina to the west, and Montenegro to the southwest. Serbia claims a border with Albania through the disputed territory of Kosovo. Serbia has about 6.6 million inhabitants, excluding Kosovo. Its capital Belgrade is also the largest city. Continuously inhabited since the Paleolithic Age, the territory of modern-day Serbia faced Slavic migrations in the 6th century, establishing several regional states in the early Middle Ages at times recognised as tributaries to the Byzantine, Frankish and Hungarian kingdoms. The Serbian Kingdom obtained recognition by the Holy See and Constantinople in 1217, reaching its territorial apex in 1346 as the Serbian Empire. By the mid-16th century, the Ottomans annexed the entirety of modern-day Serbia; their rule was at times interrupted by the Habsburg Empire, which began expanding towards Central Serbia from the end of the 17th century while maintaining a foothold in Vojvodina. In the early 19th century, the Serbian Revolution established the nation-state as the region's first constitutional monarchy, which subsequently expanded its territory. In 1918, in the aftermath of World War I, the Kingdom of Serbia united with the former Habsburg crownland of Vojvodina; later in the same year it joined with other South Slavic nations in the foundation of Yugoslavia, which existed in various political formations until the Yugoslav Wars of the 1990s. During the breakup of Yugoslavia, Serbia formed a union with Montenegro, which was peacefully dissolved in 2006, restoring Serbia's independence as a sovereign state for the first time since 1918. In 2008, representatives of the Assembly of Kosovo unilaterally declared independence, with mixed responses from the international community while Serbia continues to claim it as part of its own sovereign territory. Serbia is an upper-middle income economy, ranked "very high" in the Human Development Index domain. It is a unitary parliamentary constitutional republic, member of the UN, CoE, OSCE, PfP, BSEC, CEFTA, and is acceding to the WTO. Since 2014, the country has been negotiating its EU accession, with the possibility of joining the European Union by 2030. Serbia formally adheres to the policy of military neutrality. The country provides universal health care and free primary and secondary education to its citizens. The origin of the name Serbia is unclear. Historically, authors have mentioned the Serbs (Serbian: Srbi / Срби) and the Sorbs of Eastern Germany (Upper Sorbian: Serbja; Lower Sorbian: Serby) in a variety of ways: Cervetiis (Servetiis), gentis (S)urbiorum, Suurbi, Sorabi, Soraborum, Sorabos, Surpe, Sorabici, Sorabiet, Sarbin, Swrbjn, Servians, Sorbi, Sirbia, Sribia, Zirbia, Zribia, Suurbelant, Surbia, Serbulia / Sorbulia among others. These authors used these names to refer to Serbs and Sorbs in areas where their historical and current presence is not disputable (notably in the Balkans and Lusatia). However, there are also sources that mention the same or similar names in other parts of the World (most notably in the Asiatic Sarmatia in the Caucasus). There exist two prevailing theories on the origin of the ethnonym *Sŕbъ (plur. *Sŕby), one from a Proto-Slavic language with an appellative meaning of a "family kinship" and "alliance", while another from an Iranian-Sarmatian language with various meanings. In his work, De Administrando Imperio, Constantine VII Porphyrogenitus suggests that the Serbs originated from White Serbia near Francia. According to the recorded tradition the White Serbs split in two, with the half that became known as the Serbs coming down to settle Byzantine land. From 1815 to 1882, the official name for Serbia was the Principality of Serbia. From 1882 to 1918, it was renamed to the Kingdom of Serbia, later from 1945 to 1963, the official name for Serbia was the People's Republic of Serbia. This was again renamed the Socialist Republic of Serbia from 1963 to 1990. Since 1990, the official name of the country has been the Republic of Serbia. Archaeological evidence of Paleolithic settlements on the territory of present-day Serbia is scarce. A fragment of a human jaw was found in Sićevo (Mala Balanica) and is believed to be up to 525,000–397,000 years old. Approximately around 6,500 years BC, during the Neolithic, the Starčevo and Vinča cultures existed in the region of modern-day Belgrade. They dominated much of Southeast Europe (as well as parts of Central Europe and Anatolia). Several important archaeological sites from this era, including Lepenski Vir and Vinča-Belo Brdo, still exist near the banks of the Danube. During the Iron Age, local tribes of Triballi, Dardani, and Autariatae were encountered by the Ancient Greeks during their cultural and political expansion into the region, from the 5th up to the 2nd century BC. The Celtic tribe of Scordisci settled throughout the area in the 3rd century BC. It formed a tribal state, building several fortifications, including their capital at Singidunum (present-day Belgrade) and Naissos (present-day Niš). The Romans conquered much of the territory in the 2nd century BC. In 167 BC, the Roman province of Illyricum was established; the remainder was conquered around 75 BC, forming the Roman province of Moesia Superior; the modern-day Srem region was conquered in 9 BC; and Bačka and Banat in 106 AD after the Dacian Wars. As a result of this, contemporary Serbia extends fully or partially over several former Roman provinces, including Moesia, Pannonia, Praevalitana, Dalmatia, Dacia, and Macedonia. The chief towns of Upper Moesia (and broader) were: Singidunum (Belgrade), Viminacium (now Old Kostolac), Remesiana (now Bela Palanka), Naissos (Niš), and Sirmium (now Sremska Mitrovica), the latter of which served as a Roman capital during the Tetrarchy. Seventeen Roman Emperors were born in the area of modern-day Serbia, second only to contemporary Italy. The most famous of these was Constantine the Great, the first Christian Emperor, who issued an edict ordering religious tolerance throughout the Empire. When the Roman Empire was divided in 395, most of Serbia remained under the Byzantine Empire. At the same time, its northwestern parts were included in the Western Roman Empire. By the 6th century, South Slavs migrated into the European provinces of the Byzantine Empire in large numbers. They merged with the local Romanised population that was gradually assimilated. White Serbs, an early Slavic tribe from White Serbia eventually settled in an area between the Sava river and the Dinaric Alps. By the beginning of the 9th century, Serbia achieved a level of statehood. Christianization of Serbia was a gradual process, finalized by the middle of the 9th century. In the mid-10th-century, the Serbian state stretched between the Adriatic Sea, the Neretva, the Sava, the Morava, and Skadar. During the 11th and 12th century, Serbian state frequently fought with the neighbouring Byzantine Empire. Between 1166 and 1371, Serbia was ruled by the Nemanjić dynasty (whose legacy is especially cherished), under whom the state was elevated to a kingdom in 1217, and an empire in 1346, under Stefan Dušan. Serbian Orthodox Church was organized as an autocephalous archbishopric in 1219, through the effort of Sava, the country's patron saint, and in 1346 it was raised to the Patriarchate. Monuments of the Nemanjić period survive in many monasteries (several being World Heritage sites) and fortifications. During these centuries the Serbian state (and influence) expanded significantly. The northern part (modern Vojvodina), was ruled by the Kingdom of Hungary. The period after 1371, known as the Fall of the Serbian Empire saw the once-powerful state fragmented into several principalities, culminating in the Battle of Kosovo (1389) against the rising Ottoman Empire. By the end of the 14th century, the Turks had conquered and ruled the territories south of the Šar Mountains. At the same time, the political center of Serbia shifted northwards, when the capital of the newly established Serbian Despotate was transferred to Belgrade in 1403, before moving to Smederevo in 1430. The Despotate was then under the double vassalage of Hungary and the Ottoman Empire. The fall of Smederevo on June 20, 1459, which marked the full conquest of the Serbian Despotate by the Ottomans, also symbolically signified the end of the Serbian state. In all Serbian lands conquered by the Ottomans, the native nobility was eliminated and the peasantry was enserfed to Ottoman rulers, while much of the clergy fled or were confined to the isolated monasteries. Under the Ottoman system, Serbs, as well as Christians, were considered an inferior class of people and subjected to heavy taxes, and a portion of the Serbian population experienced Islamization. Many Serbs were recruited during the devshirme system, a form of slavery in the Ottoman Empire, in which boys from Balkan Christian families were forcibly converted to Islam and trained for infantry units of the Ottoman army known as the Janissaries. The Serbian Patriarchate of Peć was extinguished in 1463, but reestablished in 1557, providing for limited continuation of Serbian cultural traditions within the Ottoman Empire, under the Millet system. After the loss of statehood to the Ottoman Empire, Serbian resistance continued in northern regions (modern Vojvodina), under titular despots (until 1537), and popular leaders like Jovan Nenad (1526–1527). From 1521 to 1552, Ottomans conquered Belgrade and regions of Syrmia, Bačka, and Banat. Continuing wars and various rebellions constantly challenged Ottoman rule. One of the most significant was the Banat Uprising in 1594 and 1595, which was part of the Long War (1593–1606) between the Habsburgs and the Ottomans. The area of modern Vojvodina endured a century-long Ottoman occupation before being ceded to the Habsburg monarchy, partially by the Treaty of Karlovci (1699), and fully by the Treaty of Požarevac (1718). During the Habsburg-Ottoman war (1683–1699), much of Serbia switched from Ottoman rule to Habsburg control from 1688 to 1690, including the cities of Belgrade, Čačak, Užice and Niš, as well as the area of present-day Kosovo. However, the Ottoman army reconquered a large part of Serbia in the winter of 1689/1690, leading to a brutal massacre of the civilian population by uncontrolled Albanian and Tatar units. As a result of the persecutions, several tens of thousands of Serbs, led by the patriarch, Arsenije III Crnojević, fled northwards across the Sava river, to settle in Hungary, an event known as the Great Migration of 1690. In August 1690, following several petitions, the Emperor Leopold I formally granted Serbs from the Habsburg monarchy a first set of “privileges”, primarily to guarantee them freedom of religion. As a consequence, the ecclesiastical centre of the Serbs also moved northwards, to the Metropolitanate of Karlovci, and the Serbian Patriarchate of Peć was once-again abolished by the Ottomans in 1766. In 1718–39, the Habsburg monarchy occupied much of Central Serbia and established the Kingdom of Serbia as crownland. Those gains were lost by the Treaty of Belgrade in 1739, when the Ottomans retook the region. Apart from territory of modern Vojvodina which remained under the Habsburg Empire, central regions of Serbia were occupied once again by the Habsburgs in 1788–1792. The Serbian Revolution for independence from the Ottoman Empire lasted eleven years, from 1804 until 1815. The revolution comprised two separate uprisings which gained autonomy from the Ottoman Empire (1830) that eventually evolved towards full independence (1878). During the First Serbian Uprising (1804–1813), led by vožd Karađorđe Petrović, Serbia was independent for almost a decade before the Ottoman army was able to reoccupy the country. Shortly after this, the Second Serbian Uprising began in 1815. Led by Miloš Obrenović, it ended with a compromise between Serbian revolutionaries and Ottoman authorities. Likewise, Serbia was one of the first nations in the Balkans to abolish feudalism. The Akkerman Convention in 1826, the Treaty of Adrianople in 1829 and finally, the Hatt-i Sharif, recognised the suzerainty of Serbia. The First Serbian Constitution was adopted on 15 February 1835 (the anniversary of the outbreak of the First Serbian Uprising), making the country one of the first to adopt a democratic constitution in Europe. 15 February is now commemorated as Statehood Day, a public holiday. Following the clashes between the Ottoman army and Serbs in Belgrade in 1862, and under pressure from the Great Powers, by 1867 the last Turkish soldiers left the Principality, making the country de facto independent. By enacting a new constitution in 1869, without consulting the Porte, Serbian diplomats confirmed the de facto independence of the country. In 1876, Serbia declared war on the Ottoman Empire, siding with the ongoing Christian uprisings in Bosnia-Herzegovina and Bulgaria. The formal independence of the country was internationally recognised at the Congress of Berlin in 1878, which ended the Russo-Turkish War; this treaty, however, prohibited Serbia from uniting with other Serbian regions by placing Bosnia and Herzegovina under Austro-Hungarian occupation, alongside the occupation of the region of Raška. From 1815 to 1903, the Principality of Serbia was ruled by the House of Obrenović, save for the rule of Prince Aleksandar Karađorđević between 1842 and 1858. In 1882, Principality of Serbia became the Kingdom of Serbia, ruled by King Milan I. The House of Karađorđević, descendants of the revolutionary leader Karađorđe Petrović, assumed power in 1903 following the May Overthrow. In the north, the 1848 revolution in Austria led to the establishment of the autonomous territory of Serbian Vojvodina; by 1849, the region was transformed into the Voivodeship of Serbia and Banat of Temeschwar. In the course of the First Balkan War in 1912, the Balkan League defeated the Ottoman Empire and captured its European territories, which enabled territorial expansion of the Kingdom of Serbia into regions of Raška, Kosovo, Metohija, and Vardarian Macedonia. The Second Balkan War soon ensued when Bulgaria turned on its former allies, but was defeated, resulting in the Treaty of Bucharest. In two years, Serbia enlarged its territory by 80% and its population by 50%, it also suffered high casualties on the eve of World War I, with more than 36,000 dead. Austria-Hungary became wary of the rising regional power on its borders and its potential to become an anchor for unification of Serbs and other South Slavs, and the relationship between the two countries became tense. The assassination of Archduke Franz Ferdinand of Austria on 28 June 1914 in Sarajevo by Gavrilo Princip, a member of the Young Bosnia organisation, led to Austria-Hungary declaring war on Serbia, on 28 July 1914. Local war escalated when Germany declared war on Russia and invaded France and Belgium, thus drawing Great Britain into the conflict that became the First World War. Serbia won the first major battles of World War I, including the Battle of Cer, and the Battle of Kolubara, marking the first Allied victories against the Central Powers in World War I. Despite initial success, it was eventually overpowered by the Central Powers in 1915 and Austro-Hungarian occupation of Serbia followed. Most of its army and some people retreated through Albania to Greece and Corfu, suffering immense losses on the way. Serbia was occupied by the Central Powers. After the Central Powers military situation on other fronts worsened, the remains of the Serb army returned east and led a final breakthrough through enemy lines on 15 September 1918, liberating Serbia and defeating Bulgaria and Austria-Hungary. Serbia, with its campaign, was a major Balkan Entente Power which contributed significantly to the Allied victory in the Balkans in November 1918, especially by helping France force Bulgaria's capitulation. Serbia's casualties accounted for 8% of the total Entente military deaths; 58% (243,600) soldiers of the Serbian army perished in the war. The total number of casualties is placed around 700,000, more than 16% of Serbia's prewar size, and a majority (57%) of its overall male population. Serbia suffered the biggest casualty rate in World War I. The beginnings of the idea of the first common South Slavic state were the signing of a declaration on the island of Corfu in 1917. The Corfu Declaration was a formal agreement between the government-in-exile of the Kingdom of Serbia and the Yugoslav Committee (anti-Habsburg South Slav émigrés) that pledged to unify Kingdom of Serbia and Kingdom of Montenegro with Austria-Hungary's South Slav autonomous crown lands: Kingdom of Croatia-Slavonia, Kingdom of Dalmatia, Slovenia, Vojvodina (then part of the Kingdom of Hungary) and Bosnia and Herzegovina in a post-war Yugoslav state. It was signed on 20 July 1917 on Corfu. As the Austro-Hungarian Empire collapsed, the territory of Syrmia united with Serbia on 24 November 1918. Just a day later, on 25 November 1918, the Great People's Assembly of Serbs, Bunjevci and other Slavs in Banat, Bačka and Baranja declared the unification of these regions (Banat, Bačka, and Baranja) with the Kingdom of Serbia. On 26 November 1918, the Podgorica Assembly deposed the House of Petrović-Njegoš and united Montenegro with Serbia. On 1 December 1918, in Belgrade, Serbian Prince Regent Alexander Karađorđević proclaimed the Kingdom of the Serbs, Croats, and Slovenes, under King Peter I of Serbia. King Peter was succeeded by his son, Alexander, in August 1921. Serb centralists and Croat autonomists clashed in the parliament, and most governments were fragile and short-lived. Nikola Pašić, a conservative prime minister, headed or dominated most governments until his death. King Alexander established a dictatorship in 1929 with the aim of establishing the Yugoslav ideology and single Yugoslav nation, changed the name of the country to Yugoslavia and changed the internal divisions from the 33 oblasts to nine new banovinas. The effect of Alexander's dictatorship was to further alienate the non-Serbs living in Yugoslavia from the idea of unity. Alexander was assassinated in Marseille, during an official visit in 1934 by Vlado Chernozemski, member of the IMRO. Alexander was succeeded by his eleven-year-old son Peter II and a regency council was headed by his cousin, Prince Paul. In August 1939 the Cvetković–Maček Agreement established an autonomous Banate of Croatia as a solution to Croatian concerns. In 1941, in spite of Yugoslav attempts to remain neutral in the war, the Axis powers invaded Yugoslavia. The territory of modern Serbia was divided between Hungary, Bulgaria, the Independent State of Croatia, Greater Albania and Montenegro, while the remaining part of the occupied Serbia was placed under the military administration of Nazi Germany, with Serbian puppet governments led by Milan Aćimović and Milan Nedić assisted by Dimitrije Ljotić's fascist organization Yugoslav National Movement (Zbor). The Yugoslav territory was the scene of a civil war between royalist Chetniks commanded by Draža Mihailović and communist partisans commanded by Josip Broz Tito. Axis auxiliary units of the Serbian Volunteer Corps and the Serbian State Guard fought against both of these forces. The siege of Kraljevo was a major battle of the uprising in Serbia, led by Chetnik forces against the Nazis. Several days after the battle began the German forces committed a massacre of approximately 2,000 civilians in an event known as the Kraljevo massacre, in a reprisal for the attack. Draginac and Loznica massacre of 2,950 villagers in Western Serbia in 1941 was the first large execution of civilians in occupied Serbia by Germans, with Kragujevac massacre and Novi Sad Raid of Jews and Serbs by Hungarian fascists being the most notorious, with over 3,000 victims in each case. After one year of occupation, around 16,000 Serbian Jews were murdered in the area, or around 90% of its pre-war Jewish population during The Holocaust in Serbia. Many concentration camps were established across the area. Banjica concentration camp was the largest concentration camp and jointly run by the German army and Nedić's regime, with primary victims being Serbian Jews, Roma, and Serb political prisoners. During this period, hundreds of thousands of ethnic Serbs fled the Axis puppet state known as the Independent State of Croatia and sought refuge in German-occupied Serbia, seeking to escape the large-scale persecution and Genocide of Serbs, Jews, and Roma being committed by the Ustaše regime. The number of Serb victims was approximately 300,000 to 350,000. According to Tito himself, Serbs made up the vast majority of anti-fascist fighters and Yugoslav Partisans for the whole course of World War II. The Republic of Užice was a short-lived liberated territory established by the Partisans and the first liberated territory in World War II Europe, organised as a military mini-state that existed in the autumn of 1941 in the west of occupied Serbia. By late 1944, the Belgrade Offensive swung in favour of the partisans in the civil war; the partisans subsequently gained control of Yugoslavia. Following the Belgrade Offensive, the Syrmian Front was the last major military action of World War II in Serbia. A study by Vladimir Žerjavić estimates total war related deaths in Yugoslavia at 1,027,000, including 273,000 in Serbia. The victory of the Communist Partisans resulted in the abolition of the monarchy and a subsequent constitutional referendum. A one-party state was soon established in Yugoslavia by the Communist Party of Yugoslavia. It is claimed between 60,000 and 70,000 people died in Serbia during the 1944–45 communist takeover and purge. All opposition was suppressed and people deemed to be promoting opposition to socialism or promoting separatism were imprisoned or executed for sedition. Serbia became a constituent republic within the Federal People's Republic of Yugoslavia known as the People's Republic of Serbia, and had a republic-branch of the federal communist party, the League of Communists of Serbia. Serbia's most powerful and influential politician in Tito-era Yugoslavia was Aleksandar Ranković, one of the "big four" Yugoslav leaders, alongside Tito, Edvard Kardelj, and Milovan Đilas. Ranković was later removed from the office because of the disagreements regarding Kosovo's nomenklatura and the unity of Serbia. Ranković's dismissal was highly unpopular among Serbs. Pro-decentralisation reformers in Yugoslavia succeeded in the late 1960s in attaining substantial decentralisation of powers, creating substantial autonomy in Kosovo and Vojvodina, and recognising a distinctive "Muslim" nationality. As a result of these reforms, there was a massive overhaul of Kosovo's nomenklatura and police, that shifted from being Serb-dominated to ethnic Albanian-dominated through firing Serbs on a large scale. Further concessions were made to the ethnic Albanians of Kosovo in response to unrest, including the creation of the University of Pristina as an Albanian language institution. These changes created widespread fear among Serbs of being treated as second-class citizens. Belgrade, the capital of FPR Yugoslavia and PR Serbia, hosted the first Non-Aligned Movement Summit in September 1961, as well as the first major gathering of the Organization for Security and Co-operation in Europe (OSCE) with the aim of implementing the Helsinki Accords from October 1977 to March 1978. The 1972 smallpox outbreak in SAP Kosovo and other parts of SR Serbia was the last major outbreak of smallpox in Europe since World War II. In 1989, Slobodan Milošević rose to power in Serbia. Milošević promised a reduction of powers for the autonomous provinces of Kosovo and Vojvodina, where his allies subsequently took over power, during the Anti-bureaucratic revolution. This ignited tensions between the communist leadership of the other republics of Yugoslavia and awoke ethnic nationalism across Yugoslavia that eventually resulted in its breakup, with Slovenia, Croatia, Bosnia and Herzegovina, and Macedonia declaring independence during 1991 and 1992. Serbia and Montenegro remained together as the Federal Republic of Yugoslavia (FRY). However, according to the Badinter Commission, the country was not legally considered a continuation of the former SFRY, but a new state. Fueled by ethnic tensions, the Yugoslav Wars (1991–2001) erupted, with the most severe conflicts taking place in Croatia and Bosnia, where the large ethnic Serb communities opposed independence from Yugoslavia. The FRY remained outside the conflicts, but provided logistic, military and financial support to Serb forces in the wars. In response, the UN imposed sanctions against Yugoslavia which led to political isolation and the collapse of the economy (GDP decreased from $24 billion in 1990 to under $10 billion in 1993). Serbia was in the 2000s sued on the charges of alleged genocide by neighbouring Bosnia and Herzegovina and Croatia but in both cases the main charges against Serbia were dismissed. Multi-party democracy was introduced in Serbia in 1990, officially dismantling the one-party system. Despite constitutional changes, Milošević maintained strong political influence over the state media and security apparatus. When the ruling Socialist Party of Serbia refused to accept its defeat in municipal elections in 1996, Serbians engaged in large protests against the government. In 1998, continued clashes between the Albanian guerilla Kosovo Liberation Army and Yugoslav security forces led to the short Kosovo War (1998–99), in which NATO intervened, leading to the withdrawal of Serbian forces and the establishment of UN administration in the province. After the Yugoslav Wars, Serbia became home to highest number of refugees and internally displaced persons in Europe. After presidential elections in September 2000, opposition parties accused Milošević of electoral fraud, with the government claiming that his main challenger, Vojislav Koštunica, did not gain the majority of votes needed to avoid a run-off against Milošević. A campaign of civil resistance followed, led by the Democratic Opposition of Serbia (DOS), a broad coalition of anti-Milošević parties. This culminated on 5 October when half a million people from all over the country congregated in Belgrade, compelling Milošević to concede defeat. The fall of Milošević ended Yugoslavia's international isolation. Milošević was sent to the International Criminal Tribunal for the former Yugoslavia. The DOS announced that FR Yugoslavia would seek to join the European Union. In 2003, the Federal Republic of Yugoslavia was renamed Serbia and Montenegro; the EU opened negotiations with the country for the Stabilisation and Association Agreement. Serbia's political climate remained tense and in 2003, the Prime Minister Zoran Đinđić was assassinated as result of a plot originating from circles of organised crime and former security officials. In 2004 unrest in Kosovo took place, leaving 19 people dead and a number of Serbian Orthodox churches and monasteries destroyed or damaged. On 21 May 2006, Montenegro held a referendum to determine whether to end its union with Serbia. The results showed 55.4% of voters in favour of independence, which was just above the 55% required by the referendum. This was followed on 5 June 2006 by Serbia's declaration of independence, marking the final dissolution of the State Union of Serbia and Montenegro, and the re-emergence of Serbia as an independent state, for the first time since 1918. On the same occasion, the National Assembly of Serbia declared Serbia to be the legal successor to the former state union. The Assembly of Kosovo unilaterally declared independence from Serbia on 17 February 2008. Serbia immediately condemned the declaration and continues to deny any statehood to Kosovo. The declaration has sparked varied responses from the international community, some welcoming it, while others condemned the unilateral move. Status-neutral talks between Serbia and Kosovo-Albanian authorities are held in Brussels, mediated by the EU. Serbia officially applied for membership in the European Union on 22 December 2009, and received candidate status on 1 March 2012, following a delay in December 2011. Following a positive recommendation of the European Commission and European Council in June 2013, negotiations to join the EU commenced in January 2014. Since Aleksandar Vučić and his Serbian Progressive Party came to power in 2012, Serbia has suffered from democratic backsliding into authoritarianism, followed by a decline in media freedom and civil liberties. After the COVID-19 pandemic spread to Serbia in March 2020, a state of emergency was declared and a curfew was introduced for the first time in Serbia since World War II. In January and February 2021, Serbia carried the second-fastest vaccine rollout in Europe. In April 2022, President Aleksandar Vučić was re-elected. Serbia drew western criticism for not joining EU sanctions against Russia and maintaining bilateral relations after the 2022 Russian invasion of Ukraine. However, Serbia condemned Russia at the United Nations General Assembly and Human Rights Council. In December 2023, ruling Serbian Progressive Party (SNS) of President Vucic won a snap parliamentary election, gaining an absolute majority with more than half of the 250 seats in the National Assembly. The election resulted in protests, with opposition supporters claiming that the election result was fraudulent. A landlocked country situated at the crossroads between Central and Southeastern Europe, Serbia is located in the Balkan peninsula and the Pannonian Plain. Serbia lies between latitudes 41° and 47° N, and longitudes 18° and 23° E. The country covers a total of 88,499 km (34,170 sq mi) (including Kosovo), which places it at 111th place in the world; with Kosovo excluded, the total area is 77,474 km (29,913 sq mi), which would make it 117th. Its total border length amounts to 2,027 km (1,260 mi): Albania 115 km (71 mi), Bosnia and Herzegovina 302 km (188 mi), Bulgaria 318 km (198 mi), Croatia 241 km (150 mi), Hungary 151 km (94 mi), North Macedonia 221 km (137 mi), Montenegro 203 km (126 mi) and Romania 476 km (296 mi). All of Kosovo's border with Albania (115 km (71 mi)), North Macedonia (159 km (99 mi)) and Montenegro (79 km (49 mi)) are under control of the Kosovo border police. Serbia treats the 352 km (219 mi) long border between Kosovo and rest of Serbia as an "administrative line"; it is under shared control of Kosovo border police and Serbian police forces, and there are 11 crossing points. The Pannonian Plain covers the northern third of the country (Vojvodina and Mačva) while the easternmost tip of Serbia extends into the Wallachian Plain. The terrain of the central part of the country, with the region of Šumadija at its heart, consists chiefly of hills traversed by rivers. Mountains dominate the southern third of Serbia. Dinaric Alps stretch in the west and the southwest, following the flow of the rivers Drina and Ibar. The Carpathian Mountains and Balkan Mountains stretch in a north–south direction in eastern Serbia. Ancient mountains in the southeast corner of the country belong to the Rilo-Rhodope Mountain system. Elevation ranges from the Midžor peak of the Balkan Mountains at 2,169 metres (7,116 feet) (the highest peak in Serbia, excluding Kosovo) to the lowest point of just 17 metres (56 feet) near the Danube river at Prahovo. The largest lake is Đerdap Lake (163 square kilometres (63 sq mi)) and the longest river passing through Serbia is the Danube (587.35 kilometres (364.96 mi)). The climate of Serbia is under the influences of the landmass of Eurasia and the Atlantic Ocean and Mediterranean Sea. With mean January temperatures around 0 °C (32 °F), and mean July temperatures of 22 °C (72 °F), it can be classified as a warm-humid continental or humid subtropical climate. In the north, the climate is more continental, with cold winters, and hot, humid summers along with well-distributed rainfall patterns. In the south, summers and autumns are drier, and winters are relatively cold, with heavy inland snowfall in the mountains. Differences in elevation, proximity to the Adriatic Sea and large river basins, as well as exposure to the winds account for climate variations. Southern Serbia is subject to Mediterranean influences. The Dinaric Alps and other mountain ranges contribute to the cooling of most of the warm air masses. Winters are quite harsh in the Pešter plateau, because of the mountains which encircle it. One of the climatic features of Serbia is Košava, a cold and very squally southeastern wind which starts in the Carpathian Mountains and follows the Danube northwest through the Iron Gate where it gains a jet effect and continues to Belgrade and can spread as far south as Niš. The average annual air temperature for the period 1961–1990 for the area with an elevation of up to 300 m (984 ft) is 10.9 °C (51.6 °F). The areas with an elevation of 300 to 500 m (984 to 1,640 ft) have an average annual temperature of around 10.0 °C (50.0 °F), and over 1,000 m (3,281 ft) of elevation around 6.0 °C (42.8 °F). The lowest recorded temperature in Serbia was −39.5 °C (−39.1 °F) on 13 January 1985, Karajukića Bunari in Pešter, and the highest was 44.9 °C (112.8 °F), on 24 July 2007, recorded in Smederevska Palanka. Serbia is one of few European countries with very high risk exposure to natural hazards (earthquakes, storms, floods, droughts). It is estimated that potential floods, particularly in areas of Central Serbia, threaten over 500 larger settlements and an area of 16,000 square kilometres. The most disastrous were the floods in May 2014, when 57 people died and a damage of over a 1.5 billion euro was inflicted. Almost all of Serbia's rivers drain to the Black Sea, by way of the Danube river. The Danube, the second largest European river, passes through Serbia with 588 kilometres (21% of its overall length) and represents the major source of fresh water. It is joined by its biggest tributaries, the Great Morava (longest river entirely in Serbia with 493 km (306 mi) of length), Sava and Tisza rivers. One notable exception is the Pčinja which flows into the Aegean. Drina river forms the natural border between Bosnia and Herzegovina and Serbia, and represents the main kayaking and rafting attraction in both countries. Due to configuration of the terrain, natural lakes are sparse and small; most of them are located in the lowlands of Vojvodina, like the aeolian lake Palić or numerous oxbow lakes along river flows (like Zasavica and Carska Bara). However, there are numerous artificial lakes, mostly due to hydroelectric dams, the biggest being Đerdap (Iron Gates) on the Danube with 163 km (63 sq mi) on the Serbian side (a total area of 253 km (98 sq mi) is shared with Romania); Perućac on the Drina, and Vlasina. The largest waterfall, Jelovarnik, located in Kopaonik, is 71 m high. Abundance of relatively unpolluted surface waters and numerous underground natural and mineral water sources of high water quality presents a chance for export and economy improvement; however, more extensive exploitation and production of bottled water began only recently. Serbia is a country of rich ecosystem and species diversity—covering only 1.9% of the whole European territory, Serbia is home to 39% of European vascular flora, 51% of European fish fauna, 40% of European reptiles and amphibian fauna, 74% of European bird fauna, and 67% European mammal fauna. Its abundance of mountains and rivers make it an ideal environment for a variety of animals, many of which are protected including wolves, lynx, bears, foxes, and stags. There are 17 snake species living all over the country, 8 of them are venomous. Mountain of Tara in western Serbia is one of the last regions in Europe where bears can still live in absolute freedom. Serbia is home to about 380 species of birds. In Carska Bara, there are over 300 bird species on just a few square kilometres. Uvac Gorge is considered one of the last habitats of the Griffon vulture in Europe. In area around the city of Kikinda, in the northernmost part of the country, some 145 endangered long-eared owls are noted, making it the world's biggest settlement of these species. The country is considerably rich with threatened species of bats and butterflies as well. There are 380 protected areas of Serbia, encompassing 4,947 square kilometres or 6.4% of the country. The "Spatial plan of the Republic of Serbia" states that the total protected area should be increased to 12% by 2021. Those protected areas include 5 national parks (Đerdap, Tara, Kopaonik, Fruška Gora and Šar Mountain), 15 nature parks, 15 "landscapes of outstanding features", 61 nature reserves, and 281 natural monuments. With 29.1% of its territory covered by forest, Serbia is considered to be a middle-forested country, compared on a global scale to world forest coverage at 30%, and European average of 35%. The total forest area in Serbia is 2,252,000 ha (1,194,000 ha or 53% are state-owned, and 1,058,387 ha or 47% are privately owned) or 0.3 ha per inhabitant. It had a 2019 Forest Landscape Integrity Index mean score of 5.29/10, ranking it 105th globally out of 172 countries. The most common trees are oak, beech, pines, and firs. Air pollution is a significant problem in Bor area, due to work of large copper mining and smelting complex, and Pančevo where oil and petrochemical industry is based. Some cities suffer from water supply problems, due to mismanagement and low investments in the past, as well as water pollution (like the pollution of the Ibar River from the Trepča zinc-lead combinate, affecting the city of Kraljevo, or the presence of natural arsenic in underground waters in Zrenjanin). Poor waste management has been identified as one of the most important environmental problems in Serbia and the recycling is a fledgling activity, with only 15% of its waste being turned back for reuse. The 1999 NATO bombing caused serious damage to the environment, with several thousand tonnes of toxic chemicals stored in targeted factories and refineries released into the soil and water basins. Serbia is a parliamentary republic, with the government divided into legislative, executive, and judiciary branches. Serbia had one of the first modern constitutions in Europe, the 1835 Constitution (known as the Sretenje Constitution), which was at the time considered among the most progressive and liberal constitutions in Europe. Since then it has adopted 10 different constitutions. The current constitution was adopted in 2006 in the aftermath of the Montenegro independence referendum which by consequence renewed the independence of Serbia itself. The Constitutional Court rules on matters regarding the Constitution. The President of the Republic (Predsednik Republike) is the head of state, is elected by popular vote to a five-year term and is limited by the Constitution to a maximum of two terms. In addition to being the commander in chief of the armed forces, the president has the procedural duty of appointing the prime minister with the consent of the parliament, and has some influence on foreign policy. Aleksandar Vučić of the Serbian Progressive Party is the current president following the 2017 presidential election. Seat of the presidency is Novi Dvor. The Government (Vlada) is composed of the prime minister and cabinet ministers. The Government is responsible for proposing legislation and a budget, executing the laws, and guiding the foreign and internal policies. The current prime minister is Ana Brnabić, nominated by the Serbian Progressive Party. The National Assembly (Narodna skupština) is a unicameral legislative body. The National Assembly has the power to enact laws, approve the budget, schedule presidential elections, select and dismiss the Prime Minister and other ministers, declare war, and ratify international treaties and agreements. It is composed of 250 proportionally elected members who serve four-year terms. Following the 2020 parliamentary election, the largest political parties in the National Assembly are the populist Serbian Progressive Party and Socialist Party of Serbia, that with its partners, hold more than a supermajority number of seats. In 2021, Serbia was the 5th country in Europe by the number of women holding high-ranking public functions. Serbia is the fourth modern-day European country, after France, Austria and the Netherlands, to have a codified legal system. The country has a three-tiered judicial system, made up of the Supreme Court of Cassation as the court of the last resort, Courts of Appeal as the appellate instance, and Basic and High courts as the general jurisdictions at first instance. Courts of special jurisdictions are the Administrative Court, commercial courts (including the Commercial Court of Appeal at second instance) and misdemeanor courts (including High Misdemeanor Court at second instance). The judiciary is overseen by the Ministry of Justice. Serbia has a typical civil law legal system. Law enforcement is the responsibility of the Serbian Police, which is subordinate to the Ministry of the Interior. Serbian Police fields 27,363 uniformed officers. National security and counterintelligence are the responsibility of the Security Intelligence Agency (BIA). Serbia has established diplomatic relations with 191 UN member states, the Holy See, the Sovereign Military Order of Malta and the European Union. Foreign relations are conducted through the Ministry of Foreign Affairs. Serbia has a network of 65 embassies and 23 consulates internationally. There are 69 foreign embassies, 5 consulates and 4 liaison offices in Serbia. Serbian foreign policy is focused on achieving the strategic goal of becoming a member state of the European Union (EU). Serbia started the process of joining the EU by signing of the Stabilisation and Association Agreement on 29 April 2008 and officially applied for membership in the European Union on 22 December 2009. It received a full candidate status on 1 March 2012 and started accession talks on 21 January 2014. The European Commission considers accession possible by 2025. On 17 February 2008, Kosovo unilaterally declared independence from Serbia. In protest, Serbia initially recalled its ambassadors from countries that recognised Kosovo's independence. The resolution of 26 December 2007 by the National Assembly stated that both the Kosovo declaration of independence and recognition thereof by any state would be gross violation of international law. Serbia began cooperation and dialogue with NATO in 2006, when the country joined the Partnership for Peace programme and the Euro-Atlantic Partnership Council. The country's military neutrality was formally proclaimed by a resolution adopted by Serbia's parliament in December 2007, which makes joining any military alliance contingent on a popular referendum, a stance acknowledged by NATO. On the other hand, Serbia's relations with Russia are habitually described by mass media as a "centuries-old religious, ethnic and political alliance" and Russia is said to have sought to solidify its relationship with Serbia since the imposition of sanctions against Russia in 2014. During the 2022 Russian invasion of Ukraine, Serbia voted to condemn the invasion, supporting the adoption of the United Nations draft resolution demanding Russia to withdraw its military forces from Ukraine. However, Serbia is one of the only countries in Europe not to sanction Russia after the invasion. The Serbian Armed Forces are subordinate to the Ministry of Defence, and are composed of the Army and the Air Force. Although a landlocked country, Serbia operates a River Flotilla which patrols on the Danube, Sava and Tisa rivers. The Serbian Chief of the General Staff reports to the Defence Minister. The Chief of Staff is appointed by the president, who is the commander-in-chief. As of 2019, Serbian defence budget amounts to $804 million. Traditionally having relied on a large number of conscripts, Serbian Armed Forces went through a period of downsizing, restructuring and professionalisation. Conscription was abolished in 2011. Serbian Armed Forces have 28,000 active troops, supplemented by the "active reserve" which numbers 20,000 members and "passive reserve" with about 170,000. Serbia participates in the NATO Individual Partnership Action Plan programme, but has no intention of joining NATO, due to significant popular rejection, largely a legacy of the NATO bombing of Yugoslavia in 1999. It is an observer member of the Collective Security Treaty Organisation (CSTO) as of 2013. The country also signed the Stability Pact for South Eastern Europe. The Serbian Armed Forces take part in several multinational peacekeeping missions, including deployments in Lebanon, Cyprus, Ivory Coast, and Liberia. Serbia is a major producer and exporter of military equipment in the region. Defence exports totaled around $600 million in 2018. The defence industry has seen significant growth over the years and it continues to grow on a yearly basis. Serbia is one of the countries with the largest number of firearms in the civilian population in the world. Serbia is a unitary state composed of municipalities/cities, districts, and two autonomous provinces. In Serbia, excluding Kosovo, there are 145 municipalities (opštine) and 29 cities (gradovi), which form the basic units of local self-government. Apart from municipalities/cities, there are 24 districts (okruzi, 10 most populated listed below), with the City of Belgrade constituting an additional district. Except for Belgrade, which has an elected local government, districts are regional centres of state authority, but have no powers of their own; they present purely administrative divisions. The Constitution of Serbia recognizes two autonomous provinces, Vojvodina in the north, and the disputed territory of Kosovo and Metohija in the south, while the remaining area of Central Serbia never had its own regional authority. Following the Kosovo War, UN peacekeepers entered Kosovo and Metohija, as per UNSC Resolution 1244. The government of Serbia does not recognise Kosovo's February 2008 declaration of independence, considering it illegal and illegitimate. As of 2022 census, Serbia (excluding Kosovo) has a total population of 6,647,003 and the overall population density is medium as it stands at 85.8 inhabitants per square kilometre. The census was not conducted in Kosovo which held its own census that numbered their total population at 1,739,825, excluding Serb-inhabited North Kosovo, as Serbs from that area (about 50,000) boycotted the census. Serbia has been enduring a demographic crisis since the beginning of the 1990s, with a death rate that has continuously exceeded its birth rate. It is estimated that 300,000 people left Serbia during the 1990s, 20% of whom had a higher education. Serbia subsequently has one of the oldest populations in the world, with the average age of 43.3 years, and its population is shrinking at one of the fastest rates in the world. A fifth of all households consist of only one person, and just one-fourth of four and more persons. Average life expectancy in Serbia at birth is 76.1 years. During the 1990s, Serbia had the largest refugee population in Europe. Refugees and internally displaced persons (IDPs) in Serbia formed between 7% and 7.5% of its population at the time – about half a million refugees sought refuge in the country following the series of Yugoslav wars, mainly from Croatia (and to a lesser extent from Bosnia and Herzegovina) and the IDPs from Kosovo. Serbs with 5,360,239 are the largest ethnic group in Serbia, representing 81% of the total population (excluding Kosovo). Serbia is one of the European countries with the highest number of registered national minorities, while the province of Vojvodina is recognizable for its multi-ethnic and multi-cultural identity. Despite a decline in recent years, with a population of 184,442, Hungarians are still the largest ethnic minority in Serbia, concentrated predominantly in northern Vojvodina and representing 2.8% of the country's population (10.5% in Vojvodina). Romani population stands at 131,936 according to the 2022 census but unofficial estimates place their actual number between 400,000 and 500,000. Bosniaks with 153,801 and Muslims by nationality with 13,011 are concentrated in Raška (Sandžak), in the southwest. Other minority groups include Albanians, Croats and Bunjevci, Slovaks, Yugoslavs, Montenegrins, Romanians and Vlachs, Macedonians and Bulgarians. Chinese, estimated at 15,000, are the only significant non-European immigrant minority. Most recently, tens of thousands of Russians and Ukrainians have immigrated to Serbia following the Russian Invasion of Ukraine. The majority of the population, or 59.4%, reside in urban areas and some 16.1% in Belgrade alone. Belgrade is the only city with more than a million inhabitants and there are four more with over 100,000 inhabitants. The Constitution of Serbia defines it as a secular state with guaranteed religious freedom. Orthodox Christians with 6,079,396 comprise 84.5% of country's population. The Serbian Orthodox Church is the largest and traditional church of the country, adherents of which are overwhelmingly Serbs. Other Orthodox Christian communities in Serbia include Montenegrins, Romanians, Vlachs, Macedonians and Bulgarians. In 2011, Roman Catholics numbered 356,957 in Serbia, or roughly 6% of the population, mostly in northern Vojvodina which is home to ethnic minority groups such as Hungarians, Croats, and Bunjevci, as well as to some Slovaks and Czechs. Greek Catholic Church is adhered by around 25,000 citizens (0.37% of the population), mostly Rusyns in Vojvodina. Protestantism accounts for about 1% of the country's population, chiefly Lutheranism among Slovaks in Vojvodina as well as Calvinism among Reformed Hungarians. Muslims, with 222,282 or 3% of the population, form the third largest religious group. Islam has a strong historic following in the southern regions of Serbia, primarily in southern Raška. Bosniaks are the largest Islamic community in Serbia, followed by Albanians; estimates are that around a third of the country's Roma people are Muslim. In 2011, there were only 578 Jews in Serbia, compared to over 30,000 prior to World War II. Atheists numbered 80,053, or 1.1% of the population, and an additional 4,070 declared themselves to be agnostics. The official language is Serbian, native to 88% of the population. Serbian is the only European language with active digraphia, using both Cyrillic and Latin alphabets. Serbian Cyrillic is designated in the Constitution as the "official script" and was devised in 1814 by Serbian philologist Vuk Karadžić, who based it on phonemic principles. A survey from 2014 showed that 47% of Serbians favour the Latin alphabet, 36% favour the Cyrillic one and 17% have no preference. Standard Serbian is mutually intelligible with recognised minority languages of Bosnian and Croatian, as all three are based on the most widespread Shtokavian dialect from Eastern Herzegovina. Other recognised minority languages are: Hungarian, Slovak, Albanian, Romanian, Bulgarian, Rusyn, and Macedonian. All these languages are in official use in municipalities or cities where the ethnic minority exceeds 15% of the total population. In Vojvodina, the provincial administration co-officially uses, besides Serbian, five other languages (Slovak, Hungarian, Croatian, Romanian and Rusyn). The healthcare system in Serbia is organized and managed by the three primary institutions: The Ministry of Health, The Institute of Public Health of Serbia "Dr Milan Jovanović Batut" and the Military Medical Academy. The right to healthcare protections is defined as a constitutional right in Serbia. The Serbian public health system is based on the principles of equity and solidarity, organized on the model of compulsory health insurance contributions. Private health care is not integrated into the public health system, but certain services may be included by contracting. The Ministry of Health determines the healthcare policy and adopts standards for the work of the healthcare service. The Ministry is also in charge of the health care system, health insurance, preservation and improvement of health of citizens, health inspection, supervision over the work of the healthcare service and other tasks in the field of health care. The Institute of Public Health of Serbia "Dr Milan Jovanović Batut" is responsible for medical statistics, epidemiology and hygiene. This central, tertiary institution manages and coordinates a dense network of municipal and regional Centers of Public Health, spread across the entire country, that provide services in the domain of epidemiology and hygiene on the primary and secondary level. The Republic Health Insurance Institute finances the functioning of health care at all levels. One of the most important health institutions in Serbia is the Military Medical Academy in Belgrade. It takes care of about 30,000 patients a year (military and civilian insured). The academy performs around 30,000 surgical interventions and more than 500,000 specialist examinations. The Clinical Centre of Serbia spreads over 34 hectares in Belgrade and consists of about 50 buildings, while also has 3,150 beds considered to be the highest number in Europe, and among highest in the world. Other important health institutions include: KBC Dr Dragiša Mišović, Cardiovascular institute Detinje, Clinical Centre of Kragujevac, Clinical Centre of Niš, Clinical Center of Vojvodina and others. Medical specialists from Serbia have performed a number of operations which have been described as "pioneer works". Serbia has an emerging market economy in upper-middle income range. According to the International Monetary Fund, Serbian nominal GDP in 2022 is officially estimated at $65.697 billion or $9,561 per capita while purchasing power parity GDP stood at $153.076 billion or $22,278 per capita. The economy is dominated by services which accounts for 67.9% of GDP, followed by industry with 26.1% of GDP, and agriculture at 6% of GDP. The official currency of Serbia is Serbian dinar (ISO code: RSD), and the central bank is National Bank of Serbia. The Belgrade Stock Exchange is the only stock exchange in the country, with market capitalisation of $8.65 billion and BELEX15 as the main index representing the 15 most liquid stocks. The country is ranked 52nd on the Social Progress Index as well as 51st on the Global Peace Index. The economy has been affected by the global economic crisis. After almost a decade of strong economic growth (average of 4.45% per year), Serbia entered the recession in 2009 with negative growth of −3% and again in 2012 and 2014 with −1% and −1.8%, respectively. As the government was fighting effects of crisis the public debt has more than doubled: from pre-crisis level of just under 30% to about 70% of GDP and trending downwards recently to around 50%. Labour force stands at 3.2 million, with 56% employed in services sector, 28.1% in industry and 15.9% in the agriculture. The average monthly net salary in May 2019 stood at 47,575 dinars or $525. The unemployment remains an acute problem, with rate of 12.7% as of 2018. Since 2000, Serbia has attracted over $40 billion in foreign direct investment (FDI). Blue-chip corporations making investments include: Fiat Chrysler Automobiles, Siemens, Bosch, Philip Morris, Michelin, Coca-Cola, Carlsberg and others. In the energy sector, Russian energy giants, Gazprom and Lukoil have made large investments. In metallurgy sector, Chinese steel and copper giants, Hesteel and Zijin Mining have acquired key complexes. Serbia has an unfavourable trade balance: imports exceed exports by 25%. Serbia's exports, however, recorded a steady growth in last couple of years reaching $19.2 billion in 2018. The country has free trade agreements with the EFTA and CEFTA, a preferential trade regime with the European Union, a Generalised System of Preferences with the United States, and individual free trade agreements with Russia, Belarus, Kazakhstan, and Turkey. Serbia has very favourable natural conditions (land and climate) for varied agricultural production. It has 5,056,000 ha of agricultural land (0.7 ha per capita), out of which 3,294,000 ha is arable land (0.45 ha per capita). In 2016, Serbia exported agricultural and food products worth $3.2 billion, and the export-import ratio was 178%. Agricultural exports constitute more than one-fifth of all Serbia's sales on the world market. Serbia is one of the largest provider of frozen fruit to the EU (largest to the French market, and 2nd largest to the German market). Agricultural production is most prominent in Vojvodina on the fertile Pannonian Plain. Other agricultural regions include Mačva, Pomoravlje, Tamnava, Rasina, and Jablanica. In the structure of the agricultural production, 70% is from the crop field production and 30% is from the livestock production. Serbia is world's second largest producer of plums (582,485 tonnes; second to China), second largest of raspberries (89,602 tonnes, second to Poland), it is also a significant producer of maize (6.48 million tonnes, ranked 32nd in the world) and wheat (2.07 million tonnes, ranked 35th in the world). Other important agricultural products are: sunflower, sugar beet, soybean, potato, apple, pork meat, beef, poultry and dairy. There are 56,000 ha of vineyards in Serbia, producing about 230 million litres of wine annually. The most famous viticulture regions are located in Vojvodina and Šumadija. The industry was the economic sector hardest hit by the UN sanctions and trade embargo and NATO bombing during the 1990s and transition to market economy during the 2000s. The industrial output saw dramatic downsizing: in 2013 it was expected to be only a half of that of 1989. Main industrial sectors include: automotive, mining, non-ferrous metals, food-processing, electronics, pharmaceuticals, clothes. Serbia has 14 free economic zones as of September 2017, in which many foreign direct investments are realised. Automotive industry (with Fiat Chrysler Automobiles as a forebear) is dominated by cluster located in Kragujevac and its vicinity, and contributes to export with about $2 billion. Country is a leading steel producer in the wider region of Southeast Europe and had production of nearly 2 million tonnes of raw steel in 2018, coming entirely from Smederevo steel mill, owned by the Chinese Hesteel. Serbia's mining industry is comparatively strong: Serbia is the 18th largest producer of coal (7th in Europe) extracted from large deposits in Kolubara and Kostolac basins; it is also world's 23rd largest (3rd in Europe) producer of copper which is extracted by Zijin Bor Copper, a large copper mining company, acquired by Chinese Zijin Mining in 2018; significant gold extraction is developed around Majdanpek. Serbia notably manufactures intel smartphones named Tesla smartphones. Food industry is well known both regionally and internationally and is one of the strong points of the economy. Some of the international brand-names established production in Serbia: PepsiCo and Nestlé in food-processing sector; Coca-Cola (Belgrade), Heineken (Novi Sad) and Carlsberg (Bačka Palanka) in beverage industry; Nordzucker in sugar industry. Serbia's electronics industry had its peak in the 1980s and the industry today is only a third of what it was back then, but has witnessed a something of revival in last decade with investments of companies such as Siemens (wind turbines) in Subotica, Panasonic (lighting devices) in Svilajnac, and Gorenje (electrical home appliances) in Valjevo. The pharmaceutical industry in Serbia comprises a dozen manufacturers of generic drugs, of which Hemofarm in Vršac and Galenika in Belgrade, account for 80% of production volume. Domestic production meets over 60% of the local demand. The energy sector is one of the largest and most important sectors to the country's economy. Serbia is a net exporter of electricity and importer of key fuels (such as oil and gas). Serbia has an abundance of coal, and significant reserves of oil and gas. Serbia's proven reserves of 5.5 billion tonnes of coal lignite are the fifth largest in the world (second in Europe, after Germany). Coal is found in two large deposits: Kolubara (4 billion tonnes of reserves) and Kostolac (1.5 billion tonnes). Despite being small on a world scale, Serbia's oil and gas resources (77.4 million tonnes of oil equivalent and 48.1 billion cubic metres, respectively) have a certain regional importance since they are largest in the region of former Yugoslavia as well as the Balkans (excluding Romania). Almost 90% of the discovered oil and gas are to be found in Banat and those oil and gas fields are by size among the largest in the Pannonian basin but are average on a European scale. The production of electricity in 2015 in Serbia was 36.5 billion kilowatt-hours (KWh), while the final electricity consumption amounted to 35.5 billion kilowatt-hours (KWh). Most of the electricity produced comes from thermal-power plants (72.7% of all electricity) and to a lesser degree from hydroelectric-power plants (27.3%). There are 6 lignite-operated thermal-power plants with an installed power of 3,936 MW; largest of which are 1,502 MW-Nikola Tesla 1 and 1,160 MW-Nikola Tesla 2, both in Obrenovac. Total installed power of 9 hydroelectric-power plants is 2,831 MW, largest of which is Đerdap 1 with capacity of 1,026 MW. In addition to this, there are mazute and gas-operated thermal-power plants with an installed power of 353 MW. The entire production of electricity is concentrated in Elektroprivreda Srbije (EPS), public electric-utility power company. The current oil production in Serbia amounts to over 1.1 million tonnes of oil equivalent and satisfies some 43% of country's needs while the rest is imported. National petrol company, Naftna Industrija Srbije (NIS), was acquired in 2008 by Gazprom Neft. The company's refinery in Pančevo (capacity of 4.8 million tonnes) is one of the most modern oil-refineries in Europe; it also operates network of 334 filling stations in Serbia (74% of domestic market) and additional 36 stations in Bosnia and Herzegovina, 31 in Bulgaria, and 28 in Romania. There are 155 kilometres of crude oil pipelines connecting Pančevo and Novi Sad refineries as a part of trans-national Adria oil pipeline. Serbia is heavily dependent on foreign sources of natural gas, with only 17% coming from domestic production (totalling 491 million cubic metres in 2012) and the rest is imported, mainly from Russia (via gas pipelines that run through Ukraine and Hungary). Srbijagas, public company, operates the natural gas transportation system which comprise 3,177 kilometres (1,974 mi) of trunk and regional natural gas pipelines and a 450 million cubic metre underground gas storage facility at Banatski Dvor. In 2021, Balkan Stream gas pipeline opened through Serbia. Serbia has a strategic transportation location since the country's backbone, Morava Valley, represents the easiest land route from continental Europe to Asia Minor and the Near East. Serbian road network carries the bulk of traffic in the country. Total length of roads is 45,419 km (28,222 mi) of which 962 km (598 mi) are "class-IA state roads" (i.e. motorways); 4,517 km (2,807 mi) are "class-IB state roads" (national roads); 10,941 km (6,798 mi) are "class-II state roads" (regional roads) and 23,780 km (14,780 mi) are "municipal roads". The road network, except for the most of class-IA roads, are of comparatively lower quality to the Western European standards because of lack of financial resources for their maintenance in the last 20 years. Over 300 km (190 mi) of new motorways were constructed in the last decade and additional 154 km (96 mi) are currently under construction: A5 motorway (from north of Kruševac to Čačak) and 31 km (19 mi)-long segment of A2 (between Čačak and Požega). Coach transport is very extensive: almost every place in the country is connected by bus, from largest cities to the villages; in addition there are international routes (mainly to countries of Western Europe with large Serb diaspora). Routes, both domestic and international, are served by more than hundred intercity coach services, biggest of which are Lasta and Niš-Ekspres. As of 2018, there were 1,999,771 registered passenger cars or 1 passenger car per 3.5 inhabitants. Serbia has 3,819 km (2,373 mi) of rail tracks, of which 1,279 km (795 mi) are electrified and 283 km (176 mi) are double-track railroad. The major rail hub is Belgrade (and to a lesser degree Niš), while the most important railroads include: Belgrade–Subotica–Budapest (Hungary) (currently upgraded to high-speed status), Belgrade–Bar (Montenegro), Belgrade–Šid–Zagreb (Croatia)/Belgrade–Niš–Sofia (Bulgaria) (part of Pan-European Corridor X), and Niš–Thessaloniki (Greece). Some 75 km (46 mi) of new high-speed rail line between Belgrade and Novi Sad was opened in 2022 and additional 108 km (67 mi) from Novi Sad to Subotica and border with Hungary are currently under construction and due to open in 2025. Construction work for 212 km-long prolongation of the high-speed rail line to the south, to the city of Niš, is set to commence in 2024 and with its planned completion by the end of the decade four of country's five largest cities will be connected by the high-speed rail lines. Rail services are operated by Srbija Voz (passenger transport) and Srbija Kargo (freight transport). There are three airports with regular passenger services reaching over 6 million passengers in 2022 with Belgrade Nikola Tesla Airport serving bulk of it, being a hub of flagship carrier Air Serbia which flies to 80 destinations in 32 countries (including intercontinental flights to New York City, Chicago and Tianjin) and carried 2.75 million passengers in 2022. Serbia has a developed inland water transport since there are 1,716 km (1,066 mi) of navigable inland waterways (1,043 km, 648 mi of navigable rivers and 673 km, 418 mi of navigable canals), which are almost all located in northern third of the country. The most important inland waterway is the Danube (part of Pan-European Corridor VII). Other navigable rivers include Sava, Tisza, Begej and Timiş Rivers, all of which connect Serbia with Northern and Western Europe through the Rhine–Main–Danube Canal and North Sea route, to Eastern Europe via the Tisza, Begej and Danube Black Sea routes, and to Southern Europe via the Sava river. More than 8 million tonnes of cargo were transported on Serbian rivers and canals in 2018 while the largest river ports are: Novi Sad, Belgrade, Pančevo, Smederevo, Prahovo and Šabac. Fixed telephone lines connect 81% of households in Serbia, and with about 9.1 million users the number of cellphones surpasses the total population of by 28%. The largest mobile operator is Telekom Srbija with 4.2 million subscribers, followed by Telenor with 2.8 million users and A1 with about 2 million. Some 58% of households have fixed-line (non-mobile) broadband Internet connection while 67% are provided with pay television services (i.e. 38% cable television, 17% IPTV, and 10% satellite). Digital television transition has been completed in 2015 with DVB-T2 standard for signal transmission. Serbia is not a mass-tourism destination but nevertheless has a diverse range of touristic products. In 2019, total of over 3.6 million tourists were recorded in accommodations, of which half were foreign. Foreign exchange earnings from tourism were estimated at $1.5 billion. Tourism is mainly focused on the mountains and spas of the country, which are mostly visited by domestic tourists, as well as Belgrade and, to a lesser degree, Novi Sad, which are preferred choices of foreign tourists (almost two-thirds of all foreign visits are made to these two cities). The most famous mountain resorts are Kopaonik, Stara Planina and Zlatibor. There are also many spas in Serbia, the biggest of which are Vrnjačka Banja, Soko Banja, and Banja Koviljača. City-break and conference tourism is developed in Belgrade and Novi Sad. Other touristic products that Serbia offer are natural wonders like Đavolja varoš, Christian pilgrimage to the many Orthodox monasteries across the country and the river cruising along the Danube. There are several internationally popular music festivals held in Serbia, such as EXIT (with 25–30,000 foreign visitors coming from 60 countries) and the Guča trumpet festival. According to 2011 census, literacy in Serbia stands at 98% of population while computer literacy is at 49% (complete computer literacy is at 34.2%). Same census showed the following levels of education: 16.2% of inhabitants have higher education (10.6% have bachelors or master's degrees, 5.6% have an associate degree), 49% have a secondary education, 20.7% have an elementary education, and 13.7% have not completed elementary education. Education in Serbia is regulated by the Ministry of Education and Science. Education starts in either preschools or elementary schools. Children enroll in elementary schools at the age of seven. Compulsory education consists of eight grades of elementary school. Students have the opportunity to attend gymnasiums and vocational schools for another four years, or to enroll in vocational training for 2 to 3 years. Following the completion of gymnasiums or vocational schools, students have the opportunity to attend university. Elementary and secondary education are also available in languages of recognised minorities in Serbia, where classes are held in Hungarian, Slovak, Albanian, Romanian, Rusyn, Bulgarian as well as Bosnian and Croatian languages. Petnica Science Center is a notable institution for extracurricular science education focusing on gifted students. There are 19 universities in Serbia (nine public universities with a total number of 86 faculties and ten private universities with 51 faculties). In 2018/2019 academic year, 210,480 students attended 19 universities (181,310 at public universities and some 29,170 at private universities) while 47,169 attended 81 "higher schools". Public universities in Serbia are: the University of Belgrade (oldest, founded in 1808, and largest university with 97,696 undergraduates and graduates), University of Novi Sad (founded in 1960 and with student body of 42,489), University of Niš (founded in 1965; 20,559 students), University of Kragujevac (founded in 1976; 14,053 students), University of Priština (located in North Mitrovica), Public University of Novi Pazar as well as three specialist universities – University of Arts, University of Defence and University of Criminal Investigation and Police Studies. Largest private universities include Megatrend University and Singidunum University, both in Belgrade, and Educons University in Novi Sad. The University of Belgrade (placed in 301–400 bracket on 2013 Shanghai Ranking of World Universities, being best-placed university in Southeast Europe after those in Athens and Thessaloniki) and University of Novi Sad are generally considered the best institutions of higher learning in the country. Serbia spent 0.9% of GDP on scientific research in 2017, which is slightly below the European average. Serbia was ranked 53rd in the Global Innovation Index in 2023, up from 57th in 2019. Since 2018, Serbia is a full member of CERN. Serbia has a long history of excellence in maths and computer sciences which has created a strong pool of engineering talent, although economic sanctions during the 1990s and chronic underinvestment in research forced many scientific professionals to leave the country. Nevertheless, there are several areas in which Serbia still excels such as growing information technology sector, which includes software development as well as outsourcing. It generated over $1.2 billion in exports in 2018, both from international investors and a significant number of dynamic homegrown enterprises. Serbia is one of the countries with the highest proportion of women in science. Among the scientific institutes operating in Serbia, the largest are the Mihajlo Pupin Institute and Vinča Nuclear Institute, both in Belgrade. The Serbian Academy of Sciences and Arts is a learned society promoting science and arts from its inception in 1841. For centuries straddling the boundaries between East and West, the territory of Serbia had been divided among the Eastern and Western halves of the Roman Empire; then between Byzantium and the Kingdom of Hungary; and in the early modern period between the Ottoman Empire and the Habsburg Empire. These overlapping influences have resulted in cultural varieties throughout Serbia; its north leans to the profile of Central Europe, while the south is characteristic of the wider Balkans and even the Mediterranean. The Byzantine influence on Serbia was profound, first through the introduction of Eastern Christianity in the Early Middle Ages. The Serbian Orthodox Church has many monasteries built in the Serbian Middle Ages. Serbia was influenced by the Republic of Venice as well, mainly though trade, literature and romanesque architecture. Serbia has five cultural monuments inscribed in the list of UNESCO World Heritage: the early medieval capital Stari Ras and the 13th-century monastery Sopoćani; the 12th-century Studenica monastery; the Roman complex of Gamzigrad–Felix Romuliana; medieval tombstones Stećci; and finally the endangered Medieval Monuments in Kosovo (the monasteries of Visoki Dečani, Our Lady of Ljeviš, Gračanica and Patriarchal Monastery of Peć). There are two literary works on UNESCO's Memory of the World Programme: the 12th-century Miroslav Gospel, and scientist Nikola Tesla's archive. The slava (patron saint veneration), kolo (traditional folk dance), singing to the accompaniment of the gusle, Zlakusa pottery and slivovitz (plum brandy) are inscribed on UNESCO Intangible Cultural Heritage Lists. The Ministry of Culture and Information is tasked with preserving the nation's cultural heritage and overseeing its development, with further activities undertaken by local governments. Traces of Roman and early Byzantine Empire architectural heritage are found in many royal cities and palaces in Serbia, such as Sirmium, Felix Romuliana and Justiniana Prima, since 535 the seat of the Archbishopric of Justiniana Prima. Serbian monasteries were under the influence of Byzantine Art, particularly after the fall of Constantinople in 1204 when many Byzantine artists fled to Serbia. The monasteries include Studenica (built around 1190), which was a model for such later monasteries as Mileševa, Sopoćani, Žiča, Gračanica and Visoki Dečani. Numerous monuments and cultural sites were destroyed at various stages of Serbian history, including destruction in Kosovo. In the late 14th and the 15th centuries, an autochthonous architectural style known as Morava style evolved in the area around Morava Valley. A characteristic of this style was the wealthy decoration of the frontal church walls. Examples of this include Manasija, Ravanica and Kalenić monasteries. Frescos include White Angel (Mileševa monastery), Crucifixion (Studenica monastery) and Dormition of the Virgin (Sopoćani). The country is dotted with many well-preserved medieval fortifications and castles such as Smederevo Fortress (largest lowland fortress in Europe), Golubac, Maglič, Soko grad, Belgrade Fortress, Ostrvica and Ram. Under Ottoman occupation, Serbian art was virtually non-existent outside the lands ruled by the Habsburg monarchy. Traditional Serbian art showed Baroque influences at the end of the 18th century as shown in the works of Nikola Nešković, Teodor Kračun, Zaharije Orfelin and Jakov Orfelin. Serbian painting showed the influence of Biedermeier and Neoclassicism as seen in works by Konstantin Danil, Arsenije Teodorović and Pavel Đurković. Many painters followed the artistic trends set in the 19th century Romanticism, notably Đura Jakšić, Stevan Todorović, Katarina Ivanović and Novak Radonić. Serbian painters of the first half of the 20th century include Paja Jovanović and Uroš Predić of Realism, Cubist Sava Šumanović, Milena Pavlović-Barili and Nadežda Petrović of Impressionism, Expressionist Milan Konjović. Painters of the second half of 20th century include Marko Čelebonović, Petar Lubarda, Milo Milunović, Ljubomir Popović and Vladimir Veličković. Anastas Jovanović was one of the earliest photographers in the world. Marina Abramović is a performance artist. Pirot carpet is a traditional handicraft in Serbia. There are around 180 museums in Serbia, including the National Museum of Serbia, founded in 1844, houses one of the largest art collections in the Balkans, including many foreign pieces. Other art museums include the Museum of Contemporary Art in Belgrade, the Museum of Vojvodina and the Gallery of Matica Srpska in Novi Sad. Serbian uses the Cyrillic alphabet created by the students of the brothers Cyril and Methodius at the Preslav Literary School in Bulgaria. Serbian works from the early 11th century are written in Glagolitic. Starting in the 12th century, books were written in Cyrillic. The Miroslav Gospels from 1186 are considered to be the oldest book of Serbian medieval history and are listed in UNESCO's Memory of the World Register. Medieval authors include Saint Sava, Jefimija, Stefan Lazarević, Constantine of Kostenets and others. Under Ottoman occupation, when Serbia was not part of the European Renaissance, the tradition of oral story-telling through epic poetry was inspired by the Kosovo battle and folk tales rooted in Slavic mythology. Serbian epic poetry in those times was seen as the most effective way in preserving the national identity. The oldest known, entirely fictional poems, make up the Non-historic cycle, which is followed by poems inspired by events before, during and after the Battle of Kosovo. Some cycles are dedicated to Serbian legendary hero, Marko Kraljević, others are about hajduks and uskoks, and the last one is dedicated to the liberation of Serbia in the 19th century. Folk ballads include The Death of the Mother of the Jugović Family and The Mourning Song of the Noble Wife of the Asan Aga (1646), translated into European languages by Goethe, Walter Scott, Pushkin and Mérimée. A tale from Serbian folklore is The Nine Peahens and the Golden Apples. Baroque trends in Serbian literature emerged in the late 17th century. Baroque-influenced authors include Gavril Stefanović Venclović, Jovan Rajić, Zaharije Orfelin and Andrija Zmajević. Dositej Obradović was a prominent figure of the Age of Enlightenment, while Jovan Sterija Popović was a Classicist writer whose works also contained elements of Romanticism. In the era of national revival, in the first half of the 19th century, Vuk Stefanović Karadžić collected Serbian folk literature, and reformed the Serbian language and spelling, paving the way for Serbian Romanticism. The first half of the 19th century was dominated by Romanticist writers, including Petar II Petrović-Njegoš, Branko Radičević, Đura Jakšić, Jovan Jovanović Zmaj and Laza Kostić, while the second half of the century was marked by Realist writers such as Milovan Glišić, Laza Lazarević, Simo Matavulj, Stevan Sremac, Vojislav Ilić, Branislav Nušić, Radoje Domanović and Borisav Stanković. The 20th century was dominated by the prose writers Meša Selimović (Death and the Dervish), Miloš Crnjanski (Migrations), Isidora Sekulić (The Chronicle of a Small Town Cemetery), Branko Ćopić (Eagles Fly Early), Borislav Pekić (The Time of Miracles), Danilo Kiš (The Encyclopedia of the Dead), Dobrica Ćosić (The Roots), Aleksandar Tišma (The Use of Man), Milorad Pavić and others. Notable poets include Milan Rakić, Jovan Dučić, Vladislav Petković Dis, Rastko Petrović, Stanislav Vinaver, Dušan Matić, Branko Miljković, Vasko Popa, Oskar Davičo, Miodrag Pavlović, and Stevan Raičković. Pavić is a 21st-century Serbian author whose Dictionary of the Khazars has been translated into 38 languages. Contemporary authors include David Albahari, Svetislav Basara, Goran Petrović, Gordana Kuić, Vuk Drašković and Vladislav Bajac. Serbian comics emerged in the 1930s and the medium remains popular today. Ivo Andrić (The Bridge on the Drina) is a Serbian author who won the Nobel Prize in Literature in 1961. Another writer was Desanka Maksimović, who for seven decades was the leading lady of Yugoslav poetry. She is honoured with statues, postage stamps, and the names of streets across Serbia. There are 551 public libraries, the largest of which are: the National Library of Serbia in Belgrade with about 6 million items, and Matica Srpska (the oldest matica and Serbian cultural institution, founded in 1826) in Novi Sad with nearly 3.5 million volumes. In 2010, there were 10,989 books and brochures published. The book publishing market is dominated by several major publishers such as Laguna and Vulkan (both of which operate their own bookstore chains) and the industry's centrepiece event, annual Belgrade Book Fair, is the most visited cultural event in Serbia with 158,128 visitors in 2013. The highlight of the literary scene is awarding of NIN Prize, given every January since 1954 for the best newly published novel in Serbian. Composer and musicologist Stevan Stojanović Mokranjac is considered the founder of modern Serbian music. The Serbian composers of the first generation Petar Konjović, Stevan Hristić, and Miloje Milojević maintained the national expression and modernised the romanticism into the direction of impressionism. Other famous classical Serbian composers include Isidor Bajić, Stanislav Binički and Josif Marinković. There are three opera houses in Serbia: Opera of the National Theatre and Madlenianum Opera, both in Belgrade, and Opera of the Serbian National Theatre in Novi Sad. Four symphonic orchestra operate in the country: Belgrade Philharmonic Orchestra, Niš Symphony Orchestra, Novi Sad Philharmonic Orchestra and Symphonic Orchestra of Radio Television of Serbia. The Choir of Radio Television of Serbia is a leading vocal ensemble in the country. The BEMUS is one of the most prominent classical music festivals in the Southeastern Europe. Traditional Serbian music includes various kinds of bagpipes, flutes, horns, trumpets, lutes, psalteries, drums and cymbals. The kolo is the traditional collective folk dance, which has a number of varieties throughout the regions. The most popular are those from Užice and Morava region. Sung epic poetry has been an integral part of Serbian and Balkan music for centuries. In the highlands of Serbia these long poems are typically accompanied on a one-string fiddle called the gusle, and concern themselves with themes from history and mythology. There are records of gusle being played at the court of the 13th-century King Stefan Nemanjić. Pop music artist Željko Joksimović won second place at the 2004 Eurovision Song Contest and Marija Šerifović won the 2007 Eurovision Song Contest with the song "Molitva", and Serbia was the host of the 2008 edition of the contest. Pop singers include Zdravko Čolić, Vlado Georgiev, Aleksandra Radović, Jelena Tomašević and Nataša Bekvalac, among others. Serbian rock was part of the former Yugoslav rock scene during the 1960s, 1970s and 1980s. During the 1990s and 2000s, the popularity of rock music declined in Serbia, and although several major mainstream acts managed to sustain their popularity, an underground and independent music scene developed. The 2000s saw a revival of the mainstream scene and the appearance of a large number of notable acts. Serbian rock acts include Atheist Rap, Bajaga i Instruktori, Đorđe Balašević, Bjesovi, Block Out, Crni Biseri, Darkwood Dub, Disciplina Kičme, Elipse, Ekatarina Velika, Električni Orgazam, Eva Braun, Galija, Generacija 5, Goblini, Idoli, Kanda, Kodža i Nebojša, Kerber, Korni Grupa, Laboratorija Zvuka, Slađana Milošević, Neverne Bebe, Obojeni Program, Orthodox Celts, Partibrejkers, Pekinška Patka, Piloti, Riblja Čorba, Ritam Nereda, Rambo Amadeus, S.A.R.S., Siluete, S Vremena Na Vreme, Šarlo Akrobata, Pop Mašina, Smak, U Škripcu, Van Gogh, YU Grupa, Zana and others. Folk music in its original form has been a prominent music style since World War I following the early success of Sofka Nikolić. The music has been further promoted by Danica Obrenić, Anđelija Milić, Nada Mamula, and during the 60s and 70s with performers like Silvana Armenulić, Toma Zdravković, Lepa Lukić, Vasilija Radojčić, Vida Pavlović and Gordana Stojićević. Turbo-folk music is a subgenre that was developed in Serbia in the late 1980s and the beginning of the 1990s and has since enjoyed an immense popularity through acts of Dragana Mirković, Zorica Brunclik, Šaban Šaulić, Ana Bekuta, Sinan Sakić, Vesna Zmijanac, Mile Kitić, Snežana Đurišić, Šemsa Suljaković, and Nada Topčagić. It is a blend of folk music with pop and dance elements and can be seen as a result of the urbanisation of folk music. In recent years, turbo-folk has featured even more pop music elements, and some of the performers have been labeled as pop-folk. The most famous among them are Ceca (often considered to be the biggest music star of Serbia), Jelena Karleuša, Aca Lukas, Seka Aleksić, Dara Bubamara, Indira Radić, Saša Matić, Viki Miljković, Stoja and Lepa Brena, arguably the most prominent performer of former Yugoslavia. Balkan Brass, or truba ("trumpet") is a popular genre, especially in Central and Southern Serbia where Balkan Brass originated. The music has its tradition from the First Serbian Uprising. The trumpet was used as a military instrument to wake and gather soldiers and announce battles, and it took on the role of entertainment during downtime, as soldiers used it to transpose popular folk songs. When the war ended and the soldiers returned to the rural life, the music entered civilian life and eventually became a music style, accompanying births, baptisms, weddings, and funerals. There are two main varieties of this genre, one from Western Serbia and the other from Southern Serbia, with brass musician Boban Marković being one of the most respected names in the world of modern brass band bandleaders. The most popular music festivals are Guča Trumpet Festival, with over 300,000 annual visitors, and EXIT in Novi Sad (won the Best Major Festival award at the European Festivals Awards for 2013 and 2017.), with 200,000 visitors in 2013. Other festivals include Nišville Jazz Festival in Niš and Gitarijada rock festival in Zaječar. Serbia has a well-established theatrical tradition with Joakim Vujić considered the founder of modern Serbian theatre. Serbia has 38 professional theatres and 11 theatres for children, the most important of which are National Theatre in Belgrade, Serbian National Theatre in Novi Sad, National Theatre in Subotica, National Theatre in Niš and Knjaževsko-srpski teatar in Kragujevac (the oldest theatre in Serbia, established in 1835). The Belgrade International Theatre Festival – BITEF, founded in 1967, is one of the oldest theatre festivals in the world, and it has become one of the five biggest European festivals. Sterijino pozorje is, on the other hand, a festival showcasing national drama plays. The most important Serbian playwrights were Jovan Sterija Popović and Branislav Nušić, while recent renowned names are Dušan Kovačević and Biljana Srbljanović. The foundation of Serbian cinema dates back to 1896. The first Serbian feature film, titled The Life and Deeds of the Immortal Leader Karađorđe, was released in 1911. Serbia's film scene is one of the most dynamic smaller European cinemas. Serbia's film industry is heavily subsidized by the government, mainly through grants approved by the Film Centre of Serbia. As of 2019, there were 26 feature films produced in Serbia, of which 14 were domestic films. There are 23 operating cinemas in the country, of which 13 are multiplexes (all but two belonging to either Cineplexx or CineStar chains), with total attendance reaching 4.8 million. A comparatively high percentage of 20% of total tickets sold were for domestic films. Modern PFI Studios located in Šimanovci is nowadays Serbia's only major film studio complex; it consists of 9 sound stages and attracts mainly international productions, primarily American and West European. The Yugoslav Film Archive used to be former Yugoslavia's and now is Serbia's national film archive – with over 100 thousand film prints, it is among the five largest film archives in the world. Famous Serbian filmmaker Emir Kusturica won two Palmes d'Or for Best Feature Film at the Cannes Film Festival, for When Father Was Away on Business in 1985 and then again for Underground in 1995; he has also won a Silver Bear at the Berlin Film Festival for Arizona Dream and a Silver Lion at the Venice Film Festival for Black Cat, White Cat. Other renowned directors include Dušan Makavejev, Želimir Žilnik (Golden Berlin Bear winner), Aleksandar Petrović, Živojin Pavlović, Goran Paskaljević, Goran Marković, Srđan Dragojević, Srdan Golubović and Mila Turajlić among others. Serbian-American screenwriter Steve Tesich won the Academy Award for Best Original Screenplay in 1979 for the movie Breaking Away. Prominent movie stars in Serbia have left a celebrated heritage in the cinematography of Yugoslavia as well. Notable mentions are Zoran Radmilović, Pavle Vuisić, Ljubiša Samardžić, Olivera Marković, Mija Aleksić, Miodrag Petrović Čkalja, Ružica Sokić, Velimir Bata Živojinović, Danilo Bata Stojković, Seka Sablić, Olivera Katarina, Dragan Nikolić, Mira Stupica, Nikola Simić, Bora Todorović and others. Milena Dravić was one of the most celebrated actresses in Serbian cinematography, winning the Best Actress Award at the Cannes Film Festival in 1980. Freedom of the press and freedom of speech are guaranteed by the constitution of Serbia. Serbia is ranked 90th out of 180 countries in the 2019 Press Freedom Index report compiled by Reporters Without Borders. The report noted that media outlets and journalists continue to face partisan and government pressure over editorial policies. Also, the media are now more heavily dependent on advertising contracts and government subsidies to survive financially. According to EBU research in 2018, Serbs on average watch five and a half hours of television per day, making it the second highest average in Europe. There are seven nationwide free-to-air television channels, with public broadcaster Radio Television of Serbia (RTS) operating three (RTS1, RTS2 and RTS3) and private broadcasters operating four (Pink, Prva, Happy, and O2). In 2019, preferred usage of these channels was as follows: 19.3% for RTS1, 17.6% for Pink, 10.5% for Prva, 6.9% for Happy, 4.1% for O2, and 1.6% for RTS2. There are 28 regional television channels and 74 local television channels. Besides terrestrial channels there are dozens of Serbian television channels available only on cable or satellite. These include regional news N1, commercial channel Nova S, and regional sports channels Sport Klub and Arena Sport, among others. There are 247 radio stations in Serbia. Out of these, six are radio stations with national coverage, including two of public broadcaster Radio Television of Serbia (Radio Belgrade 1 and Radio Belgrade 2/Radio Belgrade 3) and four private ones (Radio S1, Radio S2, Play Radio, and Radio Hit FM). Also, there are 34 regional stations and 207 local stations. There are 305 newspapers published in Serbia of which 12 are daily newspapers. Dailies Politika and Danas are Serbia's papers of record, the former being the oldest newspaper in the Balkans, founded in 1904. Highest circulation newspapers are tabloids Večernje Novosti, Blic, Kurir, and Informer, all with more than 100,000 copies sold. There is one daily newspaper devoted to sports (Sportski žurnal), one business daily (Privredni pregled), two regional newspapers (Dnevnik published in Novi Sad and Narodne novine from Niš), and one minority-language daily (Magyar Szo in Hungarian, published in Subotica). There are 1,351 magazines published in the country. These include: weekly news magazines NIN, Vreme and Nedeljnik; popular science magazine Politikin Zabavnik; women's magazine Lepota & Zdravlje; auto magazine SAT revija; and IT magazine Svet kompjutera. In addition, there is a wide selection of Serbian editions of international magazines, such as Cosmopolitan, Elle, Men's Health, National Geographic, Le Monde diplomatique, Playboy, and Hello!, among others. The main news agencies are Tanjug, Beta and Fonet. As of 2017, out of 432 web-portals (mainly on the .rs domain) the most visited are online editions of printed dailies Blic and Kurir, news web-portal B92 and classifieds KupujemProdajem. Serbian cuisine is largely heterogeneous in a way characteristic of the Balkans and, especially, the former Yugoslavia. It features foods characteristic of lands formerly under Turkish suzerainty as well as cuisine originating from other parts of Central Europe (especially Austria and Hungary). Food is very important in Serbian social life, particularly during religious holidays such as Christmas, Easter and feast days i.e. slava. Staples of the Serbian diet include bread, meat, fruits, vegetables, and dairy products. Bread plays an important role in Serbian cuisine and can be found in religious rituals. A traditional Serbian welcome is to offer bread and salt to guests. Meat is widely consumed, as is fish. The southern Serbian city of Leskovac is host to Roštiljijada, a yearly grilled meat barbecue-based festival that is considered the biggest barbecue festival in the Balkans. Other Serbian specialties include ćevapčići (grilled and seasoned caseless sausages made from minced meat), pljeskavica (grilled spiced meat patty made from a mixture of pork, beef and lamb), gibanica (cheese pie), burek (baked pastry made from a thin flaky dough that is stuffed with meat, cheese or vegetables), sarma (stuffed cabbage), punjena paprika (stuffed pepper), moussaka (casserole made from minced meat, eggs, and potatoes), Karađorđeva šnicla (veal or pork schnitzel that is stuffed with kajmak), đuveč (meat and vegetable stew), pasulj (bean soup), podvarak (roast meat with sauerkraut), ajvar (roasted red pepper spread), kajmak (dairy product similar to clotted cream), čvarci (variant of pork rinds), proja (cornbread) and kačamak (corn-flour porridge). Serbians claim their country as the birthplace of rakia (rakija), a highly alcoholic drink primarily distilled from fruit. Rakia in various forms is found throughout the Balkans, notably in Bulgaria, Croatia, Slovenia, Montenegro, Hungary and Turkey. Slivovitz (šljivovica), a plum brandy, is a type of rakia which is considered the national drink of Serbia. In 2021, Serbia's sljivovica was added to the United Nations Intangible Cultural Heritage List as a "cherished tradition to be preserved by humanity". Winemaking traditions in Serbia dates back to Roman times. Serbian wines are produced in 22 different geographical regions, with white wine dominating the total amount. Besides rakia and wine, beer is a very popular alcoholic beverage in the country. Pale lagers are currently and have been the traditional beer choice for Serbians. Meanwhile, dark lagers, while still being popular, are produced and consumed in much smaller quantities. The most popular domestic brands of beer are Jelen, followed by Lav, which are both pale lagers. As in the rest of the former Yugoslavia, coffee drinking is an important cultural and social practice and Serbian coffee (a local variant of Turkish coffee) is the most commonly consumed non-alcoholic beverage in Serbia. Sports play an important role in Serbian society, and the country has a strong sporting history. The most popular sports in Serbia are football, basketball, tennis, volleyball, water polo and handball. Professional sports in Serbia are organised by sporting federations and leagues (in the case of team sports). One of the particularities of Serbian professional sports is the existence of many multi-sport clubs (called "sports societies"), the biggest and most successful of which are Red Star, Partizan, and Beograd in Belgrade; Vojvodina in Novi Sad; Radnički in Kragujevac; and Spartak in Subotica. Football is the most popular sport in Serbia, and the Football Association of Serbia with 146,845 registered players, is the largest sporting association in the country. Dragan Džajić was officially recognised as "the best Serbian player of all time" by the Football Association of Serbia, and more recently the likes of Nemanja Vidić, Dejan Stanković, Branislav Ivanović, Aleksandar Kolarov and Nemanja Matić play for the elite European clubs, developing the nation's reputation as one of the world's biggest exporters of footballers. The Serbia national football team lacks relative success although it qualified for three of the last four FIFA World Cups. The two main football clubs in Serbia are Red Star (winner of the 1991 European Cup) and Partizan (a finalist at the 1966 European Cup), both from Belgrade. The rivalry between the two clubs is known as the "Eternal Derby", and is often cited as one of the most exciting sports rivalries in the world. Serbia is one of the traditional powerhouses of world basketball, as Serbia men's national basketball team have won two World Championships (in 1998 and 2002), three European Championships (1995, 1997, and 2001) and two Olympic silver medals (in 1996 and 2016) as well. The women's national basketball team have won two European Championships (2015, 2021) and an Olympic bronze medal in 2016. A total of 31 Serbian players have played in the NBA in the last three decades, including Nikola Jokić (2020–21, 2021–22 NBA MVP, 2023 NBA champion, 2023 NBA Finals MVP and five-time NBA All-Star), Predrag "Peja" Stojaković (2011 NBA champion and three-time NBA All-Star), and Vlade Divac (2001 NBA All-Star and Basketball Hall of Famer). The renowned "Serbian coaching school" produced many of the most successful European basketball coaches of all time, such as Željko Obradović (who won a record 9 Euroleague titles as a coach), Dušan Ivković, Svetislav Pešić, and Igor Kokoškov (the first coach born and raised outside of North America to be hired as a head coach in the NBA). KK Partizan basketball club was the 1992 European champion. The Serbia men's national water polo team is one of the most successful national teams, having won an Olympic gold medal in 2016 and 2020, three World Championships (2005, 2009 and 2015), and seven European Championships (2001, 2003, 2006, 2012, 2014, 2016 and 2018). VK Partizan has won a joint-record seven European champion titles. The recent success of Serbian tennis players has led to an immense growth in the popularity of tennis in the country. Novak Djokovic has won an all-time record of 24 Grand Slam singles titles and has held the No. 1 spot in the ATP rankings for a record duration. He became the eighth player in history to achieve the Career Grand Slam, the third man to hold all four major titles at once, the first ever to do so on three different surfaces, and the first ever to achieve a triple Career Grand Slam. Ana Ivanovic (champion of 2008 French Open) and Jelena Janković were both ranked No. 1 in the WTA rankings. There were two No. 1 ranked-tennis double players as well: Nenad Zimonjić (three-time men's double and four-time mixed double Grand Slam champion) and Slobodan Živojinović. The Serbia men's tennis national team won the 2010 Davis Cup and 2020 ATP Cup, while Serbia women's tennis national team reached the final at 2012 Fed Cup. Serbia is one of the leading volleyball countries in the world. Its men's national team won the gold medal at the 2000 Olympics, the European Championship three times, as well as the 2016 FIVB World League. The women's national volleyball team are current world Champions, have won European Championship three times (2011, 2017 and 2019), as well as an Olympic silver medal in 2016. Jasna Šekarić, sport shooter, is one of the athletes with the most appearances at the Olympic Games. She has won a total of five Olympic medals and three World Championship gold medals. Other noted Serbian athletes include: swimmers Milorad Čavić (2009 World championships gold and silver medalist as well as 2008 Olympic silver medalist on 100-metre butterfly in historic race with American swimmer Michael Phelps) and Nađa Higl (2009 World champion in 200-metre breaststroke); track and field athletes Vera Nikolić (former world record holder in 800 metres) and Ivana Španović (long-jumper; four-time European champion, World indoor champion and bronze medalist at the 2016 Olympics); wrestler Davor Štefanek (2016 Olympic gold medalist and 2014 World champion), and taekwondoist Milica Mandić (2012 Olympic gold medalist and 2017 world champion). Serbia has hosted several major sport competitions, including the 2005 Men's European Basketball Championship, 2005 Men's European Volleyball Championship, 2006 and 2016 Men's European Water Polo Championships, 2009 Summer Universiade, 2012 European Men's Handball Championship, and 2013 World Women's Handball Championship. The most important annual sporting events held in the country are the Belgrade Marathon and the Tour de Serbie cycling race. 44°N 21°E / 44°N 21°E / 44; 21
{{refimprove|date=October 2014}} [[File:What is cancer.webm|thumb|start=7.2|end=438|Lecture by Dr. Noel de Miranda about cancer]] '''Cancer''' is a type of [[disease]] where [[cells]] grow out of control, divide and invade other [[tissues]]. In a person without cancer, [[cell division]] is under control. In most [[tissues]], healthy cells divide in a controlled way and copy themselves to create new healthy cells. With cancer, this normal cell division goes out of control. Cells change their nature because [[mutations]] have occurred in their [[genes]]. All the daughter cells of cancer cells are also cancerous. Such cells are responsible for growing more cancer cells in the body. If the abnormal cells do not invade other tissues or organs, but just divide and swell up their original tissue, this is not called "cancer". It is called a [[tumour]]. Tumours are usually not a threat to life because they can be cut out. However, some tumours occur in places where they cannot be cut out, and they can be fatal. Some [[brain tumor]]s are of this type. The [[symptoms]] of cancer are caused by the cancerous cells invading other tissues. This is called [[metastasis]]. Metastasis is when cancer cells move through the [[bloodstream]] or [[lymphatic system]]. When this happens, a person's cancer can be spread to other places in the body. Eventually those other tissues cannot work as well, and the whole body begins to get worse, and may die. Cancer can affect anybody at any age. Most types of cancer are more likely to affect people as they get older. This is because as a person's [[DNA]] gets older, their DNA may become damaged, or damage that happened in the past may get worse. One type of cancer that is more common in young men, rather than older people, is [[testicular cancer]] (cancer of the [[testicles]]). Cancer is one of the biggest and most researched causes of death in [[developed country|developed countries]]. Studying cancer and its treatment is called [[oncology]]. == Causes == Cancer is one of the most common causes of death around the world. It causes about 16% (or one out of every six) of all deaths worldwide, according to the [[World Health Organization]].<ref name="WHOfact">{{cite web |title=Cancer |url=https://www.who.int/en/news-room/fact-sheets/detail/cancer |website=[[World Health Organization]] |accessdate=2022-10-20 |date=2022-02-03}}</ref> Different types of cancer have different causes. Some things are known to cause cancer in a specific body part; other things are known to be able to cause many different types of cancer. For example, using [[tobacco]] (smoked or smokeless) can cause many types of cancers, such as [[Lung cancer|lung]], [[Oral cancer|mouth, tongue, and throat cancers]].<ref>{{cite web | title = Marijuana Damages DNA And May Cause Cancer, New Test Reveals | url = http://www.sciencedaily.com/releases/2009/06/090615095940.htm | publisher = ScienceDaily.com | date = 2009-06-15 }}</ref> Other things that are known to be able to cause cancer - or make a person more likely to get cancer - include: [[radiation]] including [[sunlight]] and [[X-ray]]s in large or many doses, and [[wikt:exposure|exposure]] to radiation (for example in a [[nuclear power plant]]); [[chemical]]s and materials used in [[construction|building]] and [[manufacturing]] (for example, [[asbestos]] and [[benzene]]); high-[[fat]] or low-fiber diets; air and water [[pollution]]; eating very little [[fruit]]s and [[vegetable]]s; [[obesity]]; not enough [[wikt:physical|physical]] activity; drinking too much [[alcohol]]; and certain [[chemicals]] commonly used at home. Some cancers can also be caused by [[virus]]es. Many people who are exposed to these things do get cancer - but some do not. == Kinds == There are many different kinds of cancers. The most common cancers in the world are: [[breast cancer]], [[lung cancer]], [[colorectal cancer]], [[prostate cancer]] and [[stomach cancer]].<ref name="WHOfact"/> == Treatment of cancer == There is no sure cure for cancer. It can only be cured if all of the cancerous cells are cut out or killed in place. This means that the earlier the cancer is treated, the better the chances are for a cure (because the cancer cells may not have had enough time to copy themselves and spread so much that the person cannot be cured). There are a few different types of treatments that may kill cancer cells. These treatments are: * [[Radiotherapy]] (radiation therapy), which uses [[radiation]] to kill cancer cells * [[Chemotherapy]], which uses strong medications to kill cancer cells * [[Surgery]] to take out part or all of a tumor. After surgery, many patients may need radiotherapy or chemotherapy to keep the tumor from growing again. * [[Immunotherapy]] works by "inducing, enhancing, or suppressing an [[immune response]]".<ref>{{cite web |url=http://dictionary.reference.com/browse/immunotherapies?qsrc=2446 |title=immunotherapies definition |work=Dictionary.com |accessdate=2009-06-02}}</ref> === Treating cancer is complicated === There are a few reasons why treating cancer is complicated. For example: * Most things that kill cancer cells also kill normal, healthy [[cell]]s. This can cause many [[Adverse effect (medicine)|side effects]], like hair loss and [[vomiting]]. * The body's [[immune system]] usually will not attack cancer cells, even though they could easily kill the body. This is because the cancer has actually become a part of the body by invading cells and tissues. So the immune system sees the cancer as part of the body it is trying to protect, not as a threat to be attacked. * There are many different types of cancer, and each has its own symptoms and causes. Even with the same type of cancer, different people may have different symptoms, and may react to treatments differently; their cancer also may grow or spread at different speeds. Treatment has to be a good fit to both the type of cancer and the individual [[patient]] who has the cancer. Many people in countries around the world study cancer and work on finding treatments. There has been some good progress in finding treatments, and many kinds of cancers are treated with success. Along with looking for different medical treatments to treat cancer, some studies also look for things that people with cancer can do themselves to try to make themselves healthier. For example, one study showed that if a person with lymphedema (a [[Inflammation|swelling]] of the arm linked to breast cancer) lifts weights, he may be able to fight his cancer better than somebody who does not lift weights. == History == Cancer has been around for many thousands of years. Today, a lot of the medical terms used to describe cancer come from [[ancient Greek]] and [[Latin]]. For example, the Latinized Greek word [[carcinoma]] is used to describe a [[malignant]] [[tumor]] - a tumor made up of cancer cells. The Greeks also used the word "karkinos", which would be translated by Aulus Cornelius Celsus into the [[Latin]] word ''cancer''. The [[prefix]] 'carcino' is still used in medical words like carcinoma and [[carcinogenic]]. A famous Greek doctor, [[Galen]], helped create another word that is very important to medicine today by using the word "''onkos''" to [[wikt:describe|describe]] ''all'' tumours. This is where the word [[oncology]], the branch of medicine that deals with cancer, comes from.<ref name="Galen">{{cite journal |author=Karpozilos A, Pavlidis N |title=The treatment of cancer in Greek antiquity |journal=European Journal of Cancer |volume=40 |issue=14 |pages=2033–40 |year=2004 |pmid=15341975 |doi=10.1016/j.ejca.2004.04.036 |issn=0959-8049 }}</ref> === Ancient history === [[Hippocrates]] (a very famous ancient doctor who is often called the father of modern medicine) named many kinds of cancer. He called [[benign tumour]]s (tumors that are not made up of cancer cells) ''oncos''. In [[Greek language|Greek]], ''onkos'' means 'swelling'. He called malignant tumours ''karkinos''. This means [[crab]] or [[crayfish]] in Greek. He used this term because he thought that if a solid malignant tumor was cut into, its veins looked like a crab: "the [[vein]]s stretched on all sides as the animal the crab has its feet, whence it derives (gets) its name".<ref>{{cite web |first = Ralph W. |last = Moss |title = Galen on Cancer |url = http://www.cancerdecisions.com/speeches/galen1989.html |publisher = CancerDecisions |year = 2004 |access-date = 2013-02-27 |archive-date = 2011-07-16 |archive-url = https://web.archive.org/web/20110716111312/http://www.cancerdecisions.com/speeches/galen1989.html |url-status = dead }}</ref> Hippocrates later added ''-oma'' (Greek for 'swelling') after the word 'carcinos'. This is how the word ''carcinoma'' came about. Because the [[Ancient Greece|ancient Greeks]] did not believe in [[Autopsy|opening up dead bodies to study them]], Hippocrates was only able to describe and make drawings of tumors he saw from the outside of the body. He drew tumors that had been on the skin, nose, and breasts. Hippocrates and other doctors at that time treated people based on the [[Humorism|humor theory]]. This theory said that there were four types of fluid in the body (black, yellow bile, blood, and phlegm). Doctors tried to figure out whether these four "humors" (or body fluids) were in balance. They would then use treatments like [[blood-letting]] (cutting the patient and letting him [[Bleeding|bleed]] so that he would lose blood); [[laxative]]s (giving the patient foods or herbs to make him [[Defecation|go to the bathroom]]), and/or changing the patient's diet. The doctors thought that these treatments would work to get the patient's four humors back into the right balance. The humor theory treatment was popular until the 19th century (the 1800s), when [[cell (biology)|cells]] were discovered. By this time, people had realized that cancer can happen anywhere in the body. === Early surgery === The oldest known document that talks about cancer was discovered in [[Egypt]] and is thought to be from about 1600 B.C. The document talks about using surgery to treat eight cases of [[ulcer]]s of the breast. These were treated by cauterization - by burning them - using a tool called "the fire drill". The document also says about cancer, "There is no treatment".<ref name="CancerOrgHistory">{{cite web|date=September 2009|title=The History of Cancer|url=http://www.cancer.org/docroot/CRI/content/CRI_2_6x_the_history_of_cancer_72.asp|publisher=American Cancer Society|access-date=2010-03-14|archive-date=2010-03-02|archive-url=https://web.archive.org/web/20100302154211/http://www.cancer.org/docroot/CRI/content/CRI_2_6x_the_history_of_cancer_72.asp|url-status=dead}}</ref> Another very early type of surgery used to treat cancer was written about in the 1020s. In ''The Canon of Medicine'', [[Avicenna]] (Ibn Sina) said that treatment should involve cutting out all diseased tissue. This included the use of [[amputation]] (removing a part of the body completely) or removing veins that ran in the direction of the tumor. Avicenna also suggested that the area that had been treated should be cauterized (or burned) if needed.<ref name=Patricia>Patricia Skinner (2001), [https://archive.today/20120629144443/findarticles.com/p/articles/mi_g2603/is_0007/ai_2603000716 Unani-tibbi], ''Encyclopedia of Alternative Medicine''</ref> === The 16th and 17th centuries === In the [[16th century|16th]] and [[17th century|17th centuries]] (the 1500s and 1600s), doctors started to be allowed to [[autopsy|dissect bodies]] (or cut them open after death) in order to figure out the cause of death. Around this time, there were many different ideas about what caused cancer. The German professor Wilhelm Fabry believed that breast cancer was caused by a clot of milk in the part of a woman's breast that produces milk. The Dutch professor Francois de la Boe Sylvius believed that all disease was caused by chemical processes. He thought that cancer, in particular, was caused by acidic [[lymph]]. Nicolaes Tulp, who lived at the same time as Sylvius, believed that cancer was a poison that slowly spreads and was [[infectious disease|contagious]].<ref name="Marilyn Yalom">Marilyn Yalom "A history of the breast" 1997. New York: Alfred A. Knopf. {{ISBN|0-679-43459-3}}</ref> A British surgeon named Percivall Pott was the first person to figure out one of the real causes of cancer. In 1775, he discovered that cancer of the [[scrotum]] was a common disease among [[chimney sweep]]s (people who cleaned out [[chimney]]s). Other doctors started studying this topic and coming up with other ideas about what causes cancer. Doctors then started working together and coming up with better ideas. === The 18th century === In the 18th century (the 1700s), many people started to use the [[microscope]], and this made a big difference in helping doctors and scientists understand more about cancer. Using the microscope, scientists were able to see that the 'cancer poison' spread from one tumor through the [[lymphatic system|lymph nodes]] to other sites ("[[metastasis]]"). This was first made clear by the English surgeon Campbell De Morgan, between 1871 and 1874.<ref>{{cite journal|author=Grange JM|last2=Stanford JL|last3=Stanford CA|year=2002|title=Campbell De Morgan's 'Observations on cancer', and their relevance today|journal=Journal of the Royal Society of Medicine|volume=95|issue=6|pages=296–9|doi=10.1177/014107680209500609|pmc=1279913|pmid=12042378}}</ref> Before the 19th century (the 1800s), using [[surgery]] to treat cancer usually had bad results. Doctors did not understand how important [[hygiene]] (or keeping things clean) is for preventing disease, especially after surgery. Because things were not kept clean during or after surgery, patients often got [[Infectious disease|infections]] and died. For example, one well-known Scottish surgeon, Alexander Monro, kept records and found that 58 patients out of every 60 who had surgery for breast tumors died within the next two years. === The 19th century === In the 19th century, surgical hygiene got better because of [[asepsis]]. Doctors realized that dirtiness and [[Pathogen|germs]] cause infections, so they started to keep things cleaner and do things to kill germs in order to prevent their patients from getting infections. It became more common for people to survive after having surgery. Surgical removal of the tumor (taking the tumor out of the body by doing surgery) became the first-choice treatment for cancer. For this kind of treatment to work, the surgeon doing the operation had to be very good at removing tumors. (This meant that even if people had the same kind of cancer, they could get very different results, with some getting good treatment that worked and others getting treatment that did not work, because of differences in how good different surgeons were.) In the late 1800s, doctors and scientists started to realize that the body is made up of many kinds of [[tissue (biology)|tissues]], which in turn are made up of millions of cells. The discovery started the age of cellular [[pathology]] (studying cells to learn about diseases and figure out what is wrong with the body). ==== Discovery of radiation ==== In the 1890s, French scientists discovered [[radioactive decay]]. [[Radiation therapy]] became the first cancer treatment that worked and did not involve surgery. It required a new [[multi-disciplinary]] approach to cancer treatment (people doing different jobs were working together to treat patients). The surgeon was no longer working by himself - he worked together with hospital [[Radiology|radiologists]] (people who gave and read [[X-ray]]s) to help patients. This team approach meant changes in how they worked. The different people on the team had to communicate with each other and work together, which they were not used to doing. It also meant that treatment had to be done in a hospital rather than at the patient's home. Because of this, patients' information had to be put together into files kept at the hospital (called "medical records"). Because this information was now being kept and written down, scientists were able to do the first [[statistics|statistical]] patient studies using numbers to study questions like how many people who have a certain type of cancer or get a certain treatment survive. === The 20th century === Another important step forward in understanding cancer happened in 1926, when Janet Lane-Claypon published a paper on cancer [[epidemiology]]. (Epidemiology is a field of study which looks at how common a disease is, what patterns the disease takes in different kinds of people, and what this means for understanding and treating the disease.) This historic paper was a [[wikt:comparative|comparative]] study, which tries to find out what causes a disease by looking at a group of people who have the disease and figuring out how they are different from another group that does not have the disease. Lane-Clayton's study looked at 1000 people who all had the same background and lifestyle (or way of living): 500 people with breast cancer and 500 control patients (people without breast cancer). These people were the same in many ways, but some got breast cancer and some did not. To figure out what might be causing certain people to get breast cancer, the study looked at what was different about these people when they were compared to (or looked at alongside) the people who did not get cancer. Lane-Clayton's study was published by the British Ministry of Health. Her work on cancer epidemiology was continued by Richard Doll and Austin Bradford Hill. They used the same ways of studying cancer as Lane-Clayton, but they looked at a different kind of cancer: [[lung cancer]]. In 1956, they published their results in a paper called "Lung Cancer and Other Causes of Death In Relation to Smoking. A Second Report on the Mortality of British Doctors" (also called the [[British doctors study]]). Later, Richard Doll left the London Medical Research Center (MRC), and started the Oxford unit for Cancer epidemiology in 1968. By using [[computer]]s, this unit was able to do something new and very important: it brought together large amounts of cancer data (pieces of information about cancer). This way of studying cancer is very important to cancer epidemiology today, and it has also been very important in shaping what we now know about cancer and what the rules and laws about the disease and [[public health]] are today. Over the past 50&nbsp;years, many different people have done a lot of work to collect data from different doctors, hospitals, areas, states, and even countries. This data is used to study whether different kinds of cancer are more or less common in different areas, [[environment]]s (for example, in big cities compared to the countryside), or [[culture]]s. This helps people who study cancer to figure out what makes people more or less likely to get different kinds of cancer. ==== Effects of World War II ==== Before World War II, doctors and hospitals were getting better at collecting (or getting and keeping) data about their patients who had cancer, but it was rare for this data to be shared with other doctors or hospitals. This changed after WWII, when medical research centers found out that different countries had very different number of cases of cancer. Because of this, many countries created national public health organizations (which studied public health issues in an entire country). These national public health organizations began to bring together health data from many different doctors and hospitals. This helped them figure out some of the reasons why cancer was so much more common in certain places. For example, in Japan, people studying cancer found out that people who had survived the [[atomic bombings of Hiroshima and Nagasaki]] had [[bone marrow]] that was completely destroyed. This helped them realize that diseased [[bone marrow]] could also be destroyed with radiation, which was a very important step in figuring out that [[leukemia]] (a blood cancer) can be treated with bone marrow transplants. Since World War II, scientists have kept finding better cancer treatments. However, there are some things that still need to get better. For example, while there are good treatments for many kinds of cancer, there are still no treatments for certain kinds of cancer, or for some cancers once they progress (or get worse) to a certain stage of the disease. Also, the cancer treatments that do exist are not all standardized (there is not one agreed-upon way of giving every treatment which is used each time the treatment is given). Cancer treatments are also not available everywhere in the world. People need to keep studying cancer epidemiology and forming international partnerships (where different countries work together) to find cures and make cancer treatments available everywhere. == Other websites == * [http://www.cancer.org/cancer/cancerbasics/what-is-cancer What is Cancer? (Simple English)] - American Cancer Society == References == {{reflist|2}} {{Tumours}} [[Category:Cancer| ]]
Cancer is a group of diseases involving abnormal cell growth with the potential to invade or spread to other parts of the body. These contrast with benign tumors, which do not spread. Possible signs and symptoms include a lump, abnormal bleeding, prolonged cough, unexplained weight loss, and a change in bowel movements. While these symptoms may indicate cancer, they can also have other causes. Over 100 types of cancers affect humans. Tobacco use is the cause of about 22% of cancer deaths. Another 10% are due to obesity, poor diet, lack of physical activity or excessive alcohol consumption. Other factors include certain infections, exposure to ionizing radiation, and environmental pollutants. In the developing world, 15% of cancers are due to infections such as Helicobacter pylori, hepatitis B, hepatitis C, human papillomavirus infection, Epstein–Barr virus and human immunodeficiency virus (HIV). These factors act, at least partly, by changing the genes of a cell. Typically, many genetic changes are required before cancer develops. Approximately 5–10% of cancers are due to inherited genetic defects. Cancer can be detected by certain signs and symptoms or screening tests. It is then typically further investigated by medical imaging and confirmed by biopsy. The risk of developing certain cancers can be reduced by not smoking, maintaining a healthy weight, limiting alcohol intake, eating plenty of vegetables, fruits, and whole grains, vaccination against certain infectious diseases, limiting consumption of processed meat and red meat, and limiting exposure to direct sunlight. Early detection through screening is useful for cervical and colorectal cancer. The benefits of screening for breast cancer are controversial. Cancer is often treated with some combination of radiation therapy, surgery, chemotherapy and targeted therapy. Pain and symptom management are an important part of care. Palliative care is particularly important in people with advanced disease. The chance of survival depends on the type of cancer and extent of disease at the start of treatment. In children under 15 at diagnosis, the five-year survival rate in the developed world is on average 80%. For cancer in the United States, the average five-year survival rate is 66% for all ages. In 2015, about 90.5 million people worldwide had cancer. In 2019, annual cancer cases grew by 23.6 million people, and there were 10 million deaths worldwide, representing over the previous decade increases of 26% and 21%, respectively. The most common types of cancer in males are lung cancer, prostate cancer, colorectal cancer, and stomach cancer. In females, the most common types are breast cancer, colorectal cancer, lung cancer, and cervical cancer. If skin cancer other than melanoma were included in total new cancer cases each year, it would account for around 40% of cases. In children, acute lymphoblastic leukemia and brain tumors are most common, except in Africa, where non-Hodgkin lymphoma occurs more often. In 2012, about 165,000 children under 15 years of age were diagnosed with cancer. The risk of cancer increases significantly with age, and many cancers occur more commonly in developed countries. Rates are increasing as more people live to an old age and as lifestyle changes occur in the developing world. The global total economic costs of cancer were estimated at US$1.16 trillion (equivalent to $1.56 trillion in 2022) per year as of 2010. The word comes from the ancient Greek καρκίνος, meaning 'crab' and 'tumor'. Greek physicians Hippocrates and Galen, among others, noted the similarity of crabs to some tumors with swollen veins. The word was introduced in English in the modern medical sense around 1600. Cancers comprise a large family of diseases that involve abnormal cell growth with the potential to invade or spread to other parts of the body. They form a subset of neoplasms. A neoplasm or tumor is a group of cells that have undergone unregulated growth and will often form a mass or lump, but may be distributed diffusely. All tumor cells show the six hallmarks of cancer. These characteristics are required to produce a malignant tumor. They include: The progression from normal cells to cells that can form a detectable mass to cancer involves multiple steps known as malignant progression. When cancer begins, it produces no symptoms. Signs and symptoms appear as the mass grows or ulcerates. The findings that result depend on cancer's type and location. Few symptoms are specific. Many frequently occur in individuals who have other conditions. Cancer can be difficult to diagnose and can be considered a "great imitator." People may become anxious or depressed post-diagnosis. The risk of suicide in people with cancer is approximately double. Local symptoms may occur due to the mass of the tumor or its ulceration. For example, mass effects from lung cancer can block the bronchus resulting in cough or pneumonia; esophageal cancer can cause narrowing of the esophagus, making it difficult or painful to swallow; and colorectal cancer may lead to narrowing or blockages in the bowel, affecting bowel habits. Masses in breasts or testicles may produce observable lumps. Ulceration can cause bleeding that can lead to symptoms such as coughing up blood (lung cancer), anemia or rectal bleeding (colon cancer), blood in the urine (bladder cancer), or abnormal vaginal bleeding (endometrial or cervical cancer). Although localized pain may occur in advanced cancer, the initial tumor is usually painless. Some cancers can cause a buildup of fluid within the chest or abdomen. Systemic symptoms may occur due to the body's response to the cancer. This may include fatigue, unintentional weight loss, or skin changes. Some cancers can cause a systemic inflammatory state that leads to ongoing muscle loss and weakness, known as cachexia. Some cancers, such as Hodgkin's disease, leukemias, and liver or kidney cancers, can cause a persistent fever. Shortness of breath, called dyspnea, is a common symptom of cancer and its treatment. The causes of cancer-related dyspnea can include tumors in or around the lung, blocked airways, fluid in the lungs, pneumonia, or treatment reactions including an allergic response. Treatment for dyspnea in patients with advanced cancer can include fans, bilevel ventilation, acupressure/reflexology and multicomponent nonpharmacological interventions. Some systemic symptoms of cancer are caused by hormones or other molecules produced by the tumor, known as paraneoplastic syndromes. Common paraneoplastic syndromes include hypercalcemia, which can cause altered mental state, constipation and dehydration, or hyponatremia, which can also cause altered mental status, vomiting, headaches, or seizures. Metastasis is the spread of cancer to other locations in the body. The dispersed tumors are called metastatic tumors, while the original is called the primary tumor. Almost all cancers can metastasize. Most cancer deaths are due to cancer that has metastasized. Metastasis is common in the late stages of cancer and it can occur via the blood or the lymphatic system or both. The typical steps in metastasis are local invasion, intravasation into the blood or lymph, circulation through the body, extravasation into the new tissue, proliferation and angiogenesis. Different types of cancers tend to metastasize to particular organs, but overall the most common places for metastases to occur are the lungs, liver, brain, and the bones. While some cancers can be cured if detected early, metastatic cancer is more difficult to treat and control. Nevertheless, some recent treatments are demonstrating encouraging results. The majority of cancers, some 90–95% of cases, are due to genetic mutations from environmental and lifestyle factors. The remaining 5–10% are due to inherited genetics. Environmental refers to any cause that is not inherited, such as lifestyle, economic, and behavioral factors and not merely pollution. Common environmental factors that contribute to cancer death include tobacco use (25–30%), diet and obesity (30–35%), infections (15–20%), radiation (both ionizing and non-ionizing, up to 10%), lack of physical activity, and pollution. Psychological stress does not appear to be a risk factor for the onset of cancer, though it may worsen outcomes in those who already have cancer. It is not generally possible to prove what caused a particular cancer because the various causes do not have specific fingerprints. For example, if a person who uses tobacco heavily develops lung cancer, then it was probably caused by the tobacco use, but since everyone has a small chance of developing lung cancer as a result of air pollution or radiation, the cancer may have developed for one of those reasons. Excepting the rare transmissions that occur with pregnancies and occasional organ donors, cancer is generally not a transmissible disease, however factors that may have contributed to the development of cancer can be transmissible; such as oncoviruses like hepatitis B, Epstein-Barr virus and HIV. Exposure to particular substances have been linked to specific types of cancer. These substances are called carcinogens. Tobacco smoke, for example, causes 90% of lung cancer. Tobacco use can cause cancer throughout the body including in the mouth and throat, larynx, esophagus, stomach, bladder, kidney, cervix, colon/rectum, liver and pancreas. Tobacco smoke contains over fifty known carcinogens, including nitrosamines and polycyclic aromatic hydrocarbons. Tobacco is responsible for about one in five cancer deaths worldwide and about one in three in the developed world. Lung cancer death rates in the United States have mirrored smoking patterns, with increases in smoking followed by dramatic increases in lung cancer death rates and, more recently, decreases in smoking rates since the 1950s followed by decreases in lung cancer death rates in men since 1990. In Western Europe, 10% of cancers in males and 3% of cancers in females are attributed to alcohol exposure, especially liver and digestive tract cancers. Cancer from work-related substance exposures may cause between 2 and 20% of cases, causing at least 200,000 deaths. Cancers such as lung cancer and mesothelioma can come from inhaling tobacco smoke or asbestos fibers, or leukemia from exposure to benzene. Exposure to perfluorooctanoic acid (PFOA), which is predominantly used in the production of Teflon, is known to cause two kinds of cancer. Diet, physical inactivity, and obesity are related to up to 30–35% of cancer deaths. In the United States, excess body weight is associated with the development of many types of cancer and is a factor in 14–20% of cancer deaths. A UK study including data on over 5 million people showed higher body mass index to be related to at least 10 types of cancer and responsible for around 12,000 cases each year in that country. Physical inactivity is believed to contribute to cancer risk, not only through its effect on body weight but also through negative effects on the immune system and endocrine system. More than half of the effect from the diet is due to overnutrition (eating too much), rather than from eating too few vegetables or other healthful foods. Some specific foods are linked to specific cancers. A high-salt diet is linked to gastric cancer. Aflatoxin B1, a frequent food contaminant, causes liver cancer. Betel nut chewing can cause oral cancer. National differences in dietary practices may partly explain differences in cancer incidence. For example, gastric cancer is more common in Japan due to its high-salt diet while colon cancer is more common in the United States. Immigrant cancer profiles mirror those of their new country, often within one generation. Worldwide, approximately 18% of cancer deaths are related to infectious diseases. This proportion ranges from a high of 25% in Africa to less than 10% in the developed world. Viruses are the usual infectious agents that cause cancer but cancer bacteria and parasites may also play a role. Oncoviruses (viruses that can cause cancer) include: Bacterial infection may also increase the risk of cancer, as seen in Helicobacter pylori-induced gastric carcinoma. Parasitic infections associated with cancer include: Radiation exposure such as ultraviolet radiation and radioactive material is a risk factor for cancer. Many non-melanoma skin cancers are due to ultraviolet radiation, mostly from sunlight. Sources of ionizing radiation include medical imaging and radon gas. Ionizing radiation is not a particularly strong mutagen. Residential exposure to radon gas, for example, has similar cancer risks as passive smoking. Radiation is a more potent source of cancer when combined with other cancer-causing agents, such as radon plus tobacco smoke. Radiation can cause cancer in most parts of the body, in all animals and at any age. Children are twice as likely to develop radiation-induced leukemia as adults; radiation exposure before birth has ten times the effect. Medical use of ionizing radiation is a small but growing source of radiation-induced cancers. Ionizing radiation may be used to treat other cancers, but this may, in some cases, induce a second form of cancer. It is also used in some kinds of medical imaging. Prolonged exposure to ultraviolet radiation from the sun can lead to melanoma and other skin malignancies. Clear evidence establishes ultraviolet radiation, especially the non-ionizing medium wave UVB, as the cause of most non-melanoma skin cancers, which are the most common forms of cancer in the world. Non-ionizing radio frequency radiation from mobile phones, electric power transmission and other similar sources has been described as a possible carcinogen by the World Health Organization's International Agency for Research on Cancer. Evidence, however, has not supported a concern. This includes that studies have not found a consistent link between mobile phone radiation and cancer risk. The vast majority of cancers are non-hereditary (sporadic). Hereditary cancers are primarily caused by an inherited genetic defect. Less than 0.3% of the population are carriers of a genetic mutation that has a large effect on cancer risk and these cause less than 3–10% of cancer. Some of these syndromes include: certain inherited mutations in the genes BRCA1 and BRCA2 with a more than 75% risk of breast cancer and ovarian cancer, and hereditary nonpolyposis colorectal cancer (HNPCC or Lynch syndrome), which is present in about 3% of people with colorectal cancer, among others. Statistically for cancers causing most mortality, the relative risk of developing colorectal cancer when a first-degree relative (parent, sibling or child) has been diagnosed with it is about 2. The corresponding relative risk is 1.5 for lung cancer, and 1.9 for prostate cancer. For breast cancer, the relative risk is 1.8 with a first-degree relative having developed it at 50 years of age or older, and 3.3 when the relative developed it when being younger than 50 years of age. Taller people have an increased risk of cancer because they have more cells than shorter people. Since height is genetically determined to a large extent, taller people have a heritable increase of cancer risk. Some substances cause cancer primarily through their physical, rather than chemical, effects. A prominent example of this is prolonged exposure to asbestos, naturally occurring mineral fibers that are a major cause of mesothelioma (cancer of the serous membrane) usually the serous membrane surrounding the lungs. Other substances in this category, including both naturally occurring and synthetic asbestos-like fibers, such as wollastonite, attapulgite, glass wool and rock wool, are believed to have similar effects. Non-fibrous particulate materials that cause cancer include powdered metallic cobalt and nickel and crystalline silica (quartz, cristobalite and tridymite). Usually, physical carcinogens must get inside the body (such as through inhalation) and require years of exposure to produce cancer. Physical trauma resulting in cancer is relatively rare. Claims that breaking bones resulted in bone cancer, for example, have not been proven. Similarly, physical trauma is not accepted as a cause for cervical cancer, breast cancer or brain cancer. One accepted source is frequent, long-term application of hot objects to the body. It is possible that repeated burns on the same part of the body, such as those produced by kanger and kairo heaters (charcoal hand warmers), may produce skin cancer, especially if carcinogenic chemicals are also present. Frequent consumption of scalding hot tea may produce esophageal cancer. Generally, it is believed that cancer arises, or a pre-existing cancer is encouraged, during the process of healing, rather than directly by the trauma. However, repeated injuries to the same tissues might promote excessive cell proliferation, which could then increase the odds of a cancerous mutation. Chronic inflammation has been hypothesized to directly cause mutation. Inflammation can contribute to proliferation, survival, angiogenesis and migration of cancer cells by influencing the tumor microenvironment. Oncogenes build up an inflammatory pro-tumorigenic microenvironment. Some hormones play a role in the development of cancer by promoting cell proliferation. Insulin-like growth factors and their binding proteins play a key role in cancer cell proliferation, differentiation and apoptosis, suggesting possible involvement in carcinogenesis. Hormones are important agents in sex-related cancers, such as cancer of the breast, endometrium, prostate, ovary and testis and also of thyroid cancer and bone cancer. For example, the daughters of women who have breast cancer have significantly higher levels of estrogen and progesterone than the daughters of women without breast cancer. These higher hormone levels may explain their higher risk of breast cancer, even in the absence of a breast-cancer gene. Similarly, men of African ancestry have significantly higher levels of testosterone than men of European ancestry and have a correspondingly higher level of prostate cancer. Men of Asian ancestry, with the lowest levels of testosterone-activating androstanediol glucuronide, have the lowest levels of prostate cancer. Other factors are relevant: obese people have higher levels of some hormones associated with cancer and a higher rate of those cancers. Women who take hormone replacement therapy have a higher risk of developing cancers associated with those hormones. On the other hand, people who exercise far more than average have lower levels of these hormones and lower risk of cancer. Osteosarcoma may be promoted by growth hormones. Some treatments and prevention approaches leverage this cause by artificially reducing hormone levels and thus discouraging hormone-sensitive cancers. There is an association between celiac disease and an increased risk of all cancers. People with untreated celiac disease have a higher risk, but this risk decreases with time after diagnosis and strict treatment, probably due to the adoption of a gluten-free diet, which seems to have a protective role against development of malignancy in people with celiac disease. However, the delay in diagnosis and initiation of a gluten-free diet seems to increase the risk of malignancies. Rates of gastrointestinal cancers are increased in people with Crohn's disease and ulcerative colitis, due to chronic inflammation. Also, immunomodulators and biologic agents used to treat these diseases may promote developing extra-intestinal malignancies. Cancer is fundamentally a disease of tissue growth regulation. For a normal cell to transform into a cancer cell, the genes that regulate cell growth and differentiation must be altered. The affected genes are divided into two broad categories. Oncogenes are genes that promote cell growth and reproduction. Tumor suppressor genes are genes that inhibit cell division and survival. Malignant transformation can occur through the formation of novel oncogenes, the inappropriate over-expression of normal oncogenes, or by the under-expression or disabling of tumor suppressor genes. Typically, changes in multiple genes are required to transform a normal cell into a cancer cell. Genetic changes can occur at different levels and by different mechanisms. The gain or loss of an entire chromosome can occur through errors in mitosis. More common are mutations, which are changes in the nucleotide sequence of genomic DNA. Large-scale mutations involve the deletion or gain of a portion of a chromosome. Genomic amplification occurs when a cell gains copies (often 20 or more) of a small chromosomal locus, usually containing one or more oncogenes and adjacent genetic material. Translocation occurs when two separate chromosomal regions become abnormally fused, often at a characteristic location. A well-known example of this is the Philadelphia chromosome, or translocation of chromosomes 9 and 22, which occurs in chronic myelogenous leukemia and results in production of the BCR-abl fusion protein, an oncogenic tyrosine kinase. Small-scale mutations include point mutations, deletions, and insertions, which may occur in the promoter region of a gene and affect its expression, or may occur in the gene's coding sequence and alter the function or stability of its protein product. Disruption of a single gene may also result from integration of genomic material from a DNA virus or retrovirus, leading to the expression of viral oncogenes in the affected cell and its descendants. Replication of the data contained within the DNA of living cells will probabilistically result in some errors (mutations). Complex error correction and prevention are built into the process and safeguard the cell against cancer. If a significant error occurs, the damaged cell can self-destruct through programmed cell death, termed apoptosis. If the error control processes fail, then the mutations will survive and be passed along to daughter cells. Some environments make errors more likely to arise and propagate. Such environments can include the presence of disruptive substances called carcinogens, repeated physical injury, heat, ionising radiation, or hypoxia. The errors that cause cancer are self-amplifying and compounding, for example: The transformation of a normal cell into cancer is akin to a chain reaction caused by initial errors, which compound into more severe errors, each progressively allowing the cell to escape more controls that limit normal tissue growth. This rebellion-like scenario is an undesirable survival of the fittest, where the driving forces of evolution work against the body's design and enforcement of order. Once cancer has begun to develop, this ongoing process, termed clonal evolution, drives progression towards more invasive stages. Clonal evolution leads to intra-tumour heterogeneity (cancer cells with heterogeneous mutations) that complicates designing effective treatment strategies and requires an evolutionary approach to designing treatment. Characteristic abilities developed by cancers are divided into categories, specifically evasion of apoptosis, self-sufficiency in growth signals, insensitivity to anti-growth signals, sustained angiogenesis, limitless replicative potential, metastasis, reprogramming of energy metabolism and evasion of immune destruction. The classical view of cancer is a set of diseases driven by progressive genetic abnormalities that include mutations in tumor-suppressor genes and oncogenes, and in chromosomal abnormalities. A role for epigenetic alterations was identified in the early 21st century. Epigenetic alterations are functionally relevant modifications to the genome that do not change the nucleotide sequence. Examples of such modifications are changes in DNA methylation (hypermethylation and hypomethylation), histone modification and changes in chromosomal architecture (caused by inappropriate expression of proteins such as HMGA2 or HMGA1). Each of these alterations regulates gene expression without altering the underlying DNA sequence. These changes may remain through cell divisions, endure for multiple generations, and can be considered as equivalent to mutations. Epigenetic alterations occur frequently in cancers. As an example, one study listed protein coding genes that were frequently altered in their methylation in association with colon cancer. These included 147 hypermethylated and 27 hypomethylated genes. Of the hypermethylated genes, 10 were hypermethylated in 100% of colon cancers and many others were hypermethylated in more than 50% of colon cancers. While epigenetic alterations are found in cancers, the epigenetic alterations in DNA repair genes, causing reduced expression of DNA repair proteins, may be of particular importance. Such alterations may occur early in progression to cancer and are a possible cause of the genetic instability characteristic of cancers. Reduced expression of DNA repair genes disrupts DNA repair. This is shown in the figure at the 4th level from the top. (In the figure, red wording indicates the central role of DNA damage and defects in DNA repair in progression to cancer.) When DNA repair is deficient DNA damage remains in cells at a higher than usual level (5th level) and causes increased frequencies of mutation and/or epimutation (6th level). Mutation rates increase substantially in cells defective in DNA mismatch repair or in homologous recombinational repair (HRR). Chromosomal rearrangements and aneuploidy also increase in HRR defective cells. Higher levels of DNA damage cause increased mutation (right side of figure) and increased epimutation. During repair of DNA double strand breaks, or repair of other DNA damage, incompletely cleared repair sites can cause epigenetic gene silencing. Deficient expression of DNA repair proteins due to an inherited mutation can increase cancer risks. Individuals with an inherited impairment in any of 34 DNA repair genes (see article DNA repair-deficiency disorder) have increased cancer risk, with some defects ensuring a 100% lifetime chance of cancer (e.g. p53 mutations). Germ line DNA repair mutations are noted on the figure's left side. However, such germline mutations (which cause highly penetrant cancer syndromes) are the cause of only about 1 percent of cancers. In sporadic cancers, deficiencies in DNA repair are occasionally caused by a mutation in a DNA repair gene but are much more frequently caused by epigenetic alterations that reduce or silence expression of DNA repair genes. This is indicated in the figure at the 3rd level. Many studies of heavy metal-induced carcinogenesis show that such heavy metals cause a reduction in expression of DNA repair enzymes, some through epigenetic mechanisms. DNA repair inhibition is proposed to be a predominant mechanism in heavy metal-induced carcinogenicity. In addition, frequent epigenetic alterations of the DNA sequences code for small RNAs called microRNAs (or miRNAs). miRNAs do not code for proteins, but can "target" protein-coding genes and reduce their expression. Cancers usually arise from an assemblage of mutations and epimutations that confer a selective advantage leading to clonal expansion (see Field defects in progression to cancer). Mutations, however, may not be as frequent in cancers as epigenetic alterations. An average cancer of the breast or colon can have about 60 to 70 protein-altering mutations, of which about three or four may be "driver" mutations and the remaining ones may be "passenger" mutations. Metastasis is the spread of cancer to other locations in the body. The dispersed tumors are called metastatic tumors, while the original is called the primary tumor. Almost all cancers can metastasize. Most cancer deaths are due to cancer that has metastasized. Metastasis is common in the late stages of cancer and it can occur via the blood or the lymphatic system or both. The typical steps in metastasis are local invasion, intravasation into the blood or lymph, circulation through the body, extravasation into the new tissue, proliferation and angiogenesis. Different types of cancers tend to metastasize to particular organs, but overall the most common places for metastases to occur are the lungs, liver, brain and the bones. Normal cells typically generate only about 30% of energy from glycolysis, whereas most cancers rely on glycolysis for energy production (Warburg effect). But a minority of cancer types rely on oxidative phosphorylation as the primary energy source, including lymphoma, leukemia, and endometrial cancer. Even in these cases, however, the use of glycolysis as an energy source rarely exceeds 60%. A few cancers use glutamine as the major energy source, partly because it provides nitrogen required for nucleotide (DNA, RNA) synthesis. Cancer stem cells often use oxidative phosphorylation or glutamine as a primary energy source. Most cancers are initially recognized either because of the appearance of signs or symptoms or through screening. Neither of these leads to a definitive diagnosis, which requires the examination of a tissue sample by a pathologist. People with suspected cancer are investigated with medical tests. These commonly include blood tests, X-rays, (contrast) CT scans and endoscopy. The tissue diagnosis from the biopsy indicates the type of cell that is proliferating, its histological grade, genetic abnormalities and other features. Together, this information is useful to evaluate the prognosis and to choose the best treatment. Cytogenetics and immunohistochemistry are other types of tissue tests. These tests provide information about molecular changes (such as mutations, fusion genes and numerical chromosome changes) and may thus also indicate the prognosis and best treatment. Cancer diagnosis can cause psychological distress and psychosocial interventions, such as talking therapy, may help people with this. Cancers are classified by the type of cell that the tumor cells resemble and is therefore presumed to be the origin of the tumor. These types include: Cancers are usually named using -carcinoma, -sarcoma or -blastoma as a suffix, with the Latin or Greek word for the organ or tissue of origin as the root. For example, cancers of the liver parenchyma arising from malignant epithelial cells is called hepatocarcinoma, while a malignancy arising from primitive liver precursor cells is called a hepatoblastoma and a cancer arising from fat cells is called a liposarcoma. For some common cancers, the English organ name is used. For example, the most common type of breast cancer is called ductal carcinoma of the breast. Here, the adjective ductal refers to the appearance of cancer under the microscope, which suggests that it has originated in the milk ducts. Benign tumors (which are not cancers) are named using -oma as a suffix with the organ name as the root. For example, a benign tumor of smooth muscle cells is called a leiomyoma (the common name of this frequently occurring benign tumor in the uterus is fibroid). Confusingly, some types of cancer use the -noma suffix, examples including melanoma and seminoma. Some types of cancer are named for the size and shape of the cells under a microscope, such as giant cell carcinoma, spindle cell carcinoma and small-cell carcinoma. Cancer prevention is defined as active measures to decrease cancer risk. The vast majority of cancer cases are due to environmental risk factors. Many of these environmental factors are controllable lifestyle choices. Thus, cancer is generally preventable. Between 70% and 90% of common cancers are due to environmental factors and therefore potentially preventable. Greater than 30% of cancer deaths could be prevented by avoiding risk factors including: tobacco, excess weight/obesity, poor diet, physical inactivity, alcohol, sexually transmitted infections and air pollution. Further, poverty could be considered as an indirect risk factor in human cancers. Not all environmental causes are controllable, such as naturally occurring background radiation and cancers caused through hereditary genetic disorders and thus are not preventable via personal behavior. In 2019, ~44% of all cancer deaths – or ~4.5 M deaths or ~105 million lost disability-adjusted life years – were due to known clearly preventable risk factors, led by smoking, alcohol use and high BMI, according to a GBD systematic analysis. While many dietary recommendations have been proposed to reduce cancer risks, the evidence to support them is not definitive. The primary dietary factors that increase risk are obesity and alcohol consumption. Diets low in fruits and vegetables and high in red meat have been implicated but reviews and meta-analyses do not come to a consistent conclusion. A 2014 meta-analysis found no relationship between fruits and vegetables and cancer. Coffee is associated with a reduced risk of liver cancer. Studies have linked excessive consumption of red or processed meat to an increased risk of breast cancer, colon cancer and pancreatic cancer, a phenomenon that could be due to the presence of carcinogens in meats cooked at high temperatures. In 2015 the IARC reported that eating processed meat (e.g., bacon, ham, hot dogs, sausages) and, to a lesser degree, red meat was linked to some cancers. Dietary recommendations for cancer prevention typically include an emphasis on vegetables, fruit, whole grains and fish and an avoidance of processed and red meat (beef, pork, lamb), animal fats, pickled foods and refined carbohydrates. Medications can be used to prevent cancer in a few circumstances. In the general population, NSAIDs reduce the risk of colorectal cancer; however, due to cardiovascular and gastrointestinal side effects, they cause overall harm when used for prevention. Aspirin has been found to reduce the risk of death from cancer by about 7%. COX-2 inhibitors may decrease the rate of polyp formation in people with familial adenomatous polyposis; however, it is associated with the same adverse effects as NSAIDs. Daily use of tamoxifen or raloxifene reduce the risk of breast cancer in high-risk women. The benefit versus harm for 5-alpha-reductase inhibitor such as finasteride is not clear. Vitamin supplementation does not appear to be effective at preventing cancer. While low blood levels of vitamin D are correlated with increased cancer risk, whether this relationship is causal and vitamin D supplementation is protective is not determined. One 2014 review found that supplements had no significant effect on cancer risk. Another 2014 review concluded that vitamin D3 may decrease the risk of death from cancer (one fewer death in 150 people treated over 5 years), but concerns with the quality of the data were noted. Beta-Carotene supplementation increases lung cancer rates in those who are high risk. Folic acid supplementation is not effective in preventing colon cancer and may increase colon polyps. Selenium supplementation has not been shown to reduce the risk of cancer. Vaccines have been developed that prevent infection by some carcinogenic viruses. Human papillomavirus vaccine (Gardasil and Cervarix) decrease the risk of developing cervical cancer. The hepatitis B vaccine prevents infection with hepatitis B virus and thus decreases the risk of liver cancer. The administration of human papillomavirus and hepatitis B vaccinations is recommended where resources allow. Unlike diagnostic efforts prompted by symptoms and medical signs, cancer screening involves efforts to detect cancer after it has formed, but before any noticeable symptoms appear. This may involve physical examination, blood or urine tests or medical imaging. Cancer screening is not available for many types of cancers. Even when tests are available, they may not be recommended for everyone. Universal screening or mass screening involves screening everyone. Selective screening identifies people who are at higher risk, such as people with a family history. Several factors are considered to determine whether the benefits of screening outweigh the risks and the costs of screening. These factors include: The U.S. Preventive Services Task Force (USPSTF) issues recommendations for various cancers: Screens for gastric cancer using photofluorography due to the high incidence there. Genetic testing for individuals at high-risk of certain cancers is recommended by unofficial groups. Carriers of these mutations may then undergo enhanced surveillance, chemoprevention, or preventative surgery to reduce their subsequent risk. Many treatment options for cancer exist. The primary ones include surgery, chemotherapy, radiation therapy, hormonal therapy, targeted therapy and palliative care. Which treatments are used depends on the type, location and grade of the cancer as well as the patient's health and preferences. The treatment intent may or may not be curative. Chemotherapy is the treatment of cancer with one or more cytotoxic anti-neoplastic drugs (chemotherapeutic agents) as part of a standardized regimen. The term encompasses a variety of drugs, which are divided into broad categories such as alkylating agents and antimetabolites. Traditional chemotherapeutic agents act by killing cells that divide rapidly, a critical property of most cancer cells. It was found that providing combined cytotoxic drugs is better than a single drug, a process called the combination therapy, which has an advantage in the statistics of survival and response to the tumor and in the progress of the disease. A Cochrane review concluded that combined therapy was more effective to treat metastasized breast cancer. However, generally it is not certain whether combination chemotherapy leads to better health outcomes, when both survival and toxicity are considered. Targeted therapy is a form of chemotherapy that targets specific molecular differences between cancer and normal cells. The first targeted therapies blocked the estrogen receptor molecule, inhibiting the growth of breast cancer. Another common example is the class of Bcr-Abl inhibitors, which are used to treat chronic myelogenous leukemia (CML). Currently, targeted therapies exist for many of the most common cancer types, including bladder cancer, breast cancer, colorectal cancer, kidney cancer, leukemia, liver cancer, lung cancer, lymphoma, pancreatic cancer, prostate cancer, skin cancer, and thyroid cancer as well as other cancer types. The efficacy of chemotherapy depends on the type of cancer and the stage. In combination with surgery, chemotherapy has proven useful in cancer types including breast cancer, colorectal cancer, pancreatic cancer, osteogenic sarcoma, testicular cancer, ovarian cancer and certain lung cancers. Chemotherapy is curative for some cancers, such as some leukemias, ineffective in some brain tumors, and needless in others, such as most non-melanoma skin cancers. The effectiveness of chemotherapy is often limited by its toxicity to other tissues in the body. Even when chemotherapy does not provide a permanent cure, it may be useful to reduce symptoms such as pain or to reduce the size of an inoperable tumor in the hope that surgery will become possible in the future. Radiation therapy involves the use of ionizing radiation in an attempt to either cure or improve symptoms. It works by damaging the DNA of cancerous tissue, causing mitotic catastrophe resulting in the death of the cancer cells. To spare normal tissues (such as skin or organs, which radiation must pass through to treat the tumor), shaped radiation beams are aimed from multiple exposure angles to intersect at the tumor, providing a much larger dose there than in the surrounding, healthy tissue. As with chemotherapy, cancers vary in their response to radiation therapy. Radiation therapy is used in about half of cases. The radiation can be either from internal sources (brachytherapy) or external sources. The radiation is most commonly low energy X-rays for treating skin cancers, while higher energy X-rays are used for cancers within the body. Radiation is typically used in addition to surgery and or chemotherapy. For certain types of cancer, such as early head and neck cancer, it may be used alone. Radiation therapy after surgery for brain metastases has been shown to not improve overall survival in patients compared to surgery alone. For painful bone metastasis, radiation therapy has been found to be effective in about 70% of patients. Surgery is the primary method of treatment for most isolated, solid cancers and may play a role in palliation and prolongation of survival. It is typically an important part of definitive diagnosis and staging of tumors, as biopsies are usually required. In localized cancer, surgery typically attempts to remove the entire mass along with, in certain cases, the lymph nodes in the area. For some types of cancer this is sufficient to eliminate the cancer. Palliative care is treatment that attempts to help the patient feel better and may be combined with an attempt to treat the cancer. Palliative care includes action to reduce physical, emotional, spiritual and psycho-social distress. Unlike treatment that is aimed at directly killing cancer cells, the primary goal of palliative care is to improve quality of life. People at all stages of cancer treatment typically receive some kind of palliative care. In some cases, medical specialty professional organizations recommend that patients and physicians respond to cancer only with palliative care. This applies to patients who: Palliative care may be confused with hospice and therefore only indicated when people approach end of life. Like hospice care, palliative care attempts to help the patient cope with their immediate needs and to increase comfort. Unlike hospice care, palliative care does not require people to stop treatment aimed at the cancer. Multiple national medical guidelines recommend early palliative care for patients whose cancer has produced distressing symptoms or who need help coping with their illness. In patients first diagnosed with metastatic disease, palliative care may be immediately indicated. Palliative care is indicated for patients with a prognosis of less than 12 months of life even given aggressive treatment. A variety of therapies using immunotherapy, stimulating or helping the immune system to fight cancer, have come into use since 1997. Approaches include antibodies, checkpoint therapy, and adoptive cell transfer. Laser therapy uses high-intensity light to treat cancer by shrinking or destroying tumors or precancerous growths. Lasers are most commonly used to treat superficial cancers that are on the surface of the body or the lining of internal organs. It is used to treat basal cell skin cancer and the very early stages of others like cervical, penile, vaginal, vulvar, and non-small cell lung cancer. It is often combined with other treatments, such as surgery, chemotherapy, or radiation therapy. Laser-induced interstitial thermotherapy (LITT), or interstitial laser photocoagulation, uses lasers to treat some cancers using hyperthermia, which uses heat to shrink tumors by damaging or killing cancer cells. Laser are more precise than surgery and cause less damage, pain, bleeding, swelling, and scarring. A disadvantage is surgeons must have specialized training. It may be more expensive than other treatments. Complementary and alternative cancer treatments are a diverse group of therapies, practices and products that are not part of conventional medicine. "Complementary medicine" refers to methods and substances used along with conventional medicine, while "alternative medicine" refers to compounds used instead of conventional medicine. Most complementary and alternative medicines for cancer have not been studied or tested using conventional techniques such as clinical trials. Some alternative treatments have been investigated and shown to be ineffective but still continue to be marketed and promoted. Cancer researcher Andrew J. Vickers stated, "The label 'unproven' is inappropriate for such therapies; it is time to assert that many alternative cancer therapies have been 'disproven'." Survival rates vary by cancer type and by the stage at which it is diagnosed, ranging from majority survival to complete mortality five years after diagnosis. Once a cancer has metastasized, prognosis normally becomes much worse. About half of patients receiving treatment for invasive cancer (excluding carcinoma in situ and non-melanoma skin cancers) die from that cancer or its treatment. A majority of cancer deaths are due to metastases of the primary tumor. Survival is worse in the developing world, partly because the types of cancer that are most common there are harder to treat than those associated with developed countries. Those who survive cancer develop a second primary cancer at about twice the rate of those never diagnosed. The increased risk is believed to be due to the random chance of developing any cancer, the likelihood of surviving the first cancer, the same risk factors that produced the first cancer, unwanted side effects of treating the first cancer (particularly radiation therapy), and better compliance with screening. Predicting short- or long-term survival depends on many factors. The most important are the cancer type and the patient's age and overall health. Those who are frail with other health problems have lower survival rates than otherwise healthy people. Centenarians are unlikely to survive for five years even if treatment is successful. People who report a higher quality of life tend to survive longer. People with lower quality of life may be affected by depression and other complications and/or disease progression that both impairs quality and quantity of life. Additionally, patients with worse prognoses may be depressed or report poorer quality of life because they perceive that their condition is likely to be fatal. People with cancer have an increased risk of blood clots in their veins which can be life-threatening. The use of blood thinners such as heparin decrease the risk of blood clots but have not been shown to increase survival in people with cancer. People who take blood thinners also have an increased risk of bleeding. Although extremely rare, some forms of cancer, even from an advanced stage, can heal spontaneously. This phenomenon is known as the spontaneous remission. Estimates are that in 2018, 18.1 million new cases of cancer and 9.6 million deaths occur globally. About 20% of males and 17% of females will get cancer at some point in time while 13% of males and 9% of females will die from it. In 2008, approximately 12.7 million cancers were diagnosed (excluding non-melanoma skin cancers and other non-invasive cancers) and in 2010 nearly 7.98 million people died. Cancers account for approximately 16% of deaths. The most common as of 2018 are lung cancer (1.76 million deaths), colorectal cancer (860,000) stomach cancer (780,000), liver cancer (780,000), and breast cancer (620,000). This makes invasive cancer the leading cause of death in the developed world and the second leading in the developing world. Over half of cases occur in the developing world. Deaths from cancer were 5.8 million in 1990. Deaths have been increasing primarily due to longer lifespans and lifestyle changes in the developing world. The most significant risk factor for developing cancer is age. Although it is possible for cancer to strike at any age, most patients with invasive cancer are over 65. According to cancer researcher Robert A. Weinberg, "If we lived long enough, sooner or later we all would get cancer." Some of the association between aging and cancer is attributed to immunosenescence, errors accumulated in DNA over a lifetime and age-related changes in the endocrine system. Aging's effect on cancer is complicated by factors such as DNA damage and inflammation promoting it and factors such as vascular aging and endocrine changes inhibiting it. Some slow-growing cancers are particularly common, but often are not fatal. Autopsy studies in Europe and Asia showed that up to 36% of people have undiagnosed and apparently harmless thyroid cancer at the time of their deaths and that 80% of men develop prostate cancer by age 80. As these cancers do not cause the patient's death, identifying them would have represented overdiagnosis rather than useful medical care. The three most common childhood cancers are leukemia (34%), brain tumors (23%) and lymphomas (12%). In the United States cancer affects about 1 in 285 children. Rates of childhood cancer increased by 0.6% per year between 1975 and 2002 in the United States and by 1.1% per year between 1978 and 1997 in Europe. Death from childhood cancer decreased by half between 1975 and 2010 in the United States. Cancer has existed for all of human history. The earliest written record regarding cancer is from circa 1600 BC in the Egyptian Edwin Smith Papyrus and describes breast cancer. Hippocrates (c. 460 BC – c. 370 BC) described several kinds of cancer, referring to them with the Greek word καρκίνος karkinos (crab or crayfish). This name comes from the appearance of the cut surface of a solid malignant tumor, with "the veins stretched on all sides as the animal the crab has its feet, whence it derives its name". Galen stated that "cancer of the breast is so called because of the fancied resemblance to a crab given by the lateral prolongations of the tumor and the adjacent distended veins". Celsus (c. 25 BC – 50 AD) translated karkinos into the Latin cancer, also meaning crab and recommended surgery as treatment. Galen (2nd century AD) disagreed with the use of surgery and recommended purgatives instead. These recommendations largely stood for 1000 years. In the 15th, 16th and 17th centuries, it became acceptable for doctors to dissect bodies to discover the cause of death. The German professor Wilhelm Fabry believed that breast cancer was caused by a milk clot in a mammary duct. The Dutch professor Francois de la Boe Sylvius, a follower of Descartes, believed that all disease was the outcome of chemical processes and that acidic lymph fluid was the cause of cancer. His contemporary Nicolaes Tulp believed that cancer was a poison that slowly spreads and concluded that it was contagious. The physician John Hill described tobacco sniffing as the cause of nose cancer in 1761. This was followed by the report in 1775 by British surgeon Percivall Pott that chimney sweeps' carcinoma, a cancer of the scrotum, was a common disease among chimney sweeps. With the widespread use of the microscope in the 18th century, it was discovered that the 'cancer poison' spread from the primary tumor through the lymph nodes to other sites ("metastasis"). This view of the disease was first formulated by the English surgeon Campbell De Morgan between 1871 and 1874. Although many diseases (such as heart failure) may have a worse prognosis than most cases of cancer, cancer is the subject of widespread fear and taboos. The euphemism of "a long illness" to describe cancers leading to death is still commonly used in obituaries, rather than naming the disease explicitly, reflecting an apparent stigma. Cancer is also euphemised as "the C-word"; Macmillan Cancer Support uses the term to try to lessen the fear around the disease. In Nigeria, one local name for cancer translates into English as "the disease that cannot be cured". This deep belief that cancer is necessarily a difficult and usually deadly disease is reflected in the systems chosen by society to compile cancer statistics: the most common form of cancer—non-melanoma skin cancers, accounting for about one-third of cancer cases worldwide, but very few deaths—are excluded from cancer statistics specifically because they are easily treated and almost always cured, often in a single, short, outpatient procedure. Western conceptions of patients' rights for people with cancer include a duty to fully disclose the medical situation to the person, and the right to engage in shared decision-making in a way that respects the person's own values. In other cultures, other rights and values are preferred. For example, most African cultures value whole families rather than individualism. In parts of Africa, a diagnosis is commonly made so late that cure is not possible, and treatment, if available at all, would quickly bankrupt the family. As a result of these factors, African healthcare providers tend to let family members decide whether, when and how to disclose the diagnosis, and they tend to do so slowly and circuitously, as the person shows interest and an ability to cope with the grim news. People from Asian and South American countries also tend to prefer a slower, less candid approach to disclosure than is idealized in the United States and Western Europe, and they believe that sometimes it would be preferable not to be told about a cancer diagnosis. In general, disclosure of the diagnosis is more common than it was in the 20th century, but full disclosure of the prognosis is not offered to many patients around the world. In the United States and some other cultures, cancer is regarded as a disease that must be "fought" to end the "civil insurrection"; a War on Cancer was declared in the US. Military metaphors are particularly common in descriptions of cancer's human effects, and they emphasize both the state of the patient's health and the need to take immediate, decisive actions himself rather than to delay, to ignore or to rely entirely on others. The military metaphors also help rationalize radical, destructive treatments. In the 1970s, a relatively popular alternative cancer treatment in the US was a specialized form of talk therapy, based on the idea that cancer was caused by a bad attitude. People with a "cancer personality"—depressed, repressed, self-loathing and afraid to express their emotions—were believed to have manifested cancer through subconscious desire. Some psychotherapists said that treatment to change the patient's outlook on life would cure the cancer. Among other effects, this belief allowed society to blame the victim for having caused the cancer (by "wanting" it) or having prevented its cure (by not becoming a sufficiently happy, fearless and loving person). It also increased patients' anxiety, as they incorrectly believed that natural emotions of sadness, anger or fear shorten their lives. The idea was ridiculed by Susan Sontag, who published Illness as Metaphor while recovering from treatment for breast cancer in 1978. Although the original idea is now generally regarded as nonsense, the idea partly persists in a reduced form with a widespread, but incorrect, belief that deliberately cultivating a habit of positive thinking will increase survival. This notion is particularly strong in breast cancer culture. One idea about why people with cancer are blamed or stigmatized, called the just-world hypothesis, is that blaming cancer on the patient's actions or attitudes allows the blamers to regain a sense of control. This is based upon the blamers' belief that the world is fundamentally just and so any dangerous illness, like cancer, must be a type of punishment for bad choices, because in a just world, bad things would not happen to good people. The total health care expenditure on cancer in the US was estimated to be $80.2 billion in 2015. Even though cancer-related health care expenditure have increased in absolute terms during recent decades, the share of health expenditure devoted to cancer treatment has remained close to 5% between the 1960s and 2004. A similar pattern has been observed in Europe where about 6% of all health care expenditure are spent on cancer treatment. In addition to health care expenditure and financial toxicity, cancer causes indirect costs in the form of productivity losses due to sick days, permanent incapacity and disability as well as premature death during working age. Cancer causes also costs for informal care. Indirect costs and informal care costs are typically estimated to exceed or equal the health care costs of cancer. In the United States, cancer is included as a protected condition by the Equal Employment Opportunity Commission (EEOC), mainly due to the potential for cancer having discriminating effects on workers. Discrimination in the workplace could occur if an employer holds a false belief that a person with cancer is not capable of doing a job properly, and may ask for more sick leave than other employees. Employers may also make hiring or firing decisions based on misconceptions about cancer disabilities, if present. The EEOC provides interview guidelines for employers, as well as lists of possible solutions for assessing and accommodating employees with cancer. Women are six times more likely to be separated or divorced soon after a diagnosis of cancer or multiple sclerosis than men. Doctors in neuro-oncology practices noticed that divorce occurred almost exclusively when the wife was the patient. Because cancer is a class of diseases, it is unlikely that there will ever be a single "cure for cancer" any more than there will be a single treatment for all infectious diseases. Angiogenesis inhibitors were once incorrectly thought to have potential as a "silver bullet" treatment applicable to many types of cancer. Angiogenesis inhibitors and other cancer therapeutics are used in combination to reduce cancer morbidity and mortality. Experimental cancer treatments are studied in clinical trials to compare the proposed treatment to the best existing treatment. Treatments that succeeded in one cancer type can be tested against other types. Diagnostic tests are under development to better target the right therapies to the right patients, based on their individual biology. Cancer research focuses on the following issues: The improved understanding of molecular biology and cellular biology due to cancer research has led to new treatments for cancer since US President Richard Nixon declared the "War on Cancer" in 1971. Since then, the country has spent over $200 billion on cancer research, including resources from public and private sectors. The cancer death rate (adjusting for size and age of the population) declined by five percent between 1950 and 2005. Competition for financial resources appears to have suppressed the creativity, cooperation, risk-taking and original thinking required to make fundamental discoveries, unduly favoring low-risk research into small incremental advancements over riskier, more innovative research. Other consequences of competition appear to be many studies with dramatic claims whose results cannot be replicated and perverse incentives that encourage grantee institutions to grow without making sufficient investments in their own faculty and facilities. Virotherapy, which uses convert viruses, is being studied. In the wake of the COVID-19 pandemic, there has been a worry that cancer research and treatment are slowing down. Cancer affects approximately 1 in 1,000 pregnant women. The most common cancers found during pregnancy are the same as the most common cancers found in non-pregnant women during childbearing ages: breast cancer, cervical cancer, leukemia, lymphoma, melanoma, ovarian cancer and colorectal cancer. Diagnosing a new cancer in a pregnant woman is difficult, in part because any symptoms are commonly assumed to be a normal discomfort associated with pregnancy. As a result, cancer is typically discovered at a somewhat later stage than average. Some imaging procedures, such as MRIs (magnetic resonance imaging), CT scans, ultrasounds and mammograms with fetal shielding are considered safe during pregnancy; some others, such as PET scans, are not. Treatment is generally the same as for non-pregnant women. However, radiation and radioactive drugs are normally avoided during pregnancy, especially if the fetal dose might exceed 100 cGy. In some cases, some or all treatments are postponed until after birth if the cancer is diagnosed late in the pregnancy. Early deliveries are often used to advance the start of treatment. Surgery is generally safe, but pelvic surgeries during the first trimester may cause miscarriage. Some treatments, especially certain chemotherapy drugs given during the first trimester, increase the risk of birth defects and pregnancy loss (spontaneous abortions and stillbirths). Elective abortions are not required and, for the most common forms and stages of cancer, do not improve the mother's survival. In a few instances, such as advanced uterine cancer, the pregnancy cannot be continued and in others, the patient may end the pregnancy so that she can begin aggressive chemotherapy. Some treatments can interfere with the mother's ability to give birth vaginally or to breastfeed. Cervical cancer may require birth by Caesarean section. Radiation to the breast reduces the ability of that breast to produce milk and increases the risk of mastitis. Also, when chemotherapy is given after birth, many of the drugs appear in breast milk, which could harm the baby. Veterinary oncology, concentrating mainly on cats and dogs, is a growing specialty in wealthy countries and the major forms of human treatment such as surgery and radiotherapy may be offered. The most common types of cancer differ, but the cancer burden seems at least as high in pets as in humans. Animals, typically rodents, are often used in cancer research and studies of natural cancers in larger animals may benefit research into human cancer. Across wild animals, there is still limited data on cancer. Nonetheless, a study published in 2022, explored cancer risk in (non-domesticated) zoo mammals, belonging to 191 species, 110,148 individual, demonstrated that cancer is a ubiquitous disease of mammals and it can emerge anywhere along the mammalian phylogeny. This research also highlighted that cancer risk is not uniformly distributed along mammals. For instance, species in the order Carnivora are particularly prone to be affected by cancer (e.g. over 25% of clouded leopards, bat-eared foxes and red wolves die of cancer), while ungulates (especially even-toed ungulates) appear to face consistently low cancer risks. In non-humans, a few types of transmissible cancer have also been described, wherein the cancer spreads between animals by transmission of the tumor cells themselves. This phenomenon is seen in dogs with Sticker's sarcoma (also known as canine transmissible venereal tumor), and in Tasmanian devils with devil facial tumour disease (DFTD).
[[File:View from the Window at Le Gras, Joseph Nicéphore Niépce, uncompressed UMN source.png|thumb|"View from the Window at Le Gras" by [[Joseph Nicéphore Niépce]] was taken in 1826, and is the oldest known photograph.]] '''Photography''' is a way of making a [[picture]] using a [[camera]].<ref name="webster">{{cite web|url=http://www.merriam-webster.com/dictionary/photography|title=Photography|publisher=Merriam-Webster|date=2009-08-24}}</ref> A person who makes [[Picture|pictures]] using a camera is called a ''[[photographer]]''.<ref name="webster"/> A picture made using a camera is called a photograph or photo.<ref name="webster"/> Photography became popular in the middle 19th century with [[Daguerreotype]]. Later wet plate and [[dry plate]] methods were invented. Most photography in the [[20th century]] was on [[photographic film]] and most in the 21st uses [[digital camera]]s. == Camera == [[File:Colonel William Willoughby Verner, Sanger Shepherd process, by Sarah Acland 1903.png|thumb|left|Color photography was done long before color film, as seen in this 1903 portrait by [[Sarah Angelina Acland]]. In its early years, the need for special equipment, long exposures, and complex printing processes made it rare.|alt=]] {{Main|Camera}} A [[camera]] is, in its simplest form, a box with a hole in its front. There is a special piece of glass in front of the hole called a [[Lens (optics)|lens]]. To take a photograph, the lens makes a small picture of the object inside the camera, which it does by [[focus (optics)|focus]]ing light. A lens in a camera works like a lens in glasses ([[spectacles]]) or a [[magnifying glass]]. One type of camera called the [[Pinhole camera]] has no lens but uses a very small hole to focus light. To make a photograph with a camera, the ''shutter release'' button is pressed. Pressing the button opens the ''shutter''. The [[shutter (photography)|shutter]] is like a door. It covers the hole in the camera box. The shutter is behind the lens. When the shutter is closed no light can enter the camera box. When the shutter is open light can enter the camera. When the button is pressed the shutter opens and then closes. This happens very fast. The amount of time the shutter stays open for is called the ''shutter speed''. The shutter speed can change between 1/1000th of a second (0.001 s) to a few seconds. Normally the time taken for the shutter to open and close is far less than 1 second. Some cameras have an ''[[aperture]] ring''. It controls how much light enters the camera box. Photography is the art and practice of creating images by recording light on a photosensitive surface, such as [https://dev-anuphotography.pantheonsite.io/what-is-photography/ photographic] {{Webarchive|url=https://web.archive.org/web/20231227170739/https://dev-anuphotography.pantheonsite.io/what-is-photography/ |date=2023-12-27 }} film or a digital sensor. The word “shooting” comes from the Greek words “photo’s,” meaning light, and “graphs,” meaning drawing or writing. It essentially means “drawing with light.”[https://dev-anuphotography.pantheonsite.io/what-is-photography/] {{Webarchive|url=https://web.archive.org/web/20231227170739/https://dev-anuphotography.pantheonsite.io/what-is-photography/ |date=2023-12-27 }} The photograph in a camera may be made on film or, if it is a digital camera, on an [[electronics|electronic]] sensor. == Film == [[File:2001 Canon EOS 1v 2013 CP+.jpg|thumb|A Film Camera]] [[File:Bigapple.jpg|thumb|[[Manhattan]], [[New York City]] seen from the top of the now defunct [[World Trade Center]] with an [[apple]] in the foreground, an allusion to the city's nickname, ''Big Apple''.]] The picture the lens makes is recorded on [[photographic film]]. Film is placed inside the camera box. Light coming through the lens, aperture and open shutter shines on the film. Photographic film is coated with chemicals that react when light shines on it. Letting light shine on the film is called ''exposing'' the film. There are many different types of photographic film. There are films for taking colour photographs and films for taking black and white photographs. There are different sizes of film. The most common size is ''35 mm''. It is called 35&nbsp;mm because the width of the film is 35 millimetres. Another difference between films is how sensitive they are to light. Films have a number, called an ''ISO number'', which tells how fast a film reacts when light shines on it. == Film Processing == {{main|Film processing}} Once the film has been exposed it is ''processed''. Processing has to be done in total darkness or the film will be exposed too much and the picture will be lost. Processing stops the film reacting to light any more. After the film has been processed the picture can be seen on the film. A photographic ''print'' is a photograph made on paper. A light sensitive paper is used. The picture on the film is placed in an ''enlarger''. An enlarger is a machine that shines light through the film and makes a bigger picture on the light sensitive paper. A chemical reaction happens in the paper, which turns the areas hit with light black when the paper is 'developed'. (The more light, the darker the area.) ''Developing'' makes the picture appear on the paper – now it is a photograph. Then the paper is put into other chemicals that make it not sensitive to light any more. This is called "fixing". Last, the paper is washed so that there are no more chemicals on it and then dried. Then it is finished. == Digital photography == {{main|Digital camera}} [[file:M8JI1.jpg|thumb|Digital Camera]] Digital photography uses a [[digital camera]]. Sometimes it is called ''digital imaging''. Like other cameras a digital camera has a lens, aperture, and [[Shutter (photography)|shutter]]. The picture the lens makes is recorded by a light-sensitive [[Electronics|electronic]] [[sensor]]. A digital camera does not use photographic film to record a picture. Digital photographs are stored in storage devices such as [[SD card]]s. They can later be transferred to a computer. Paper prints can also be made from digital pictures. Digital cameras are also not expensive to use, as there is no film to buy. == Taking a photograph == One of the most important things when taking a photograph is focusing the lens. If the lens is not focused well, the photograph will be blurry. [[Autofocus]] cameras focus automatically when the shutter release is pressed. There are also ''manual'' focus cameras (usually older ones). Three other things are important when taking a photograph. These control how bright or dark the photograph will be. The brighter you have it the more transparent the picture comes out, if you have it to bright the whole screen can turn out to be all white, if the camera settings are set to be to dark, well, then the picture will come out to dark. So it really depends on how you want the picture, some cameras automatically come in with the setting at 50%, this is most recommended if you want to take a normal picture, but if you want to adjust the settings, it is recommended to keep it between 45% and 75%. # The shutter speed – how long the shutter is open for. This is written in the form "1/400," or one four-hundredth of a second. # The [[aperture]] – how big the aperture in the lens opens. This changes how much light is let in. This is written in the form "f/5.6," which describes the ratio between focal length and the size of the aperture opening. # The [[film speed]] – how quickly the film/sensor records the picture. This is also called ISO, and is written in the form "400." A slower shutter speed, a bigger aperture, and faster film/higher ISO sensor all make a brighter picture. A faster shutter speed, a smaller aperture, and a slower film/lower ISO sensor all make a darker picture. A good picture is not too bright and not too dark. When it is too bright it would be called "[[overexposed]]". An automatic camera changes these things by itself when the shutter release is pressed. ==Camera types== Though there are many types of cameras, all include five indispensable components: * the camera box, which holds and protects the sensitive film from all light except that entering through the lens * film, on which the image is recorded, a light-sensitive strip usually wound on a spool, either manually or automatically, as successive pictures are taken * the light control, consisting of an aperture or diaphragm and a shutter, both often adjustable * the lens, which focuses the light rays from the subject onto the film, creating the image, and which is usually adjustable by moving forward or back, changing the focus * the viewing system, which may be separate from the lens system (usually above it) or may operate through it by means of a mirror ==References== {{commonscat}} {{reflist}} {{Photography}} [[Category:Photography| ]]
Photography is the art, application, and practice of creating images by recording light, either electronically by means of an image sensor, or chemically by means of a light-sensitive material such as photographic film. It is employed in many fields of science, manufacturing (e.g., photolithography), and business, as well as its more direct uses for art, film and video production, recreational purposes, hobby, and mass communication. Typically, a lens is used to focus the light reflected or emitted from objects into a real image on the light-sensitive surface inside a camera during a timed exposure. With an electronic image sensor, this produces an electrical charge at each pixel, which is electronically processed and stored in a digital image file for subsequent display or processing. The result with photographic emulsion is an invisible latent image, which is later chemically "developed" into a visible image, either negative or positive, depending on the purpose of the photographic material and the method of processing. A negative image on film is traditionally used to photographically create a positive image on a paper base, known as a print, either by using an enlarger or by contact printing. The word "photography" was created from the Greek roots φωτός (phōtós), genitive of φῶς (phōs), "light" and γραφή (graphé) "representation by means of lines" or "drawing", together meaning "drawing with light". Several people may have coined the same new term from these roots independently. Hércules Florence, a French painter and inventor living in Campinas, Brazil, used the French form of the word, photographie, in private notes which a Brazilian historian believes were written in 1834. This claim is widely reported but is not yet largely recognized internationally. The first use of the word by Florence became widely known after the research of Boris Kossoy in 1980. The German newspaper Vossische Zeitung of 25 February 1839 contained an article entitled Photographie, discussing several priority claims – especially Henry Fox Talbot's – regarding Daguerre's claim of invention. The article is the earliest known occurrence of the word in public print. It was signed "J.M.", believed to have been Berlin astronomer Johann von Maedler. The astronomer John Herschel is also credited with coining the word, independent of Talbot, in 1839. The inventors Nicéphore Niépce, Talbot, and Louis Daguerre seem not to have known or used the word "photography", but referred to their processes as "Heliography" (Niépce), "Photogenic Drawing"/"Talbotype"/"Calotype" (Talbot), and "Daguerreotype" (Daguerre). Photography is the result of combining several technical discoveries, relating to seeing an image and capturing the image. The discovery of the camera obscura ("dark chamber" in Latin) that provides an image of a scene dates back to ancient China. Greek mathematicians Aristotle and Euclid independently described a camera obscura in the 5th and 4th centuries BCE. In the 6th century CE, Byzantine mathematician Anthemius of Tralles used a type of camera obscura in his experiments. The Arab physicist Ibn al-Haytham (Alhazen) (965–1040) also invented a camera obscura as well as the first true pinhole camera. The invention of the camera has been traced back to the work of Ibn al-Haytham. While the effects of a single light passing through a pinhole had been described earlier, Ibn al-Haytham gave the first correct analysis of the camera obscura, including the first geometrical and quantitative descriptions of the phenomenon, and was the first to use a screen in a dark room so that an image from one side of a hole in the surface could be projected onto a screen on the other side. He also first understood the relationship between the focal point and the pinhole, and performed early experiments with afterimages, laying the foundations for the invention of photography in the 19th century. Leonardo da Vinci mentions natural camerae obscurae that are formed by dark caves on the edge of a sunlit valley. A hole in the cave wall will act as a pinhole camera and project a laterally reversed, upside down image on a piece of paper. Renaissance painters used the camera obscura which, in fact, gives the optical rendering in color that dominates Western Art. It is a box with a small hole in one side, which allows specific light rays to enter, projecting an inverted image onto a viewing screen or paper. The birth of photography was then concerned with inventing means to capture and keep the image produced by the camera obscura. Albertus Magnus (1193–1280) discovered silver nitrate, and Georg Fabricius (1516–1571) discovered silver chloride, and the techniques described in Ibn al-Haytham's Book of Optics are capable of producing primitive photographs using medieval materials. Daniele Barbaro described a diaphragm in 1566. Wilhelm Homberg described how light darkened some chemicals (photochemical effect) in 1694. The fiction book Giphantie, published in 1760, by French author Tiphaigne de la Roche, described what can be interpreted as photography. In June 1802, British inventor Thomas Wedgwood made the first known attempt to capture the image in a camera obscura by means of a light-sensitive substance. He used paper or white leather treated with silver nitrate. Although he succeeded in capturing the shadows of objects placed on the surface in direct sunlight, and even made shadow copies of paintings on glass, it was reported in 1802 that "the images formed by means of a camera obscura have been found too faint to produce, in any moderate time, an effect upon the nitrate of silver." The shadow images eventually darkened all over. The first permanent photoetching was an image produced in 1822 by the French inventor Nicéphore Niépce, but it was destroyed in a later attempt to make prints from it. Niépce was successful again in 1825. In 1826 he made the View from the Window at Le Gras, the earliest surviving photograph from nature (i.e., of the image of a real-world scene, as formed in a camera obscura by a lens). Because Niépce's camera photographs required an extremely long exposure (at least eight hours and probably several days), he sought to greatly improve his bitumen process or replace it with one that was more practical. In partnership with Louis Daguerre, he worked out post-exposure processing methods that produced visually superior results and replaced the bitumen with a more light-sensitive resin, but hours of exposure in the camera were still required. With an eye to eventual commercial exploitation, the partners opted for total secrecy. Niépce died in 1833 and Daguerre then redirected the experiments toward the light-sensitive silver halides, which Niépce had abandoned many years earlier because of his inability to make the images he captured with them light-fast and permanent. Daguerre's efforts culminated in what would later be named the daguerreotype process. The essential elements—a silver-plated surface sensitized by iodine vapor, developed by mercury vapor, and "fixed" with hot saturated salt water—were in place in 1837. The required exposure time was measured in minutes instead of hours. Daguerre took the earliest confirmed photograph of a person in 1838 while capturing a view of a Paris street: unlike the other pedestrian and horse-drawn traffic on the busy boulevard, which appears deserted, one man having his boots polished stood sufficiently still throughout the several-minutes-long exposure to be visible. The existence of Daguerre's process was publicly announced, without details, on 7 January 1839. The news created an international sensation. France soon agreed to pay Daguerre a pension in exchange for the right to present his invention to the world as the gift of France, which occurred when complete working instructions were unveiled on 19 August 1839. In that same year, American photographer Robert Cornelius is credited with taking the earliest surviving photographic self-portrait. In Brazil, Hercules Florence had apparently started working out a silver-salt-based paper process in 1832, later naming it Photographie. Meanwhile, a British inventor, William Fox Talbot, had succeeded in making crude but reasonably light-fast silver images on paper as early as 1834 but had kept his work secret. After reading about Daguerre's invention in January 1839, Talbot published his hitherto secret method and set about improving on it. At first, like other pre-daguerreotype processes, Talbot's paper-based photography typically required hours-long exposures in the camera, but in 1840 he created the calotype process, which used the chemical development of a latent image to greatly reduce the exposure needed and compete with the daguerreotype. In both its original and calotype forms, Talbot's process, unlike Daguerre's, created a translucent negative which could be used to print multiple positive copies; this is the basis of most modern chemical photography up to the present day, as daguerreotypes could only be replicated by rephotographing them with a camera. Talbot's famous tiny paper negative of the Oriel window in Lacock Abbey, one of a number of camera photographs he made in the summer of 1835, may be the oldest camera negative in existence. In France, Hippolyte Bayard invented his own process for producing direct positive paper prints and claimed to have invented photography earlier than Daguerre or Talbot. British chemist John Herschel made many contributions to the new field. He invented the cyanotype process, later familiar as the "blueprint". He was the first to use the terms "photography", "negative" and "positive". He had discovered in 1819 that sodium thiosulphate was a solvent of silver halides, and in 1839 he informed Talbot (and, indirectly, Daguerre) that it could be used to "fix" silver-halide-based photographs and make them completely light-fast. He made the first glass negative in late 1839. In the March 1851 issue of The Chemist, Frederick Scott Archer published his wet plate collodion process. It became the most widely used photographic medium until the gelatin dry plate, introduced in the 1870s, eventually replaced it. There are three subsets to the collodion process; the Ambrotype (a positive image on glass), the Ferrotype or Tintype (a positive image on metal) and the glass negative, which was used to make positive prints on albumen or salted paper. Many advances in photographic glass plates and printing were made during the rest of the 19th century. In 1891, Gabriel Lippmann introduced a process for making natural-color photographs based on the optical phenomenon of the interference of light waves. His scientifically elegant and important but ultimately impractical invention earned him the Nobel Prize in Physics in 1908. Glass plates were the medium for most original camera photography from the late 1850s until the general introduction of flexible plastic films during the 1890s. Although the convenience of the film greatly popularized amateur photography, early films were somewhat more expensive and of markedly lower optical quality than their glass plate equivalents, and until the late 1910s they were not available in the large formats preferred by most professional photographers, so the new medium did not immediately or completely replace the old. Because of the superior dimensional stability of glass, the use of plates for some scientific applications, such as astrophotography, continued into the 1990s, and in the niche field of laser holography, it has persisted into the 21st century. Hurter and Driffield began pioneering work on the light sensitivity of photographic emulsions in 1876. Their work enabled the first quantitative measure of film speed to be devised. The first flexible photographic roll film was marketed by George Eastman, founder of Kodak in 1885, but this original "film" was actually a coating on a paper base. As part of the processing, the image-bearing layer was stripped from the paper and transferred to a hardened gelatin support. The first transparent plastic roll film followed in 1889. It was made from highly flammable nitrocellulose known as nitrate film. Although cellulose acetate or "safety film" had been introduced by Kodak in 1908, at first it found only a few special applications as an alternative to the hazardous nitrate film, which had the advantages of being considerably tougher, slightly more transparent, and cheaper. The changeover was not completed for X-ray films until 1933, and although safety film was always used for 16 mm and 8 mm home movies, nitrate film remained standard for theatrical 35 mm motion pictures until it was finally discontinued in 1951. Films remained the dominant form of photography until the early 21st century when advances in digital photography drew consumers to digital formats. Although modern photography is dominated by digital users, film continues to be used by enthusiasts and professional photographers. The distinctive "look" of film based photographs compared to digital images is likely due to a combination of factors, including (1) differences in spectral and tonal sensitivity (S-shaped density-to-exposure (H&D curve) with film vs. linear response curve for digital CCD sensors), (2) resolution, and (3) continuity of tone. Originally, all photography was monochrome, or black-and-white. Even after color film was readily available, black-and-white photography continued to dominate for decades, due to its lower cost, chemical stability, and its "classic" photographic look. The tones and contrast between light and dark areas define black-and-white photography. Monochromatic pictures are not necessarily composed of pure blacks, whites, and intermediate shades of gray but can involve shades of one particular hue depending on the process. The cyanotype process, for example, produces an image composed of blue tones. The albumen print process, publicly revealed in 1847, produces brownish tones. Many photographers continue to produce some monochrome images, sometimes because of the established archival permanence of well-processed silver-halide-based materials. Some full-color digital images are processed using a variety of techniques to create black-and-white results, and some manufacturers produce digital cameras that exclusively shoot monochrome. Monochrome printing or electronic display can be used to salvage certain photographs taken in color which are unsatisfactory in their original form; sometimes when presented as black-and-white or single-color-toned images they are found to be more effective. Although color photography has long predominated, monochrome images are still produced, mostly for artistic reasons. Almost all digital cameras have an option to shoot in monochrome, and almost all image editing software can combine or selectively discard RGB color channels to produce a monochrome image from one shot in color. Color photography was explored beginning in the 1840s. Early experiments in color required extremely long exposures (hours or days for camera images) and could not "fix" the photograph to prevent the color from quickly fading when exposed to white light. The first permanent color photograph was taken in 1861 using the three-color-separation principle first published by Scottish physicist James Clerk Maxwell in 1855. The foundation of virtually all practical color processes, Maxwell's idea was to take three separate black-and-white photographs through red, green and blue filters. This provides the photographer with the three basic channels required to recreate a color image. Transparent prints of the images could be projected through similar color filters and superimposed on the projection screen, an additive method of color reproduction. A color print on paper could be produced by superimposing carbon prints of the three images made in their complementary colors, a subtractive method of color reproduction pioneered by Louis Ducos du Hauron in the late 1860s. Russian photographer Sergei Mikhailovich Prokudin-Gorskii made extensive use of this color separation technique, employing a special camera which successively exposed the three color-filtered images on different parts of an oblong plate. Because his exposures were not simultaneous, unsteady subjects exhibited color "fringes" or, if rapidly moving through the scene, appeared as brightly colored ghosts in the resulting projected or printed images. Implementation of color photography was hindered by the limited sensitivity of early photographic materials, which were mostly sensitive to blue, only slightly sensitive to green, and virtually insensitive to red. The discovery of dye sensitization by photochemist Hermann Vogel in 1873 suddenly made it possible to add sensitivity to green, yellow and even red. Improved color sensitizers and ongoing improvements in the overall sensitivity of emulsions steadily reduced the once-prohibitive long exposure times required for color, bringing it ever closer to commercial viability. Autochrome, the first commercially successful color process, was introduced by the Lumière brothers in 1907. Autochrome plates incorporated a mosaic color filter layer made of dyed grains of potato starch, which allowed the three color components to be recorded as adjacent microscopic image fragments. After an Autochrome plate was reversal processed to produce a positive transparency, the starch grains served to illuminate each fragment with the correct color and the tiny colored points blended together in the eye, synthesizing the color of the subject by the additive method. Autochrome plates were one of several varieties of additive color screen plates and films marketed between the 1890s and the 1950s. Kodachrome, the first modern "integral tripack" (or "monopack") color film, was introduced by Kodak in 1935. It captured the three color components in a multi-layer emulsion. One layer was sensitized to record the red-dominated part of the spectrum, another layer recorded only the green part and a third recorded only the blue. Without special film processing, the result would simply be three superimposed black-and-white images, but complementary cyan, magenta, and yellow dye images were created in those layers by adding color couplers during a complex processing procedure. Agfa's similarly structured Agfacolor Neu was introduced in 1936. Unlike Kodachrome, the color couplers in Agfacolor Neu were incorporated into the emulsion layers during manufacture, which greatly simplified the processing. Currently, available color films still employ a multi-layer emulsion and the same principles, most closely resembling Agfa's product. Instant color film, used in a special camera which yielded a unique finished color print only a minute or two after the exposure, was introduced by Polaroid in 1963. Color photography may form images as positive transparencies, which can be used in a slide projector, or as color negatives intended for use in creating positive color enlargements on specially coated paper. The latter is now the most common form of film (non-digital) color photography owing to the introduction of automated photo printing equipment. After a transition period centered around 1995–2005, color film was relegated to a niche market by inexpensive multi-megapixel digital cameras. Film continues to be the preference of some photographers because of its distinctive "look". In 1981, Sony unveiled the first consumer camera to use a charge-coupled device for imaging, eliminating the need for film: the Sony Mavica. While the Mavica saved images to disk, the images were displayed on television, and the camera was not fully digital. The first digital camera to both record and save images in a digital format was the Fujix DS-1P created by Fujfilm in 1988. In 1991, Kodak unveiled the DCS 100, the first commercially available digital single-lens reflex camera. Although its high cost precluded uses other than photojournalism and professional photography, commercial digital photography was born. Digital imaging uses an electronic image sensor to record the image as a set of electronic data rather than as chemical changes on film. An important difference between digital and chemical photography is that chemical photography resists photo manipulation because it involves film and photographic paper, while digital imaging is a highly manipulative medium. This difference allows for a degree of image post-processing that is comparatively difficult in film-based photography and permits different communicative potentials and applications. Digital photography dominates the 21st century. More than 99% of photographs taken around the world are through digital cameras, increasingly through smartphones. A large variety of photographic techniques and media are used in the process of capturing images for photography. These include the camera; dualphotography; full-spectrum, ultraviolet and infrared media; light field photography; and other imaging techniques. The camera is the image-forming device, and a photographic plate, photographic film or a silicon electronic image sensor is the capture medium. The respective recording medium can be the plate or film itself, or a digital magnetic or electronic memory. Photographers control the camera and lens to "expose" the light recording material to the required amount of light to form a "latent image" (on plate or film) or RAW file (in digital cameras) which, after appropriate processing, is converted to a usable image. Digital cameras use an electronic image sensor based on light-sensitive electronics such as charge-coupled device (CCD) or complementary metal–oxide–semiconductor (CMOS) technology. The resulting digital image is stored electronically, but can be reproduced on a paper. The camera (or 'camera obscura') is a dark room or chamber from which, as far as possible, all light is excluded except the light that forms the image. It was discovered and used in the 16th century by painters. The subject being photographed, however, must be illuminated. Cameras can range from small to very large, a whole room that is kept dark while the object to be photographed is in another room where it is properly illuminated. This was common for reproduction photography of flat copy when large film negatives were used (see Process camera). As soon as photographic materials became "fast" (sensitive) enough for taking candid or surreptitious pictures, small "detective" cameras were made, some actually disguised as a book or handbag or pocket watch (the Ticka camera) or even worn hidden behind an Ascot necktie with a tie pin that was really the lens. The movie camera is a type of photographic camera that takes a rapid sequence of photographs on recording medium. In contrast to a still camera, which captures a single snapshot at a time, the movie camera takes a series of images, each called a "frame". This is accomplished through an intermittent mechanism. The frames are later played back in a movie projector at a specific speed, called the "frame rate" (number of frames per second). While viewing, a person's eyes and brain merge the separate pictures to create the illusion of motion. Photographs, both monochrome and color, can be captured and displayed through two side-by-side images that emulate human stereoscopic vision. Stereoscopic photography was the first that captured figures in motion. While known colloquially as "3-D" photography, the more accurate term is stereoscopy. Such cameras have long been realized by using film and more recently in digital electronic methods (including cell phone cameras). Dualphotography consists of photographing a scene from both sides of a photographic device at once (e.g. camera for back-to-back dualphotography, or two networked cameras for portal-plane dualphotography). The dualphoto apparatus can be used to simultaneously capture both the subject and the photographer, or both sides of a geographical place at once, thus adding a supplementary narrative layer to that of a single image. Ultraviolet and infrared films have been available for many decades and employed in a variety of photographic avenues since the 1960s. New technological trends in digital photography have opened a new direction in full spectrum photography, where careful filtering choices across the ultraviolet, visible and infrared lead to new artistic visions. Modified digital cameras can detect some ultraviolet, all of the visible and much of the near infrared spectrum, as most digital imaging sensors are sensitive from about 350 nm to 1000 nm. An off-the-shelf digital camera contains an infrared hot mirror filter that blocks most of the infrared and a bit of the ultraviolet that would otherwise be detected by the sensor, narrowing the accepted range from about 400 nm to 700 nm. Replacing a hot mirror or infrared blocking filter with an infrared pass or a wide spectrally transmitting filter allows the camera to detect the wider spectrum light at greater sensitivity. Without the hot-mirror, the red, green and blue (or cyan, yellow and magenta) colored micro-filters placed over the sensor elements pass varying amounts of ultraviolet (blue window) and infrared (primarily red and somewhat lesser the green and blue micro-filters). Uses of full spectrum photography are for fine art photography, geology, forensics and law enforcement. Layering is a photographic composition technique that manipulates the foreground, subject or middle-ground, and background layers in a way that they all work together to tell a story through the image. Layers may be incorporated by altering the focal length, distorting the perspective by positioning the camera in a certain spot. People, movement, light and a variety of objects can be used in layering. Digital methods of image capture and display processing have enabled the new technology of "light field photography" (also known as synthetic aperture photography). This process allows focusing at various depths of field to be selected after the photograph has been captured. As explained by Michael Faraday in 1846, the "light field" is understood as 5-dimensional, with each point in 3-D space having attributes of two more angles that define the direction of each ray passing through that point. These additional vector attributes can be captured optically through the use of microlenses at each pixel point within the 2-dimensional image sensor. Every pixel of the final image is actually a selection from each sub-array located under each microlens, as identified by a post-image capture focus algorithm. Besides the camera, other methods of forming images with light are available. For instance, a photocopy or xerography machine forms permanent images but uses the transfer of static electrical charges rather than photographic medium, hence the term electrophotography. Photograms are images produced by the shadows of objects cast on the photographic paper, without the use of a camera. Objects can also be placed directly on the glass of an image scanner to produce digital pictures. Amateur photographers take photos for personal use, as a hobby or out of casual interest, rather than as a business or job. The quality of amateur work can be comparable to that of many professionals. Amateurs can fill a gap in subjects or topics that might not otherwise be photographed if they are not commercially useful or salable. Amateur photography grew during the late 19th century due to the popularization of the hand-held camera. Twenty-first century social media and near-ubiquitous camera phones have made photographic and video recording pervasive in everyday life. In the mid-2010s smartphone cameras added numerous automatic assistance features like color management, autofocus face detection and image stabilization that significantly decreased skill and effort needed to take high quality images. Commercial photography is probably best defined as any photography for which the photographer is paid for images rather than works of art. In this light, money could be paid for the subject of the photograph or the photograph itself. Wholesale, retail, and professional uses of photography would fall under this definition. The commercial photographic world could include: During the 20th century, both fine art photography and documentary photography became accepted by the English-speaking art world and the gallery system. In the United States, a handful of photographers, including Alfred Stieglitz, Edward Steichen, John Szarkowski, F. Holland Day, and Edward Weston, spent their lives advocating for photography as a fine art. At first, fine art photographers tried to imitate painting styles. This movement is called Pictorialism, often using soft focus for a dreamy, 'romantic' look. In reaction to that, Weston, Ansel Adams, and others formed the Group f/64 to advocate 'straight photography', the photograph as a (sharply focused) thing in itself and not an imitation of something else. The aesthetics of photography is a matter that continues to be discussed regularly, especially in artistic circles. Many artists argued that photography was the mechanical reproduction of an image. If photography is authentically art, then photography in the context of art would need redefinition, such as determining what component of a photograph makes it beautiful to the viewer. The controversy began with the earliest images "written with light"; Nicéphore Niépce, Louis Daguerre, and others among the very earliest photographers were met with acclaim, but some questioned if their work met the definitions and purposes of art. Clive Bell in his classic essay Art states that only "significant form" can distinguish art from what is not art. There must be some one quality without which a work of art cannot exist; possessing which, in the least degree, no work is altogether worthless. What is this quality? What quality is shared by all objects that provoke our aesthetic emotions? What quality is common to Sta. Sophia and the windows at Chartres, Mexican sculpture, a Persian bowl, Chinese carpets, Giotto's frescoes at Padua, and the masterpieces of Poussin, Piero della Francesca, and Cezanne? Only one answer seems possible – significant form. In each, lines and colors combined in a particular way, certain forms and relations of forms, stir our aesthetic emotions. On 7 February 2007, Sotheby's London sold the 2001 photograph 99 Cent II Diptychon for an unprecedented $3,346,456 to an anonymous bidder, making it the most expensive at the time. Conceptual photography turns a concept or idea into a photograph. Even though what is depicted in the photographs are real objects, the subject is strictly abstract. In parallel to this development, the then largely separate interface between painting and photography was closed in the second half of the 20th century with the chemigram of Pierre Cordier and the chemogram of Josef H. Neumann. In 1974 the chemograms by Josef H. Neumann concluded the separation of the painterly background and the photographic layer by showing the picture elements in a symbiosis that had never existed before, as an unmistakable unique specimen, in a simultaneous painterly and at the same time real photographic perspective, using lenses, within a photographic layer, united in colors and shapes. This Neumann chemogram from the seventies of the 20th century thus differs from the beginning of the previously created cameraless chemigrams of a Pierre Cordier and the photogram Man Ray or László Moholy-Nagy of the previous decades. These works of art were almost simultaneous with the invention of photography by various important artists who characterized Hippolyte Bayard, Thomas Wedgwood, William Henry Fox Talbot in their early stages, and later Man Ray and László Moholy-Nagy in the twenties and by the painter in the thirties Edmund Kesting and Christian Schad by draping objects directly onto appropriately sensitized photo paper and using a light source without a camera. Photojournalism is a particular form of photography (the collecting, editing, and presenting of news material for publication or broadcast) that employs images in order to tell a news story. It is now usually understood to refer only to still images, but in some cases the term also refers to video used in broadcast journalism. Photojournalism is distinguished from other close branches of photography (e.g., documentary photography, social documentary photography, street photography or celebrity photography) by complying with a rigid ethical framework which demands that the work be both honest and impartial whilst telling the story in strictly journalistic terms. Photojournalists create pictures that contribute to the news media, and help communities connect with one other. Photojournalists must be well informed and knowledgeable about events happening right outside their door. They deliver news in a creative format that is not only informative, but also entertaining, including sports photography. The camera has a long and distinguished history as a means of recording scientific phenomena from the first use by Daguerre and Fox-Talbot, such as astronomical events (eclipses for example), small creatures and plants when the camera was attached to the eyepiece of microscopes (in photomicroscopy) and for macro photography of larger specimens. The camera also proved useful in recording crime scenes and the scenes of accidents, such as the Wootton bridge collapse in 1861. The methods used in analysing photographs for use in legal cases are collectively known as forensic photography. Crime scene photos are usually taken from three vantage points: overview, mid-range, and close-up. In 1845 Francis Ronalds, the Honorary Director of the Kew Observatory, invented the first successful camera to make continuous recordings of meteorological and geomagnetic parameters. Different machines produced 12- or 24- hour photographic traces of the minute-by-minute variations of atmospheric pressure, temperature, humidity, atmospheric electricity, and the three components of geomagnetic forces. The cameras were supplied to numerous observatories around the world and some remained in use until well into the 20th century. Charles Brooke a little later developed similar instruments for the Greenwich Observatory. Science regularly uses image technology that has derived from the design of the pinhole camera to avoid distortions that can be caused by lenses. X-ray machines are similar in design to pinhole cameras, with high-grade filters and laser radiation. Photography has become universal in recording events and data in science and engineering, and at crime scenes or accident scenes. The method has been much extended by using other wavelengths, such as infrared photography and ultraviolet photography, as well as spectroscopy. Those methods were first used in the Victorian era and improved much further since that time. The first photographed atom was discovered in 2012 by physicists at Griffith University, Australia. They used an electric field to trap an "Ion" of the element, Ytterbium. The image was recorded on a CCD, an electronic photographic film. Wildlife photography involves capturing images of various forms of wildlife. Unlike other forms of photography such as product or food photography, successful wildlife photography requires a photographer to choose the right place and right time when specific wildlife are present and active. It often requires great patience and considerable skill and command of the right photographic equipment. There are many ongoing questions about different aspects of photography. In her On Photography (1977), Susan Sontag dismisses the objectivity of photography. This is a highly debated subject within the photographic community. Sontag argues, "To photograph is to appropriate the thing photographed. It means putting one's self into a certain relation to the world that feels like knowledge, and therefore like power." Photographers decide what to take a photo of, what elements to exclude and what angle to frame the photo, and these factors may reflect a particular socio-historical context. Along these lines, it can be argued that photography is a subjective form of representation. Modern photography has raised a number of concerns on its effect on society. In Alfred Hitchcock's Rear Window (1954), the camera is presented as promoting voyeurism. 'Although the camera is an observation station, the act of photographing is more than passive observing'. The camera doesn't rape or even possess, though it may presume, intrude, trespass, distort, exploit, and, at the farthest reach of metaphor, assassinate – all activities that, unlike the sexual push and shove, can be conducted from a distance, and with some detachment. Digital imaging has raised ethical concerns because of the ease of manipulating digital photographs in post-processing. Many photojournalists have declared they will not crop their pictures or are forbidden from combining elements of multiple photos to make "photomontages", passing them as "real" photographs. Today's technology has made image editing relatively simple for even the novice photographer. However, recent changes of in-camera processing allow digital fingerprinting of photos to detect tampering for purposes of forensic photography. Photography is one of the new media forms that changes perception and changes the structure of society. Further unease has been caused around cameras in regards to desensitization. Fears that disturbing or explicit images are widely accessible to children and society at large have been raised. Particularly, photos of war and pornography are causing a stir. Sontag is concerned that "to photograph is to turn people into objects that can be symbolically possessed". Desensitization discussion goes hand in hand with debates about censored images. Sontag writes of her concern that the ability to censor pictures means the photographer has the ability to construct reality. One of the practices through which photography constitutes society is tourism. Tourism and photography combine to create a "tourist gaze" in which local inhabitants are positioned and defined by the camera lens. However, it has also been argued that there exists a "reverse gaze" through which indigenous photographees can position the tourist photographer as a shallow consumer of images. Photography is both restricted and protected by the law in many jurisdictions. Protection of photographs is typically achieved through the granting of copyright or moral rights to the photographer. In the United States, photography is protected as a First Amendment right and anyone is free to photograph anything seen in public spaces as long as it is in plain view. In the UK, a recent law (Counter-Terrorism Act 2008) increases the power of the police to prevent people, even press photographers, from taking pictures in public places. In South Africa, any person may photograph any other person, without their permission, in public spaces and the only specific restriction placed on what may not be photographed by government is related to anything classed as national security. Each country has different laws.
[[File:Dipole xmting antenna animation 4 408x318x150ms.gif|thumb|Diagram of an antenna emitting radio waves]] '''Radio waves''' make up part of the [[electromagnetic spectrum]]. These [[wave (physics)|wave]]s are packets of energy with differing [[wavelength]]s, similar to [[visible light]] waves, [[X-ray]]s or [[gamma ray]]s, except much longer. A radio wave, like other [[electromagnetic wave]]s, is similar to an [[ocean surface wave]] or any other type of wave. Both types of wave have a hill and valley shape, repeating over and over. A [[wavelength]] is measured as the distance from the top of one crest to the top of its neighboring crest. While the wavelength of visible light is very very small, less than one micrometer and much less than the thickness of a human hair, radio waves can have a wavelength from a couple [[centimeter]]s to several [[metre|meter]]s. They also have a [[radio frequency]]. The smallest radio waves are called [[microwave]]s. [[Shortwave]]s are not quite so small. There are also medium and long waves. [[Antenna]]s designed to send and receive radio waves are usually similar in size to the wavelength they are to use. Many radio antennas (like those on cars) are made long because they receive signals of [[FM radio]] (a few meters, several feet) or [[AM radio]] (hundreds of meters, about a thousand feet). == Uses == [[Manmade]] radio waves have been used since the 19th century for [[communication]]. [[Radar]] was developed in the 20th century, using radio waves to 'see' distant objects by [[Reflection|bouncing]] waves off an object and seeing how long it takes for the waves to return. [[Radio]]s also use these waves to send and receive information. Radio waves from other planets were first discovered in the 1930s by [[Karl Guthe Jansky]], working for [[Bell Laboratories]]. Bell was detecting [[noise (electronics)]] on radio channels, and had Jansky try to find the source of this static, or [[interference]]. After identifying noise that came from [[lightning]], he spent much time looking into the remainder. Surprisingly, some of the interference was coming from space! This discovery eventually led [[astronomer]]s to look at radio waves along with [[light wave]]s to find things in the sky. These [[radio astronomer]]s use giant [[Radio telescope]]s, shaped like [[satellite]] [[dish]]es, to gather and study the waves. Radio waves are used for many things today. [[Broadcasting]] and [[communications satellite]]s and [[mobile phone]]s and many [[computer]]s communicate by radio waves. [[Category:Wave physics]] [[Category:Radio technology]] [[Category:Electromagnetic radiation]]
Radio waves are a type of electromagnetic radiation with the longest wavelengths in the electromagnetic spectrum, typically with frequencies of 300 gigahertz (GHz) and below. At 300 GHz, the corresponding wavelength is 1mm, which is shorter than the diameter of a grain of rice. At 30 Hz the corresponding wavelength is ~10,000 kilometers (6,200 miles), which is longer than the radius of the Earth. Wavelength of a radio wave is inversely proportional to its frequency, because its velocity is constant. Like all electromagnetic waves, radio waves in a vacuum travel at the speed of light, and in the Earth's atmosphere at a slightly slower speed. Radio waves are generated by charged particles undergoing acceleration, such as time-varying electric currents. Naturally occurring radio waves are emitted by lightning and astronomical objects, and are part of the blackbody radiation emitted by all warm objects. Radio waves are generated artificially by an electronic device called a transmitter, which is connected to an antenna which radiates the waves. They are received by another antenna connected to a radio receiver, which processes the received signal. Radio waves are very widely used in modern technology for fixed and mobile radio communication, broadcasting, radar and radio navigation systems, communications satellites, wireless computer networks and many other applications. Different frequencies of radio waves have different propagation characteristics in the Earth's atmosphere; long waves can diffract around obstacles like mountains and follow the contour of the Earth (ground waves), shorter waves can reflect off the ionosphere and return to Earth beyond the horizon (skywaves), while much shorter wavelengths bend or diffract very little and travel on a line of sight, so their propagation distances are limited to the visual horizon. To prevent interference between different users, the artificial generation and use of radio waves is strictly regulated by law, coordinated by an international body called the International Telecommunication Union (ITU), which defines radio waves as "electromagnetic waves of frequencies arbitrarily lower than 3,000 GHz, propagated in space without artificial guide". The radio spectrum is divided into a number of radio bands on the basis of frequency, allocated to different uses. Radio waves were first predicted by the theory of electromagnetism proposed in 1867 by Scottish mathematical physicist James Clerk Maxwell. His mathematical theory, now called Maxwell's equations, predicted that a coupled electric and magnetic field could travel through space as an "electromagnetic wave". Maxwell proposed that light consisted of electromagnetic waves of very short wavelength. In 1887, German physicist Heinrich Hertz demonstrated the reality of Maxwell's electromagnetic waves by experimentally generating radio waves in his laboratory, showing that they exhibited the same wave properties as light: standing waves, refraction, diffraction, and polarization. Italian inventor Guglielmo Marconi developed the first practical radio transmitters and receivers around 1894–1895. He received the 1909 Nobel Prize in physics for his radio work. Radio communication began to be used commercially around 1900. The modern term "radio wave" replaced the original name "Hertzian wave" around 1912. Radio waves are radiated by charged particles when they are accelerated. Natural sources of radio waves include radio noise produced by lightning and other natural processes in the Earth's atmosphere, and astronomical radio sources in space such as the Sun, galaxies and nebulas. All warm objects radiate high frequency radio waves (microwaves) as part of their black body radiation. Radio waves are produced artificially by time-varying electric currents, consisting of electrons flowing back and forth in a specially-shaped metal conductor called an antenna. An electronic device called a radio transmitter applies oscillating electric current to the antenna, and the antenna radiates the power as radio waves. Radio waves are received by another antenna attached to a radio receiver. When radio waves strike the receiving antenna they push the electrons in the metal back and forth, creating tiny oscillating currents which are detected by the receiver. From quantum mechanics, like other electromagnetic radiation such as light, radio waves can alternatively be regarded as streams of uncharged elementary particles called photons. In an antenna transmitting radio waves, the electrons in the antenna emit the energy in discrete packets called radio photons, while in a receiving antenna the electrons absorb the energy as radio photons. An antenna is a coherent emitter of photons, like a laser, so the radio photons are all in phase. However, from Planck's relation E = h ν {\displaystyle E=h\nu } the energy of individual radio photons is extremely small, from 10 to 10 joules. So the antenna of even a very low power transmitter emits enormous numbers of photons per second. Therefore, except for certain molecular electron transition processes such as atoms in a maser emitting microwave photons, radio wave emission and absorption is usually regarded as a continuous classical process, governed by Maxwell's equations. Radio waves in a vacuum travel at the speed of light c {\displaystyle c} . When passing through a material medium, they are slowed depending on the medium's permeability and permittivity. Air is thin enough that in the Earth's atmosphere radio waves travel very close to the speed of light. The wavelength λ {\displaystyle \lambda } is the distance from one peak (crest) of the wave's electric field to the next, and is inversely proportional to the frequency f {\displaystyle f} of the wave. The relation of frequency and wavelength in a radio wave traveling in vacuum or air is where Equivalently, c {\displaystyle \;c\;} the distance a radio wave travels in a vacuum, in one second, is 299,792,458 meters (983,571,056 ft), which is the wavelength of a 1 Hertz radio signal. A 1 megahertz radio wave (mid-AM band) has a wavelength of 299.79 meters (983.6 ft). Like other electromagnetic waves, a radio wave has a property called polarization, which is defined as the direction of the wave's oscillating electric field perpendicular to the direction of motion. A plane polarized radio wave has an electric field which oscillates in a plane along the direction of motion. In a horizontally polarized radio wave the electric field oscillates in a horizontal direction. In a vertically polarized wave the electric field oscillates in a vertical direction. In a circularly polarized wave the electric field at any point rotates about the direction of travel, once per cycle. A right circularly polarized wave rotates in a right hand sense about the direction of travel, while a left circularly polarized wave rotates in the opposite sense. The wave's magnetic field is perpendicular to the electric field, and the electric and magnetic field are oriented in a right hand sense with respect to the direction of radiation. An antenna emits polarized radio waves, with the polarization determined by the direction of the metal antenna elements. For example a dipole antenna consists of two collinear metal rods. If the rods are horizontal it radiates horizontally polarized radio waves, while if the rods are vertical it radiates vertically polarized waves. An antenna receiving the radio waves must have the same polarization as the transmitting antenna, or it will suffer a severe loss of reception. Many natural sources of radio waves, such as the sun, stars and blackbody radiation from warm objects, emit unpolarized waves, consisting of incoherent short wave trains in an equal mixture of polarization states. The polarization of radio waves is determined by a quantum mechanical property of the photons called their spin. A photon can have one of two possible values of spin; it can spin in a right hand sense about its direction of motion, or in a left hand sense. Right circularly polarized radio waves consist of photons spinning in a right hand sense. Left circularly polarized radio waves consist of photons spinning in a left hand sense. Plane polarized radio waves consist of photons in a quantum superposition of right and left hand spin states. The electric field consists of a superposition of right and left rotating fields, resulting in a plane oscillation. Radio waves are more widely used for communication than other electromagnetic waves mainly because of their desirable propagation properties, stemming from their large wavelength. Radio waves have the ability to pass through the atmosphere in any weather, foliage, and most building materials, and by diffraction longer wavelengths can bend around obstructions, and unlike other electromagnetic waves they tend to be scattered rather than absorbed by objects larger than their wavelength. The study of radio propagation, how radio waves move in free space and over the surface of the Earth, is vitally important in the design of practical radio systems. Radio waves passing through different environments experience reflection, refraction, polarization, diffraction, and absorption. Different frequencies experience different combinations of these phenomena in the Earth's atmosphere, making certain radio bands more useful for specific purposes than others. Practical radio systems mainly use three different techniques of radio propagation to communicate: At microwave frequencies, atmospheric gases begin absorbing radio waves, so the range of practical radio communication systems decreases with increasing frequency. Below about 20 GHz atmospheric attenuation is mainly due to water vapor. Above 20 GHz, in the millimeter wave band, other atmospheric gases begin to absorb the waves, limiting practical transmission distances to a kilometer or less. Above 300 GHz, in the terahertz band, virtually all the power is absorbed within a few meters, so the atmosphere is effectively opaque. In radio communication systems, information is transported across space using radio waves. At the sending end, the information to be sent, in the form of a time-varying electrical signal, is applied to a radio transmitter. The information, called the modulation signal, can be an audio signal representing sound from a microphone, a video signal representing moving images from a video camera, or a digital signal representing data from a computer. In the transmitter, an electronic oscillator generates an alternating current oscillating at a radio frequency, called the carrier wave because it creates the radio waves that "carry" the information through the air. The information signal is used to modulate the carrier, altering some aspect of it, encoding the information on the carrier. The modulated carrier is amplified and applied to an antenna. The oscillating current pushes the electrons in the antenna back and forth, creating oscillating electric and magnetic fields, which radiate the energy away from the antenna as radio waves. The radio waves carry the information to the receiver location. At the receiver, the oscillating electric and magnetic fields of the incoming radio wave push the electrons in the receiving antenna back and forth, creating a tiny oscillating voltage which is a weaker replica of the current in the transmitting antenna. This voltage is applied to the radio receiver, which extracts the information signal. The receiver first uses a bandpass filter to separate the desired radio station's radio signal from all the other radio signals picked up by the antenna, then amplifies the signal so it is stronger, then finally extracts the information-bearing modulation signal in a demodulator. The recovered signal is sent to a loudspeaker or earphone to produce sound, or a television display screen to produce a visible image, or other devices. A digital data signal is applied to a computer or microprocessor, which interacts with a human user. The radio waves from many transmitters pass through the air simultaneously without interfering with each other. They can be separated in the receiver because each transmitter's radio waves oscillate at a different rate, in other words each transmitter has a different frequency, measured in kilohertz (kHz), megahertz (MHz) or gigahertz (GHz). The bandpass filter in the receiver consists of one or more tuned circuits which act like a resonator, similarly to a tuning fork. The tuned circuit has a natural resonant frequency at which it oscillates. The resonant frequency is set equal to the frequency of the desired radio station. The oscillating radio signal from the desired station causes the tuned circuit to oscillate in sympathy, and it passes the signal on to the rest of the receiver. Radio signals at other frequencies are blocked by the tuned circuit and not passed on. Radio waves are non-ionizing radiation, which means they do not have enough energy to separate electrons from atoms or molecules, ionizing them, or break chemical bonds, causing chemical reactions or DNA damage. The main effect of absorption of radio waves by materials is to heat them, similarly to the infrared waves radiated by sources of heat such as a space heater or wood fire. The oscillating electric field of the wave causes polar molecules to vibrate back and forth, increasing the temperature; this is how a microwave oven cooks food. However, unlike infrared waves, which are mainly absorbed at the surface of objects and cause surface heating, radio waves are able to penetrate the surface and deposit their energy inside materials and biological tissues. The depth to which radio waves penetrate decreases with their frequency, and also depends on the material's resistivity and permittivity; it is given by a parameter called the skin depth of the material, which is the depth within which 63% of the energy is deposited. For example, the 2.45 GHz radio waves (microwaves) in a microwave oven penetrate most foods approximately 2.5 to 3.8 cm (1 to 1.5 inches). Radio waves have been applied to the body for 100 years in the medical therapy of diathermy for deep heating of body tissue, to promote increased blood flow and healing. More recently they have been used to create higher temperatures in hyperthermia treatment and to kill cancer cells. Looking into a source of radio waves at close range, such as the waveguide of a working radio transmitter, can cause damage to the lens of the eye by heating. A strong enough beam of radio waves can penetrate the eye and heat the lens enough to cause cataracts. Since the heating effect is in principle no different from other sources of heat, most research into possible health hazards of exposure to radio waves has focused on "nonthermal" effects; whether radio waves have any effect on tissues besides that caused by heating. Radiofrequency electromagnetic fields have been classified by the International Agency for Research on Cancer (IARC) as having "limited evidence" for its effects on humans and animals. There is weak mechanistic evidence of cancer risk via personal exposure to RF-EMF from mobile telephones. Radio waves can be shielded against by a conductive metal sheet or screen, an enclosure of sheet or screen is called a Faraday cage. A metal screen shields against radio waves as well as a solid sheet as long as the holes in the screen are smaller than about 1⁄20 of wavelength of the waves. Since radio frequency radiation has both an electric and a magnetic component, it is often convenient to express intensity of radiation field in terms of units specific to each component. The unit volts per meter (V/m) is used for the electric component, and the unit amperes per meter (A/m) is used for the magnetic component. One can speak of an electromagnetic field, and these units are used to provide information about the levels of electric and magnetic field strength at a measurement location. Another commonly used unit for characterizing an RF electromagnetic field is power density. Power density is most accurately used when the point of measurement is far enough away from the RF emitter to be located in what is referred to as the far field zone of the radiation pattern. In closer proximity to the transmitter, i.e., in the "near field" zone, the physical relationships between the electric and magnetic components of the field can be complex, and it is best to use the field strength units discussed above. Power density is measured in terms of power per unit area, for example, milliwatts per square centimeter (mW/cm). When speaking of frequencies in the microwave range and higher, power density is usually used to express intensity since exposures that might occur would likely be in the far field zone.
[[File:Ironbridge 6.jpg|thumb|[[The Iron Bridge]], 1781, was the first [[cast iron]] bridge]] [[File:Pont d'arcades de Móra d'Ebre (Ribera d'Ebre, Catalonia).jpg|thumb|Bridge crossing river d'Ebre, [[Catalonia]].]] [[File:Akashi-kaikyo bridge3.jpg|thumb|The longest suspension bridge in the world connects Honshu [[Shikoku|to Shikoku]], [[Japan]].]] {{otheruses}} A '''bridge''' is a [[structure]] to cross an open space or gap. Bridges are mostly made for crossing [[river]]s, [[valley]]s, or [[road]]s. Nowadays most big bridges are made to carry [[vehicle]]s but people have also walked across bridges for thousands of years. Bridges called [[highway]] overpasses carry a road over another road. [[Military]] bridges are [[portable]], so that they may be easily moved to where they are needed. This makes them much more [[complex]] than most [[civilian]] bridges. The first man-made bridges were probably made from cut wood or stones. Some stone bridges have lasted thousands of years. In recent centuries large bridges are made mostly of [[steel]]. They do not last as long. Many bridges are in a state of disrepair.<ref>https://usatoday30.usatoday.com/news/nation/2010-11-22-bridges22_ST_N.htm</ref> == Types of bridges == There are five major structural types of bridges: * [[Beam bridge]]s<ref name=beambridge>{{cite web | url=http://www.design-technology.org/beambridges.htm | title=Beam bridges | publisher=Design Technology | accessdate=2008-05-14}}</ref> * [[cantilever bridge]]s, * [[arch bridge]]s<ref name=dubai>{{cite web | url=http://www.arabianbusiness.com/509621-worlds-longest-arch-bridge-for-dubai?ln=en | publisher=Arabian Business | accessdate=2008-05-14 | title=Dubai to build world's longest arch bridge | first=Amy | last=Glass}}</ref> * [[suspension bridge]]s and * [[cable-stayed bridge]]s<ref name=cable>{{cite web | url=http://www.newton.dep.anl.gov/askasci/eng99/eng99373.htm | title=Cable Stay vs Suspension Bridges | first=Andy | last=Johnson | publisher=U.S. Department of Energy}}</ref> == Notes == {{commonscat-inline|Bridges}} {{reflist}} {{Civil engineering}} [[Category:Basic English 850 words]] [[Category:Bridges| ]]
A bridge is a structure built to span a physical obstacle (such as a body of water, valley, road, or railway) without blocking the way underneath. It is constructed for the purpose of providing passage over the obstacle, which is usually something that is otherwise difficult or impossible to cross. There are many different designs of bridges, each serving a particular purpose and applicable to different situations. Designs of bridges vary depending on factors such as the function of the bridge, the nature of the terrain where the bridge is constructed and anchored, the material used to make it, and the funds available to build it. The earliest bridges were likely made with fallen trees and stepping stones. The Neolithic people built boardwalk bridges across marshland. The Arkadiko Bridge (dating from the 13th century BC, in the Peloponnese) is one of the oldest arch bridges still in existence and use. The Oxford English Dictionary traces the origin of the word bridge to an Old English word brycg, of the same meaning. The Oxford English Dictionary also notes that there is some suggestion that the word can be traced directly back to Proto-Indo-European *bʰrēw-. However, they also note that "this poses semantic problems." The origin of the word for the card game of the same name is unknown. The simplest and earliest types of bridges were stepping stones. Neolithic people also built a form of boardwalk across marshes; examples of such bridges include the Sweet Track and the Post Track in England, approximately 6000 years old. Undoubtedly, ancient people would also have used log bridges; that is a timber bridge that fall naturally or are intentionally felled or placed across streams. Some of the first human-made bridges with significant span were probably intentionally felled trees. Among the oldest timber bridges is the Holzbrücke Rapperswil-Hurden bridge that crossed upper Lake Zürich in Switzerland; prehistoric timber pilings discovered to the west of the Seedamm causeway date back to 1523 BC. The first wooden footbridge there led across Lake Zürich; it was reconstructed several times through the late 2nd century AD, when the Roman Empire built a 6-metre-wide (20 ft) wooden bridge to carry transport across the lake. Between 1358 and 1360, Rudolf IV, Duke of Austria, built a 'new' wooden bridge across the lake that was used until 1878; it was approximately 1,450 metres (4,760 ft) long and 4 metres (13 ft) wide. On April 6, 2001, a reconstruction of the original wooden footbridge was opened; it is also the longest wooden bridge in Switzerland. The Arkadiko Bridge is one of four Mycenaean corbel arch bridges part of a former network of roads, designed to accommodate chariots, between the fort of Tiryns and town of Epidauros in the Peloponnese, in southern Greece. Dating to the Greek Bronze Age (13th century BC), it is one of the oldest arch bridges still in existence and use. Several intact arched stone bridges from the Hellenistic era can be found in the Peloponnese. The greatest bridge builders of antiquity were the ancient Romans. The Romans built arch bridges and aqueducts that could stand in conditions that would damage or destroy earlier designs. Some stand today. An example is the Alcántara Bridge, built over the river Tagus, in Spain. The Romans also used cement, which reduced the variation of strength found in natural stone. One type of cement, called pozzolana, consisted of water, lime, sand, and volcanic rock. Brick and mortar bridges were built after the Roman era, as the technology for cement was lost (then later rediscovered). In India, the Arthashastra treatise by Kautilya mentions the construction of dams and bridges. A Mauryan bridge near Girnar was surveyed by James Princep. The bridge was swept away during a flood, and later repaired by Puspagupta, the chief architect of emperor Chandragupta I. The use of stronger bridges using plaited bamboo and iron chain was visible in India by about the 4th century. A number of bridges, both for military and commercial purposes, were constructed by the Mughal administration in India. Although large Chinese bridges of wooden construction existed at the time of the Warring States period, the oldest surviving stone bridge in China is the Zhaozhou Bridge, built from 595 to 605 AD during the Sui dynasty. This bridge is also historically significant as it is the world's oldest open-spandrel stone segmental arch bridge. European segmental arch bridges date back to at least the Alconétar Bridge (approximately 2nd century AD), while the enormous Roman era Trajan's Bridge (105 AD) featured open-spandrel segmental arches in wooden construction. Rope bridges, a simple type of suspension bridge, were used by the Inca civilization in the Andes mountains of South America, just prior to European colonization in the 16th century. The Ashanti built bridges over streams and rivers. They were constructed by pounding four large forked tree trunks into the stream bed, placing beams along these forked pillars, then positioning cross-beams that were finally covered with four to six inches of dirt. During the 18th century, there were many innovations in the design of timber bridges by Hans Ulrich Grubenmann, Johannes Grubenmann, as well as others. The first book on bridge engineering was written by Hubert Gautier in 1716. A major breakthrough in bridge technology came with the erection of the Iron Bridge in Shropshire, England in 1779. It used cast iron for the first time as arches to cross the river Severn. With the Industrial Revolution in the 19th century, truss systems of wrought iron were developed for larger bridges, but iron does not have the tensile strength to support large loads. With the advent of steel, which has a high tensile strength, much larger bridges were built, many using the ideas of Gustave Eiffel. In Canada and the United States, numerous timber covered bridges were built in the late 1700s to the late 1800s, reminiscent of earlier designs in Germany and Switzerland. Some covered bridges were also built in Asia. In later years, some were partly made of stone or metal but the trusses were usually still made of wood; in the United States, there were three styles of trusses, the Queen Post, the Burr Arch and the Town Lattice. Hundreds of these structures still stand in North America. They were brought to the attention of the general public in the 1990s by the novel, movie and play The Bridges of Madison County. In 1927 welding pioneer Stefan Bryła designed the first welded road bridge in the world, the Maurzyce Bridge which was later built across the river Słudwia at Maurzyce near Łowicz, Poland in 1929. In 1995, the American Welding Society presented the Historic Welded Structure Award for the bridge to Poland. Bridges can be categorized in several different ways. Common categories include the type of structural elements used, by what they carry, whether they are fixed or movable, and by the materials used. Bridges may be classified by how the actions of tension, compression, bending, torsion and shear are distributed through their structure. Most bridges will employ all of these to some degree, but only a few will predominate. The separation of forces and moments may be quite clear. In a suspension or cable-stayed bridge, the elements in tension are distinct in shape and placement. In other cases the forces may be distributed among a large number of members, as in a truss. Some Engineers sub-divide 'beam' bridges into slab, beam-and-slab and box girder on the basis of their cross-section. A slab can be solid or voided (though this is no longer favored for inspectability reasons) while beam-and-slab consists of concrete or steel girders connected by a concrete slab. A box-girder cross-section consists of a single-cell or multi-cellular box. In recent years, integral bridge construction has also become popular. Most bridges are fixed bridges, meaning they have no moving parts and stay in one place until they fail or are demolished. Temporary bridges, such as Bailey bridges, are designed to be assembled, taken apart, transported to a different site, and re-used. They are important in military engineering and are also used to carry traffic while an old bridge is being rebuilt. Movable bridges are designed to move out of the way of boats or other kinds of traffic, which would otherwise be too tall to fit. These are generally electrically powered. The Tank bridge transporter (TBT) has the same cross-country performance as a tank even when fully loaded. It can deploy, drop off and load bridges independently, but it cannot recover them. Double-decked (or double-decker) bridges have two levels, such as the George Washington Bridge, connecting New York City to Bergen County, New Jersey, US, as the world's busiest bridge, carrying 102 million vehicles annually; truss work between the roadway levels provided stiffness to the roadways and reduced movement of the upper level when the lower level was installed three decades after the upper level. The Tsing Ma Bridge and Kap Shui Mun Bridge in Hong Kong have six lanes on their upper decks, and on their lower decks there are two lanes and a pair of tracks for MTR metro trains. Some double-decked bridges only use one level for street traffic; the Washington Avenue Bridge in Minneapolis reserves its lower level for automobile and light rail traffic and its upper level for pedestrian and bicycle traffic (predominantly students at the University of Minnesota). Likewise, in Toronto, the Prince Edward Viaduct has five lanes of motor traffic, bicycle lanes, and sidewalks on its upper deck; and a pair of tracks for the Bloor–Danforth subway line on its lower deck. The western span of the San Francisco–Oakland Bay Bridge also has two levels. Robert Stephenson's High Level Bridge across the River Tyne in Newcastle upon Tyne, completed in 1849, is an early example of a double-decked bridge. The upper level carries a railway, and the lower level is used for road traffic. Other examples include Britannia Bridge over the Menai Strait and Craigavon Bridge in Derry, Northern Ireland. The Oresund Bridge between Copenhagen and Malmö consists of a four-lane highway on the upper level and a pair of railway tracks at the lower level. Tower Bridge in London is different example of a double-decked bridge, with the central section consisting of a low-level bascule span and a high-level footbridge. A viaduct is made up of multiple bridges connected into one longer structure. The longest and some of the highest bridges are viaducts, such as the Lake Pontchartrain Causeway and Millau Viaduct. A multi-way bridge has three or more separate spans which meet near the center of the bridge. Multi-way bridges with only three spans appear as a "T" or "Y" when viewed from above. Multi-way bridges are extremely rare. The Tridge, Margaret Bridge, and Zanesville Y-Bridge are examples. A bridge can be categorized by what it is designed to carry, such as trains, pedestrian or road traffic (road bridge), a pipeline (Pipe bridge) or waterway for water transport or barge traffic. An aqueduct is a bridge that carries water, resembling a viaduct, which is a bridge that connects points of equal height. A road-rail bridge carries both road and rail traffic. Overway is a term for a bridge that separates incompatible intersecting traffic, especially road and rail. A bridge can carry overhead power lines as does the Storstrøm Bridge. Some bridges accommodate other purposes, such as the tower of Nový Most Bridge in Bratislava, which features a restaurant, or a bridge-restaurant which is a bridge built to serve as a restaurant. Other suspension bridge towers carry transmission antennas. Conservationists use wildlife overpasses to reduce habitat fragmentation and animal-vehicle collisions. The first animal bridges sprung up in France in the 1950s, and these types of bridges are now used worldwide to protect both large and small wildlife. Bridges are subject to unplanned uses as well. The areas underneath some bridges have become makeshift shelters and homes to homeless people, and the undertimbers of bridges all around the world are spots of prevalent graffiti. Some bridges attract people attempting suicide, and become known as suicide bridges. The materials used to build the structure are also used to categorize bridges. Until the end of the 18th century, bridges were made out of timber, stone and masonry. Modern bridges are currently built in concrete, steel, fiber reinforced polymers (FRP), stainless steel or combinations of those materials. Living bridges have been constructed of live plants such as Ficus elastica tree roots in India and wisteria vines in Japan. Unlike buildings whose design is led by architects, bridges are usually designed by engineers. This follows from the importance of the engineering requirements; namely spanning the obstacle and having the durability to survive, with minimal maintenance, in an aggressive outdoor environment. Bridges are first analysed; the bending moment and shear force distributions are calculated due to the applied loads. For this, the finite element method is the most popular. The analysis can be one-, two-, or three-dimensional. For the majority of bridges, a two-dimensional plate model (often with stiffening beams) is sufficient or an upstand finite element model. On completion of the analysis, the bridge is designed to resist the applied bending moments and shear forces, section sizes are selected with sufficient capacity to resist the stresses. Many bridges are made of prestressed concrete which has good durability properties, either by pre-tensioning of beams prior to installation or post-tensioning on site. In most countries, bridges, like other structures, are designed according to Load and Resistance Factor Design (LRFD) principles. In simple terms, this means that the load is factored up by a factor greater than unity, while the resistance or capacity of the structure is factored down, by a factor less than unity. The effect of the factored load (stress, bending moment) should be less than the factored resistance to that effect. Both of these factors allow for uncertainty and are greater when the uncertainty is greater. Most bridges are utilitarian in appearance, but in some cases, the appearance of the bridge can have great importance. Often, this is the case with a large bridge that serves as an entrance to a city, or crosses over a main harbor entrance. These are sometimes known as signature bridges. Designers of bridges in parks and along parkways often place more importance on aesthetics, as well. Examples include the stone-faced bridges along the Taconic State Parkway in New York. Generally bridges are more aesthetically pleasing if they are simple in shape, the deck is thinner (in proportion to its span), the lines of the structure are continuous, and the shapes of the structural elements reflect the forces acting on them. To create a beautiful image, some bridges are built much taller than necessary. This type, often found in east-Asian style gardens, is called a Moon bridge, evoking a rising full moon. Other garden bridges may cross only a dry bed of stream-washed pebbles, intended only to convey an impression of a stream. Often in palaces, a bridge will be built over an artificial waterway as symbolic of a passage to an important place or state of mind. A set of five bridges cross a sinuous waterway in an important courtyard of the Forbidden City in Beijing, China. The central bridge was reserved exclusively for the use of the Emperor and Empress, with their attendants. The estimated life of bridges varies between 25 and 80 years depending on location and material. However, bridges may age hundred years with proper maintenance and rehabilitation. Bridge maintenance consisting of a combination of structural health monitoring and testing. This is regulated in country-specific engineer standards and includes an ongoing monitoring every three to six months, a simple test or inspection every two to three years and a major inspection every six to ten years. In Europe, the cost of maintenance is considerable and is higher in some countries than spending on new bridges. The lifetime of welded steel bridges can be significantly extended by aftertreatment of the weld transitions. This results in a potential high benefit, using existing bridges far beyond the planned lifetime. While the response of a bridge to the applied loading is well understood, the applied traffic loading itself is still the subject of research. This is a statistical problem as loading is highly variable, particularly for road bridges. Load Effects in bridges (stresses, bending moments) are designed for using the principles of Load and Resistance Factor Design. Before factoring to allow for uncertainty, the load effect is generally considered to be the maximum characteristic value in a specified return period. Notably, in Europe, it is the maximum value expected in 1000 years. Bridge standards generally include a load model, deemed to represent the characteristic maximum load to be expected in the return period. In the past, these load models were agreed by standard drafting committees of experts but today, this situation is changing. It is now possible to measure the components of bridge traffic load, to weigh trucks, using weigh-in-motion (WIM) technologies. With extensive WIM databases, it is possible to calculate the maximum expected load effect in the specified return period. This is an active area of research, addressing issues of opposing direction lanes, side-by-side (same direction) lanes, traffic growth, permit/non-permit vehicles and long-span bridges (see below). Rather than repeat this complex process every time a bridge is to be designed, standards authorities specify simplified notional load models, notably HL-93, intended to give the same load effects as the characteristic maximum values. The Eurocode is an example of a standard for bridge traffic loading that was developed in this way. Most bridge standards are only applicable for short and medium spans - for example, the Eurocode is only applicable for loaded lengths up to 200 m. Longer spans are dealt with on a case-by-case basis. It is generally accepted that the intensity of load reduces as span increases because the probability of many trucks being closely spaced and extremely heavy reduces as the number of trucks involved increases. It is also generally assumed that short spans are governed by a small number of trucks traveling at high speed, with an allowance for dynamics. Longer spans on the other hand, are governed by congested traffic and no allowance for dynamics is needed. Calculating the loading due to congested traffic remains a challenge as there is a paucity of data on inter-vehicle gaps, both within-lane and inter-lane, in congested conditions. Weigh-in-Motion (WIM) systems provide data on inter-vehicle gaps but only operate well in free flowing traffic conditions. Some authors have used cameras to measure gaps and vehicle lengths in jammed situations and have inferred weights from lengths using WIM data. Others have used microsimulation to generate typical clusters of vehicles on the bridge. Bridges vibrate under load and this contributes, to a greater or lesser extent, to the stresses. Vibration and dynamics are generally more significant for slender structures such as pedestrian bridges and long-span road or rail bridges. One of the most famous examples is the Tacoma Narrows Bridge that collapsed shortly after being constructed due to excessive vibration. More recently, the Millennium Bridge in London vibrated excessively under pedestrian loading and was closed and retrofitted with a system of dampers. For smaller bridges, dynamics is not catastrophic but can contribute an added amplification to the stresses due to static effects. For example, the Eurocode for bridge loading specifies amplifications of between 10% and 70%, depending on the span, the number of traffic lanes and the type of stress (bending moment or shear force). There have been many studies of the dynamic interaction between vehicles and bridges during vehicle crossing events. Fryba did pioneering work on the interaction of a moving load and an Euler-Bernoulli beam. With increased computing power, vehicle-bridge interaction (VBI) models have become ever more sophisticated. The concern is that one of the many natural frequencies associated with the vehicle will resonate with the bridge first natural frequency. The vehicle-related frequencies include body bounce and axle hop but there are also pseudo-frequencies associated with the vehicle's speed of crossing and there are many frequencies associated with the surface profile. Given the wide variety of heavy vehicles on road bridges, a statistical approach has been suggested, with VBI analyses carried out for many statically extreme loading events. The failure of bridges is of special concern for structural engineers in trying to learn lessons vital to bridge design, construction and maintenance. The failure of bridges first assumed national interest in Britain during the Victorian era when many new designs were being built, often using new materials, with some of them failing catastrophically. In the United States, the National Bridge Inventory tracks the structural evaluations of all bridges, including designations such as "structurally deficient" and "functionally obsolete". There are several methods used to monitor the condition of large structures like bridges. Many long-span bridges are now routinely monitored with a range of sensors, including strain transducers, accelerometers, tiltmeters, and GPS. Accelerometers have the advantage that they are inertial, i.e., they do not require a reference point to measure from. This is often a problem for distance or deflection measurement, especially if the bridge is over water. Crowdsourcing bridge conditions by accessing data passively captured by cell phones, which routinely include accelerometers and GPS sensors, has been suggested as an alternative to including sensors during bridge construction and an augment for professional examinations. An option for structural-integrity monitoring is "non-contact monitoring", which uses the Doppler effect (Doppler shift). A laser beam from a Laser Doppler Vibrometer is directed at the point of interest, and the vibration amplitude and frequency are extracted from the Doppler shift of the laser beam frequency due to the motion of the surface. The advantage of this method is that the setup time for the equipment is faster and, unlike an accelerometer, this makes measurements possible on multiple structures in as short a time as possible. Additionally, this method can measure specific points on a bridge that might be difficult to access. However, vibrometers are relatively expensive and have the disadvantage that a reference point is needed to measure from. Snapshots in time of the external condition of a bridge can be recorded using Lidar to aid bridge inspection. This can provide measurement of the bridge geometry (to facilitate the building of a computer model) but the accuracy is generally insufficient to measure bridge deflections under load. While larger modern bridges are routinely monitored electronically, smaller bridges are generally inspected visually by trained inspectors. There is considerable research interest in the challenge of smaller bridges as they are often remote and do not have electrical power on site. Possible solutions are the installation of sensors on a specialist inspection vehicle and the use of its measurements as it drives over the bridge to infer information about the bridge condition. These vehicles can be equipped with accelerometers, gyrometers, Laser Doppler Vibrometers and some even have the capability to apply a resonant force to the road surface in order to dynamically excite the bridge at its resonant frequency.
{{about|the province of Canada}} {{Infobox province or territory of Canada | name = Ontario | settlement_type = [[Provinces and territories of Canada|Province]] | other_name = | image_flag = Flag of Ontario.svg | flag_alt = A red flag with a large Union Jack in the upper left corner and a shield in the centre-right | image_shield = Coat of arms of Ontario.svg | shield_alt = A central shield with the upper part showing the red cross of St. George and the lower part showing three golden maple leaves on a green background. There is a black bear on top of a knight's helmet above the shield with a moose to the left and a Canadian deer to the right. The province's motto "Ut incepit Fidelis sic permanet", Latin for "Loyal she began, loyal she remains" is written below the crest. | motto = {{lang-la|Ut Incepit Fidelis Sic Permanet}}<br />{{small|("Loyal she began, loyal she remains")}} | Label_map = yes | image_map = Ontario in Canada 2.svg | map_alt = Map showing Ontario's location east/central of Canada. | coordinates = {{Coord|49|15|00|N|84|30|00|W|type:adm1st_scale:30000000_region:CA-ON_source:http://www4.rncan.gc.ca/search-place-names/unique/FEHRI|display=inline,title}} | largest_city = [[Toronto]] | largest_metro = [[Greater Toronto Area]] | official_lang = [[English language|English]] (''de facto'')<ref>{{Cite web|url=https://www.ontario.ca/page/about-ontario|title=About Ontario|date=2019-03-07|website=Ontario.ca|publisher=Queen's Printer for Ontario|language=en|url-status=live|archive-url=https://web.archive.org/web/20200108211436/https://www.ontario.ca/page/about-ontario|archive-date=2020-01-08|access-date=2020-01-08}}</ref> | government_type = [[Constitutional monarchy]] | governing_body = [[Government of Ontario]] | ViceroyType = Lieutenant Governor | Viceroy = [[Edith Dumont]] | Premier = [[Doug Ford]] | PremierParty = [[Progressive Conservative Party of Ontario|PC]] | Legislature = Legislative Assembly of Ontario | HouseSeats = 121 | SenateSeats = 24 | AdmittanceDate = July 1, 1867 | AdmittanceOrder = 1st, with [[Quebec]], [[Nova Scotia]], [[New Brunswick]] | area_footnotes = <ref name="areaont">{{cite web|url=http://www.statcan.gc.ca/tables-tableaux/sum-som/l01/cst01/phys01-eng.htm|title=Land and freshwater area, by province and territory.|publisher=Statistics Canada|date=February 1, 2005|access-date=August 5, 2012|url-status=live|archive-url=https://web.archive.org/web/20121019171348/http://www.statcan.gc.ca/tables-tableaux/sum-som/l01/cst01/phys01-eng.htm|archive-date=October 19, 2012}}</ref> | area_rank = 4th | area_total_km2 = 1076395 | area_land_km2 = 917741 | area_water_km2 = 158654 | population_demonym = Ontarian<ref>{{Cite web|url=http://www.collinsdictionary.com/dictionary/english/ontarian|title=Definition of Ontarian|website=Collins Online Dictionary|publisher=HarperCollins Publishers|language=en|url-status=live|archive-url=https://web.archive.org/web/20131004234108/http://www.collinsdictionary.com/dictionary/english/ontarian|archive-date=2013-10-04|access-date=2013-10-03}}</ref> | population_rank = 1st | population_as_of = [[Canada 2016 Census|2016]] | population_total = 13448494<!--2016 StatCan federal census population only per [[WP:CANPOP]]. Do not update until 2021 census population released. Use "Population_est" below for latest StatCan quarterly estimate.--> | population_ref = <ref name=StatCan2011>{{cite web |url=http://www12.statcan.gc.ca/census-recensement/2016/dp-pd/hlt-fst/pd-pl/Table.cfm?Lang=Eng&T=101&SR=1&S=3&O=D#tPopDwell |title=Population and dwelling counts, for Canada, provinces and territories, 2016 and 2011 censuses |publisher=[[Statistics Canada]] |date=February 6, 2017 |access-date=February 8, 2017 |url-status=live |archive-url=https://web.archive.org/web/20170211082232/http://www12.statcan.gc.ca/census-recensement/2016/dp-pd/hlt-fst/pd-pl/Table.cfm?Lang=Eng&T=101&SR=1&S=3&O=D#tPopDwell |archive-date=February 11, 2017 }}</ref> | population_est = 14,733,119<!-- Latest StatCan quarterly estimate only. --> | pop_est_as_of = 2020 Q4 | pop_est_ref =<ref name="StatCan2019Q1Est">{{Cite web |url=http://www5.statcan.gc.ca/cansim/a26?lang=eng&retrLang=eng&id=0510005&paSer=&pattern=&stByVal=1&p1=1&p2=31&tabMode=dataTable&csid= |title=Population estimates, quarterly |date=June 14, 2018 |publisher=[[Statistics Canada]] |access-date=April 12, 2020}}</ref> | GDP_rank = 1st | GDP_year = 2015 | GDP_total = CA$763.276&nbsp;billion<ref name=GDP2013>{{cite web |url=http://www.statcan.gc.ca/tables-tableaux/sum-som/l01/cst01/econ15-eng.htm |title=Gross domestic product, expenditure-based, by province and territory (2015) |publisher=Statistics Canada |date=November 9, 2016 |access-date=January 26, 2017 |url-status=live |archive-url=https://web.archive.org/web/20120919211233/http://www.statcan.gc.ca/tables-tableaux/sum-som/l01/cst01/econ15-eng.htm |archive-date=September 19, 2012}}</ref> | GDP_per_capita_rank = 7th | GDP_per_capita = CA$59,879 | HDI_year = 2018 | HDI = 0.929<ref name=":0">{{Cite web|title=Sub-national HDI - Subnational HDI - Global Data Lab|url=https://globaldatalab.org/shdi|access-date=2020-06-18|website=globaldatalab.org}}</ref> — <span style="color:#090">Very high</span> | HDI_rank = 3rd | timezone_link = Time in Canada | timezone1_location = East of [[90th meridian west]] | timezone1 = [[Eastern Standard Time|EST]] | utc_offset1 = -05:00 | timezone1_DST = [[Eastern Daylight Time|EDT]] | utc_offset1_DST = -04:00 | timezone2_location = West of 90th meridian west, except [[Atikokan]] and [[Pickle Lake]] | timezone2 = [[Central Standard Time|CST]] | utc_offset2 = -06:00 | timezone2_DST = [[Central Daylight Time|CDT]] | utc_offset2_DST = -05:00 | timezone3_location = [[Atikokan]] and [[Pickle Lake]] (No DST) | timezone3 = EST | utc_offset3 = -05:00 | timezone3_DST = | utc_offset3_DST = | PostalAbbreviation = ON | PostalCodePrefix = [[List of K postal codes of Canada|K]] [[List of L postal codes of Canada|L]] [[List of M postal codes of Canada|M]] [[List of N postal codes of Canada|N]] [[List of P postal codes of Canada|P]] | iso_code = CA-ON | website = www.ontario.ca | flower = [[Trillium grandiflorum|White trillium]] | tree = [[Pinus strobus|Eastern white pine]] | bird = [[Common loon]] }} '''Ontario''' is a [[province]] of [[Canada]]. It is in the eastern half of Canada, between [[Manitoba]] and [[Quebec]]. Ontario has [[population|the most people]] of any province, with 13,150,000 in 2009, and is home to the biggest city in Canada, [[Toronto]], which is also the [[capital city|capital]] of the province. Ontario also has the second largest [[size|land area]], with 1,076,395&nbsp;km²; only [[Quebec]] is larger by size. ([[Nunavut]] and [[Northwest Territories]] are also larger, but are called ''territories'' and not ''provinces''). The province has one of the longest borders with the [[United States]] and there are several [[border crossing]]s including the one at [[Niagara Falls]]. Along this border are 4 large lakes called [[Lake Ontario]], [[Lake Erie]], [[Lake Huron]], and [[Lake Superior]]. They each are partly in Ontario and partly in the United States, and the border runs through them, but not [[Lake Michigan]], which is entirely in the United States. These five lakes are together are called the [[Great Lakes]]. There are a number of [[symbol]]s that represent the province of Ontario. The [[flag]] is red with the British [[Union Jack]] in the top left corner and the provincial [[shield]] is on the right hand side of the flag. The provincial [[bird]] is the [[loon]], and the provincial [[flower]] is the [[trillium]]. It has three flower [[petal]]s and it is usually white but some times is pink or purple. ==History== Ontario became part of Canada when it was created in 1867. Before 1840, it was known as Upper Canada, which was a [[colony]] in the [[British Empire]]. Between 1840 and 1867 it was known as Canada West. The [[government]] of Ontario sits at [[Queen's Park]] in Toronto. The leader of the government is called the ''[[Premier]]'', and [[Premier of Ontario|the current Premier]] is [[Doug Ford]]. There is also a [[Lieutenant Governor]] who represents the [[Elizabeth II|Queen]], and [[Lieutenant Governor of Ontario|the current Lieutenant Governor]] is Edith Dumont. ==Geography== Ontario is very large, so sometimes people break it into two. The two parts are called [[Northern Ontario]] and [[Southern Ontario]]. Most of the people in Ontario live in the south, and that is where the big cities are. The big cities in [[Southern Ontario]] are Toronto and the rest of the [[Greater Toronto Area]], Ottawa and the [[National Capital Region (Canada)|National Capital Region]], [[Hamilton, Ontario|Hamilton]], [[London, Ontario|London]], [[Windsor, Ontario|Windsor]], and [[Sarnia]]. The cities in the north are smaller. In the far north of Ontario hardly any people live at all, and there are no [[road]]s or [[railway]]s making it difficult to even get to those places. Much of Ontario gets lots of snow in the winter. In the summer, it can get very hot in the south parts. In some big cities, there is [[smog]] in the summer. ==References== <references /> == Other websites == {{sisterlinks|Ontario}} {{wikiatlas|Ontario}} *[http://www.gouv.qc.ca/portail/Ontario/pgs?lang=en Government of Ontario] {{Webarchive|url=https://web.archive.org/web/20160304190529/http://www.gouv.qc.ca/portail/Ontario/pgs?lang=en |date=2016-03-04 }} * {{dmoz|Regional/North_America/Canada/Ontario}} {{Provinces and territories of Canada}} [[Category:Ontario| ]] [[Category:1867 establishments in Canada]] {{Canada-stub}} {{authority control}}
Ontario (/ɒnˈtɛərioʊ/ on-TAIR-ee-oh; French: [ɔ̃taʁjo]) is one of the thirteen provinces and territories of Canada. Located in Central Canada, Ontario is the country's most populous province. As of the 2021 Canadian census, it is home to 38.5 percent of the country's population, and is the second-largest province by total area (after Quebec). Ontario is Canada's fourth-largest jurisdiction in total area of all the Canadian provinces and territories. It is home to the nation's capital, Ottawa, and its most populous city, Toronto, which is Ontario's provincial capital. Ontario is bordered by the province of Manitoba to the west, Hudson Bay and James Bay to the north, and Quebec to the east and northeast. To the south, it is bordered by the U.S. states of (from west to east) Minnesota, Michigan, Ohio, Pennsylvania, and New York. Almost all of Ontario's 2,700 km (1,700 mi) border with the United States follows inland waterways: from the westerly Lake of the Woods, eastward along the major rivers and lakes of the Great Lakes/Saint Lawrence River drainage system. There is only about 1 km (5⁄8 mi) of actual land border, made up of portages including Height of Land Portage on the Minnesota border. The great majority of Ontario's population and arable land is in Southern Ontario, and while agriculture remains a significant industry, the region's economy depends highly on manufacturing. In contrast, Northern Ontario is sparsely populated with cold winters and heavy forestation, with mining and forestry making up the region's major industries. Ontario is a term thought to be derived from Indigenous origins, either Ontarí:io, a Huron (Wyandot) word meaning "great lake", or possibly skanadario, which means "beautiful water" or "sparkling water" in the Iroquoian languages. Ontario has about 250,000 freshwater lakes. The first mention of the name Ontario was in 1641, when "Ontario" was used to describe the land on the north shore of the easternmost part of the Great Lakes. It was adopted as the official name of the new province at Confederation in 1867. The thinly populated Canadian Shield, which dominates the northwestern and central portions of the province, comprises over half the land area of Ontario. Although this area mostly does not support agriculture, it is rich in minerals, partly covered by the Central and Midwestern Canadian Shield forests, and studded with lakes and rivers. Northern Ontario is subdivided into two sub-regions: Northwestern Ontario and Northeastern Ontario. The virtually unpopulated Hudson Bay Lowlands in the extreme north and northeast are mainly swampy and sparsely forested. Southern Ontario, which is further sub-divided into four sub-regions: Central Ontario (although not actually the province's geographic centre), Eastern Ontario, Golden Horseshoe and Southwestern Ontario (parts of which were formerly referred to as Western Ontario). Despite the absence of any mountainous terrain in the province, there are large areas of uplands, particularly within the Canadian Shield which traverses the province from northwest to southeast and also above the Niagara Escarpment which crosses the south. The highest point is Ishpatina Ridge at 693 metres (2,274 ft) above sea level in Temagami, Northeastern Ontario. In the south, elevations of over 500 m (1,640 ft) are surpassed near Collingwood, above the Blue Mountains in the Dundalk Highlands and in hilltops near the Madawaska River in Renfrew County. The Carolinian forest zone covers most of the southwestern region of the province. The temperate and fertile Great Lakes-Saint Lawrence Valley in the south is part of the Eastern Great Lakes lowland forests ecoregion where the forest has now been largely replaced by agriculture, industrial and urban development. A well-known geographic feature is Niagara Falls, part of the Niagara Escarpment. The Saint Lawrence Seaway allows navigation to and from the Atlantic Ocean as far inland as Thunder Bay in Northwestern Ontario. Northern Ontario covers approximately 87% of the province's surface area; conversely, Southern Ontario contains 94% of the population. Point Pelee is a peninsula of Lake Erie in southwestern Ontario (near Windsor and Detroit, Michigan) that is the southernmost extent of Canada's mainland. Pelee Island and Middle Island in Lake Erie extend slightly farther. All are south of 42°N – slightly farther south than the northern border of California. Ontario's climate varies by season and location. Three air sources affect it: cold, dry, arctic air from the north (dominant factor during the winter months, and for a longer part of the year in far northern Ontario); Pacific polar air crossing in from the western Canadian Prairies/US Northern Plains; and warm, moist air from the Gulf of Mexico and the Atlantic Ocean. The effects of these major air masses on temperature and precipitation depend mainly on latitude, proximity to major bodies of water and to a small extent, terrain relief. In general, most of Ontario's climate is classified as humid continental. Ontario has three main climatic regions: In the northeastern parts of Ontario, extending south as far as Kirkland Lake, the cold waters of Hudson Bay depress summer temperatures, making it cooler than other locations at similar latitudes. The same is true on the northern shore of Lake Superior, which cools hot, humid air from the south, leading to cooler summer temperatures. Along the eastern shores of Lake Superior and Lake Huron winter temperatures are slightly moderated but come with frequent heavy lake-effect snow squalls that increase seasonal snowfall totals to upwards of 3 m (10 ft) in some places. These regions have higher annual precipitation, in some places over 100 cm (39 in). Severe thunderstorms peak in summer. Windsor, in Southern (Southwestern) Ontario, has the most lightning strikes per year in Canada, averaging 33 days of thunderstorm activity per year. In a typical year, Ontario averages 11 confirmed tornado touchdowns. However, over the last 4 years, it has had upwards of 20 tornado touchdowns per year, with the highest frequency in the Windsor-Essex – Chatham Kent area, though few are very destructive (the majority between F0 to F2 on the Fujita scale). Ontario had a record 29 tornadoes in both 2006 and 2009. Tropical depression remnants occasionally bring heavy rains and winds in the south, but are rarely deadly. A notable exception was Hurricane Hazel which struck Southern Ontario centred on Toronto, in October 1954. Paleo-Indians were the first people to settle on the lands of Ontario, arriving there after the Laurentide Ice Sheet melted roughly 11,000 years ago. From them, many ethnocultural groups emerged and came to exist on the lands of Ontario: the Algonquins, Mississaugas, Ojibway, Cree, Odawa, Pottowatomi, and Iroquois. In the 15th century, the Byzantine Empire fell, prompting Western Europeans to search for new sea routes to the Far East. Around 1522–1523, Giovanni da Verrazzano persuaded King Francis I of France to commission an expedition to find a western route to Cathay (China) via a Northwest Passage. Though this expedition was unsuccessful, it established the name "New France" for northeastern North America. After a few expeditions, France mostly abandoned North America for 50 years because of its financial crisis; France was involved in the Italian Wars and there were religious wars between Protestants and Catholics. Around 1580 however, the rise of the fur trade (particularly the demand for beaver pelts), reignited French interest. In 1608, Samuel de Champlain established France's first colonial settlement in New France, the Habitation de Québec (now Quebec City), in the colony of Canada (now southern Quebec). Afterwards, French explorers continued to travel west, establishing new villages along the coasts of the Saint Lawrence River. French explorers, the first of which was Étienne Brûlé who explored the Georgian Bay area in 1610–1612, mapped Southern Ontario and called the region the Pays d'en Haut ("Upper Country"), in reference to the region being upstream of the Saint Lawrence River. The colony of the Pays d'en Haut was formally established in 1610 as an administrative dependency of Canada, and was for defence and business rather than a settlement colony. The territory of the Pays-d'en-Haut was quite large and would today include the province of Ontario, as well as, in whole or in part, the American states of Minnesota, Wisconsin, Illinois, Indiana, Michigan, Ohio, Pennsylvania and New York. Indigenous peoples were the vast majority of the Pays d'en Haut population. As for Northern Ontario, the English explorer Henry Hudson sailed into Hudson Bay in 1611 and claimed its drainage basin for England. The area would become known as Rupert's Land. Samuel de Champlain reached Lake Huron in 1615, and French missionaries, such as the Jésuites and Supliciens, began to establish posts along the Great Lakes. The French allied with most Indigenous groups of Ontario, all for the fur trade and for defence against Iroquois attacks (which would later be called the Iroquois Wars). The French would declare their Indigenous allies to be subjects of the King of France and would often act as mediators between different groups. The Iroquois later allied themselves with the British. From 1634 to 1640, the Huron were devastated by European infectious diseases, such as measles and smallpox, to which they had no immunity. By 1700, the Iroquois had been driven out or left the area that would become Ontario and the Mississaugas of the Ojibwa had settled the north shore of Lake Ontario. The remaining Huron settled north of Quebec. During the French and Indian War, the North American theater of the Seven Years' War of 1754 to 1763, the British defeated the armies of New France and its Indigenous allies. In the Treaty of Paris 1763 France ceded most of its possessions in North America to Britain. Using the Quebec Act, Britain re-organised the territory into the Province of Quebec. In 1782–1784, 5,000 United Empire Loyalists entered what is now Ontario following the American Revolution. The Kingdom of Great Britain granted them 200 acres (81 ha) land and other items with which to rebuild their lives. The British also set up reserves in Ontario for the Mohawks who had fought for the British and had lost their land in New York state. Other Iroquois, also displaced from New York were resettled in 1784 at the Six Nations reserve at the west end of Lake Ontario. The Mississaugas, displaced by European settlements, would later move to Six Nations also. After the American War of Independence, the first reserves for First Nations were established. These are situated at Six Nations (1784), Tyendinaga (1793) and Akwesasne (1795). Six Nations and Tyendinaga were established by the British for those Indigenous groups who had fought on the side of the British, and were expelled from the new United States. Akwesasne was a pre-existing Mohawk community and its borders were formalized under the 1795 Jay Treaty. In 1788, while part of the province of Quebec, southern Ontario was divided into four districts: Hesse, Lunenburg, Mecklenburg, and Nassau. In 1792, the four districts were renamed: Hesse became the Western District, Lunenburg became the Eastern District, Mecklenburg became the Midland District, and Nassau became the Home District. Counties were created within the districts. The population of Canada west of the St. Lawrence-Ottawa River confluence substantially increased during this period, a fact recognized by the Constitutional Act of 1791, which split Quebec into the Canadas: Upper Canada southwest of the St. Lawrence-Ottawa River confluence, and Lower Canada east of it. John Graves Simcoe was appointed Upper Canada's first Lieutenant governor in 1793. A second wave of Americans, not all of them necessarily loyalists moved to Upper Canada after 1790 until the pre-war of 1812, many seeking available cheap land, and at the time, lower taxation. By 1798, there were eight districts: Eastern, Home, Johnstown, London, Midland, Newcastle, Niagara, and Western. By 1826, there were eleven districts: Bathurst, Eastern, Gore, Home, Johnstown, London, Midland, Newcastle, Niagara, Ottawa, and Western. By 1838, there were twenty districts: Bathurst, Brock, Colbourne, Dalhousie, Eastern, Gore, Home, Huron, Johnstown, London, Midland, Newcastle, Niagara, Ottawa, Prince Edward, Simcoe, Talbot, Victoria, Wellington, and Western. American troops in the War of 1812 invaded Upper Canada across the Niagara River and the Detroit River, but were defeated and pushed back by the British, Canadian fencibles and militias, and First Nations warriors. However, the Americans eventually gained control of Lake Erie and Lake Ontario. The 1813 Battle of York saw American troops defeat the garrison at the Upper Canada capital of York. The Americans looted the town and burned the Upper Canada Parliament Buildings during their brief occupation. The British would burn the American capital of Washington, D.C. in 1814. After the War of 1812, relative stability allowed for increasing numbers of immigrants to arrive from Europe rather than from the United States. As was the case in the previous decades, this immigration shift was encouraged by the colonial leaders. Despite affordable and often free land, many arriving newcomers, mostly from Britain and Ireland, found frontier life with the harsh climate difficult, and some of those with the means eventually returned home or went south. However, population growth far exceeded emigration in the following decades. It was a mostly agrarian-based society, but canal projects and a new network of plank roads spurred greater trade within the colony and with the United States, thereby improving previously damaged relations over time. Meanwhile, Ontario's numerous waterways aided travel and transportation into the interior and supplied water power for development. As the population increased, so did the industries and transportation networks, which in turn led to further development. By the end of the century, Ontario vied with Quebec as the nation's leader in terms of growth in population, industry, arts and communications. Unrest in the colony began to chafe against the aristocratic Family Compact who governed while benefiting economically from the region's resources, and who did not allow elected bodies power. This resentment spurred republican ideals and sowed the seeds for early Canadian nationalism. Accordingly, rebellion in favour of responsible government rose in both regions; Louis-Joseph Papineau led the Lower Canada Rebellion and William Lyon Mackenzie, first Toronto mayor, led the Upper Canada Rebellion. In Upper Canada, the rebellion was quickly a failure. William Lyon Mackenzie escaped to the United States, where he declared the Republic of Canada on Navy Island on the Niagara River. Although both rebellions were put down in short order, the British government sent Lord Durham to investigate the causes. He recommended self-government be granted and Lower and Upper Canada be re-joined in an attempt to assimilate the French Canadians. Accordingly, the two colonies were merged into the Province of Canada by the Act of Union 1840, with the capital at Kingston, and Upper Canada becoming known as Canada West. Parliamentary self-government was granted in 1848. There were heavy waves of immigration in the 1840s, and the population of Canada West more than doubled by 1851 over the previous decade. As a result, for the first time, the English-speaking population of Canada West surpassed the French-speaking population of Canada East, tilting the representative balance of power. In 1849, the districts of southern Ontario were abolished by the Province of Canada, and county governments took over certain municipal responsibilities. The Province of Canada also began creating districts in sparsely populated Northern Ontario with the establishment of Algoma District and Nipissing District in 1858. An economic boom in the 1850s coincided with railway expansion across the province, further increasing the economic strength of Central Canada. With the repeal of the Corn Laws and a reciprocity agreement in place with the United States, various industries such as timber, mining, farming and alcohol distilling benefited tremendously. A political stalemate between the French- and English-speaking legislators, as well as fear of aggression from the United States during and immediately after the American Civil War, led the political elite to hold a series of conferences in the 1860s to effect a broader federal union of all British North American colonies. The British North America Act took effect on July 1, 1867, establishing the Dominion of Canada, initially with four provinces: Nova Scotia, New Brunswick, Quebec and Ontario. The Province of Canada was divided into Ontario and Quebec so that each linguistic group would have its own province. Both Quebec and Ontario were required by section 93 of the British North America Act to safeguard existing educational rights and privileges of the Protestant and Catholic minorities. Thus, separate Catholic schools and school boards were permitted in Ontario. However, neither province had a constitutional requirement to protect its French- or English-speaking minority. Toronto was formally established as Ontario's provincial capital. The borders of Ontario, its new name in 1867, were provisionally expanded north and west. When the Province of Canada was formed, its borders were not entirely clear, and Ontario claimed eventually to reach all the way to the Rocky Mountains and Arctic Ocean. With Canada's acquisition of Rupert's Land, Ontario was interested in clearly defining its borders, especially since some of the new areas in which it was interested were rapidly growing. After the federal government asked Ontario to pay for construction in the new disputed area, the province asked for an elaboration on its limits, and its boundary was moved north to the 51st parallel north. Once constituted as a province, Ontario proceeded to assert its economic and legislative power. In 1872, the lawyer Oliver Mowat became Premier of Ontario and remained as premier until 1896. He fought for provincial rights, weakening the power of the federal government in provincial matters, usually through well-argued appeals to the Judicial Committee of the Privy Council. His battles with the federal government greatly decentralized Canada, giving the provinces far more power than John A. Macdonald had intended. He consolidated and expanded Ontario's educational and provincial institutions, created districts in Northern Ontario, and fought to ensure that those parts of Northwestern Ontario not historically part of Upper Canada (the vast areas north and west of the Lake Superior-Hudson Bay watershed, known as the District of Keewatin) would become part of Ontario, a victory embodied in the Canada (Ontario Boundary) Act, 1889. He also presided over the emergence of the province into the economic powerhouse of Canada. Mowat was the creator of what is often called Empire Ontario. Beginning with Macdonald's National Policy (1879) and the construction of the Canadian Pacific Railway (1875–1885) through Northern Ontario and the Canadian Prairies to British Columbia, Ontario manufacturing and industry flourished. However, population increases slowed after a large recession hit the province in 1893, thus slowing growth drastically but for only a few years. Many newly arrived immigrants and others moved west along the railway to the Prairie Provinces and British Columbia, sparsely settling Northern Ontario. The northern and western boundaries of Ontario were in dispute after Canadian Confederation. Ontario's right to Northwestern Ontario was determined by the Judicial Committee of the Privy Council in 1884 and confirmed by the Canada (Ontario Boundary) Act, 1889 of the Parliament of the United Kingdom. By 1899, there were seven northern districts: Algoma, Manitoulin, Muskoka, Nipissing, Parry Sound, Rainy River, and Thunder Bay. Four more northern districts were created between 1907 and 1912: Cochrane, Kenora, Sudbury and Timiskaming. Mineral exploitation accelerated in the late 19th century, leading to the rise of important mining centres in the northeast, such as Sudbury, Cobalt and Timmins. The province harnessed its water power to generate hydro-electric power and created the state-controlled Hydro-Electric Power Commission of Ontario, later Ontario Hydro. The availability of cheap electric power further facilitated the development of industry. The Ford Motor Company of Canada was established in 1904 and the McLaughlin Motor Car Company (later General Motors Canada) was founded in 1907. The motor vehicle industry became the most lucrative industry for the Ontario economy during the 20th century. In July 1912, the Conservative government of James Whitney issued Regulation 17 which severely limited the availability of French-language schooling to the province's French-speaking minority. French Canadians reacted with outrage, journalist Henri Bourassa denouncing the "Prussians of Ontario". The regulation was eventually repealed in 1927. Influenced by events in the United States, the government of William Hearst introduced prohibition of alcoholic drinks in 1916 with the passing of the Ontario Temperance Act. However, residents could distil and retain their own personal supply, and liquor producers could continue distillation and export for sale, allowing this already sizeable industry to strengthen further. Ontario became a hotbed for the illegal smuggling of liquor and the biggest supplier into the United States, which was under complete prohibition. Prohibition in Ontario came to an end in 1927 with the establishment of the Liquor Control Board of Ontario under the government of Howard Ferguson. The sale and consumption of liquor, wine, and beer are still controlled by some of the most extreme laws in North America to ensure strict community standards and revenue generation from the alcohol retail monopoly are upheld. The post-World War II period was one of exceptional prosperity and growth. Ontario has been the recipients of most immigration to Canada, largely immigrants from war-torn Europe in the 1950s and 1960s and following changes in federal immigration law, a massive influx of non-Europeans since the 1970s. From a largely ethnically British province, Ontario has rapidly become culturally very diverse. The nationalist movement in Quebec, particularly after the election of the Parti Québécois in 1976, contributed to driving many businesses and English-speaking people out of Quebec to Ontario, and as a result, Toronto surpassed Montreal as the largest city and economic centre of Canada. Depressed economic conditions in the Maritime Provinces have also resulted in de-population of those provinces in the 20th century, with heavy migration into Ontario. Ontario's official language is English, although there exists a number of French-speaking communities across Ontario. French-language services are made available for communities with a sizeable French-speaking population; a service that is ensured under the French Language Services Act of 1989. In the 2021 census, Ontario had a population of 14,223,942 living in 5,491,201 of its 5,929,250 total dwellings, a 5.8 percent change from its 2016 population of 13,448,494. With a land area of 892,411.76 km (344,562.11 sq mi), it had a population density of 15.9/km (41.3/sq mi) in 2021. The largest population centres in Ontario are Toronto, Ottawa, Hamilton, Kitchener, London and Oshawa, which all have more than 300,000 inhabitants. The percentages given below add to more than 100 per cent because of dual responses (e.g., "French and Canadian" response generates an entry both in the category "French Canadian" and in the category "Canadian"). The majority of Ontarians are of English or other European descent including large Scottish, Irish and Italian communities. Slightly less than 5 per cent of the population of Ontario is Franco-Ontarian, that is those whose native tongue is French, although those with French ancestry account for 11 per cent of the population. Compared to natural increase or interprovincial migration, immigration is a huge population growth force in Ontario, as it has been over the last two centuries. More recent sources of immigrants with large or growing communities in Ontario include East Asians, South Asians, Caribbeans, Latin Americans, Europeans, and Africans. Most populations have settled in the larger urban centres. In 2021, 34.3% of the population consisted of visible minorities and 2.9% of the population was Indigenous, mostly of First Nations and Métis descent. There was also a small number of Inuit in the province. The number of Indigenous people and visible minorities has been increasing at a faster rate than the general population of Ontario. In 2021, 52.1% of the population was Christian, with the largest religious denominations being the Roman Catholic Church (with 26.0% of the population) and the United Church of Canada with (4.1%). Other religions included Islam (6.7%), Hinduism (4.1%). 31.6% of Ontarians had no religious affiliation. The major religious groups in Ontario in 2021 were: In Ontario, Catholics are represented by the Assembly of Catholic Bishops of Ontario and the Anglican Protestants by the Ecclesiastical Province of Ontario. As of the 2021 Canadian Census, the ten most spoken languages in the province included English (13,650,230 or 97.28%), French (1,550,545 or 11.05%), Mandarin (467,420 or 3.33%), Hindi (436,125 or 3.11%), Spanish (401,205 or 2.86%), Punjabi (397,865 or 2.84%), Cantonese (352,135 or 2.51%), Arabic (342,860 or 2.44%), Italian (312,800 or 2.23%), and Urdu (295,175 or 2.1%). The principal language of Ontario is English, the province's de facto official language, with approximately 97.2 per cent of Ontarians having proficiency in the language, although only 69.5 per cent of Ontarians reported English as their mother tongue in the 2016 Census. English is one of two official languages of Canada, with the other being French. English and French are the official languages of the courts in Ontario. Approximately 4.6 per cent of the population identified as francophone, and a total of 11.5 per cent of Ontarians reported having proficiency in French. Approximately 11.2 per cent of Ontarians reported being bilingual in both English and French. Approximately 2.5 per cent of Ontarians have no proficiency in either English or French. Franco-Ontarians are concentrated in the northeastern, eastern, and extreme southern parts of the province, where under the French Language Services Act, provincial government services are required to be available in French if at least 10 per cent of a designated area's population report French as their native language or if an urban centre has at least 5,000 francophones. Other languages spoken by residents include Arabic, Bengali, Cantonese, Dutch, German, Greek, Gujarati, Hindi, Hebrew, Italian, Korean, Malayalam, Mandarin, Marathi, Persian, Polish, Portuguese, Punjabi, Russian, Sinhalese, Somali, Spanish, Tagalog, Telugu, Tamil, Tibetan, Ukrainian, Urdu, and Vietnamese. Ontario is Canada's leading manufacturing province, accounting for 52% of the total national manufacturing shipments in 2004. Ontario's largest trading partner is the American state of Michigan. As of April 2012, Moody's bond-rating agency rated Ontario debt at AA2/stable, while S&P rated it AA-. Dominion Bond Rating Service rated it AA(low) in January 2013. Long known as a bastion of Canadian manufacturing and financial solvency, Ontario's public debt-to-GDP ratio is projected to be 38.4% in fiscal year 2023–2024. Mining and the forest products industry, notably pulp and paper, are vital to the economy of Northern Ontario. As of 2011, roughly 200,000 ha are clearcut each year; herbicides for hardwood suppression are applied to a third of the total. There has been controversy over the Ring of Fire mineral deposit, and whether the province can afford to spend CAD$2.25 billion on a road from the Trans-Canada Highway near Kenora to the deposit, currently valued at CAD$60 billion. An abundance of natural resources, excellent transportation links to the North American heartland and the inland Great Lakes making ocean access possible via container ships, have all contributed to making manufacturing the principal industry of the province, found mainly in the Golden Horseshoe region, which is the largest industrialized area in Canada, the southern end of the region being part of the North American Rust Belt. Important products include motor vehicles, iron, steel, food, electrical appliances, machinery, chemicals, and paper. Hamilton is the largest steel manufacturing city in Canada followed closely by Sault Ste. Marie, and Sarnia is the centre for petrochemical production. Construction employed more than 6.5% of the province's work force in June 2011. Ontario's steel industry was once centred in Hamilton. Hamilton harbour, which can be seen from the QEW Skyway bridge, is an industrial wasteland; U.S. Steel-owned Stelco announced in the autumn of 2013 that it would close in 2014, with the loss of 875 jobs. The move flummoxed a union representative, who seemed puzzled why a plant with capacity of 2 million tons per annum would be shut while Canada imported 8 million tons of steel the previous year. Algoma Steel maintains a plant in Sault Ste Marie. Ontario surpassed Michigan in car production, assembling more than 2,696,000 vehicles in 2004. Ontario has Chrysler plants in Windsor and Bramalea, two GM plants in Oshawa and one in Ingersoll, a Honda assembly plant in Alliston, Ford plants in Oakville and St. Thomas and Toyota assembly plants in Cambridge and Woodstock. However, as a result of steeply declining sales, in 2005, General Motors announced massive layoffs at production facilities across North America, including two large GM plants in Oshawa and a drive train facility in St. Catharines, that resulted in 8,000 job losses in Ontario alone. In 2006, Ford Motor Company announced between 25,000 and 30,000 layoffs phased until 2012; Ontario was spared the worst, but job losses were announced for the St Thomas facility and the Windsor Casting plant. However, these losses will be offset by Ford's recent announcement of a hybrid vehicle facility slated to begin production in 2007 at its Oakville plant and GM's re-introduction of the Camaro which will be produced in Oshawa. On December 4, 2008, Toyota announced the grand opening of the RAV4 plant in Woodstock, and Honda also plans to add an engine plant at its facility in Alliston. Despite these new plants coming online, Ontario has not yet fully recovered following massive layoffs caused by the global recession; its unemployment rate was 7.3% in May 2013, compared to 8.7 percent in January 2010 and approximately 6% in 2007. In September 2013, the Ontario government committed CAD$70.9 million to the Ford plant in Oakville, while the federal government committed CAD$71.1mn, to secure 2,800 jobs. The province has lost 300,000 manufacturing jobs in the decade from 2003, and the Bank of Canada noted that "while the energy and mining industries have benefitted from these movements, the pressure on the manufacturing sector has intensified, since many firms in this sector were already dealing with growing competition from low-cost economies such as China." Toronto, the capital of Ontario, is the centre of Canada's financial services and banking industry. Neighbouring cities are home to product distribution, IT centres, and manufacturing industries. Canada's Federal Government is the largest single employer in the National Capital Region, which centres on the border cities of Ontario's Ottawa and Quebec's Gatineau. The information technology sector is important, particularly in the Silicon Valley North section of Ottawa, home to Canada's largest technology park. IT is also important in the Waterloo Region, where the headquarters of BlackBerry is located. Tourism contributes heavily to the economy of Central Ontario, peaking during the summer months owing to the abundance of fresh water recreation and wilderness found there in reasonable proximity to the major urban centres. At other times of the year, hunting, skiing and snowmobiling are popular. This region has some of the most vibrant fall colour displays anywhere on the continent, and tours directed at overseas visitors are organized to see them. Tourism also plays a key role in border cities with large casinos, among them Windsor, Cornwall, Sarnia and Niagara Falls, the latter of which attracts millions of US and other international visitors. Once the dominant industry, agriculture now uses a small percentage of the workforce. However, much of the land in southern Ontario is given over to agriculture. As the following table shows, while the number of individual farms has steadily decreased and their overall size has shrunk at a lower rate, greater mechanization has supported increased supply to satisfy the ever-increasing demands of a growing population base; this has also meant a gradual increase in the total amount of land used for growing crops. Common types of farms reported in the 2001 census include those for cattle, small grains and dairy. The fruit- and wine industry is primarily on the Niagara Peninsula, Prince Edward County, and along the northern shore of Lake Erie, where tobacco farms are also situated. Market vegetables grow in the rich soils of the Holland Marsh near Newmarket. The area near Windsor is also very fertile. The Heinz plant in Leamington was taken over in these autumn of 2013 by Warren Buffett and a Brazilian partner, following which it put 740 people out of work. Government subsidies followed shortly; Premier Kathleen Wynne offered CAD$200,000 to cushion the blow, and promised that another processed-food operator would soon be found. On December 10, 2013, Kellogg's announced layoffs for more than 509 workers at a cereal manufacture plant in London. The area defined as the Corn Belt covers much of the southwestern area of the province, extending as far north as close to Goderich, but corn and soy are grown throughout the southern portion of the province. Apple orchards are a common sight along the southern shore of Nottawasaga Bay (part of Georgian Bay) near Collingwood and along the northern shore of Lake Ontario near Cobourg. Tobacco production, centred in Norfolk County, has decreased, allowing an increase in alternative crops such as hazelnuts and ginseng. The Ontario origins of Massey Ferguson, once one of the largest farm-implement manufacturers in the world, indicate the importance agriculture once had to the Canadian economy. Southern Ontario's limited supply of agricultural land is going out of production at an increasing rate. Urban sprawl and farmland severances contribute to the loss of thousands of acres of productive agricultural land in Ontario each year. Over 2,000 farms and 150,000 acres (61,000 ha) of farmland in the GTA alone were lost to production in the two decades between 1976 and 1996. This loss represented approximately 18%". of Ontario's Class 1 farmland being converted to urban purposes. In addition, increasing rural severances provide ever-greater interference with agricultural production. In an effort to protect the farmland and green spaces of the National Capital Region, and Greater Toronto Area, the Federal and Provincial Governments introduced greenbelts around Ottawa and the Golden Horseshoe, limiting urban development in these areas. Ontario's rivers make it rich in hydroelectric energy. In 2009, Ontario Power Generation generated 70 percent of the province's electricity, of which 51 percent is nuclear, 39% is hydroelectric and 10% is fossil-fuel derived. By 2025, nuclear power is projected to supply 42%, while fossil-fuel-derived generation is projected to decrease slightly over the next 20 years. Much of the newer power generation coming online in the last few years is natural gas or combined-cycle natural gas plants. OPG is not, however, responsible for the transmission of power, which is under the control of Hydro One. Despite its diverse range of power options, problems related to increasing consumption, lack of energy efficiency and aging nuclear reactors, Ontario has been forced in recent years to purchase power from its neighbours Quebec and Michigan to supplement its power needs during peak consumption periods. Ontario's basic domestic rate in 2010 was 11.17 cents per kWh; by contrast. Quebec's was 6.81. In December 2013, the government projected a 42 percent hike by 2018, and 68 percent by 2033. Industrial rates are projected to rise by 33% by 2018, and 55% in 2033. The Green Energy and Green Economy Act, 2009 (GEA), takes a two-pronged approach to commercializing renewable energy; first, it aims to bring more renewable energy sources to the province; and secondly, it aims to adopt more energy-efficiency measures to help conserve energy. The bill envisaged appointing a Renewable Energy Facilitator to provide "one-window" assistance and support to project developers to facilitate project approvals. The approvals process for transmission projects would also be streamlined and (for the first time in Ontario) the bill would enact standards for renewable energy projects. Homeowners would have access to incentives to develop small-scale renewables such as low- or no-interest loans to finance the capital cost of renewable energy generating facilities like solar panels. Ontario is home to Niagara Falls, which supplies a large amount of electricity to the province. The Bruce Nuclear Generating Station, the second largest operational nuclear power plant in the world, is also in Ontario and uses 8 CANDU reactors to generate electricity for the province. Ontario had the most wind energy capacity of the country with 4,900 MW of power (41% of Canada's capacity). The British North America Act 1867 section 69 stipulated "There shall be a Legislature for Ontario consisting of the Lieutenant Governor and of One House, styled the Legislative Assembly of Ontario." The assembly currently has 124 seats (increased from 107 as of the 42nd Ontario general election) representing ridings elected in a first-past-the-post system across the province. The legislative buildings at Queen's Park are the seat of government. Following the Westminster system, the leader of the party holding the most seats in the assembly is known as the "Premier and President of the Council" (Executive Council Act R.S.O. 1990). The Premier chooses the cabinet or Executive Council whose members are deemed ministers of the Crown. Although the Legislative Assembly Act (R.S.O. 1990) refers to "members of the assembly", the legislators are now commonly called MPPs (Members of the Provincial Parliament) in English and députés de l'Assemblée législative in French, but they have also been called MLAs (Members of the Legislative Assembly); both are acceptable but the latter is uncommon. The title of Prime Minister of Ontario, correct in French (le Premier ministre), is permissible in English but now generally avoided in favour of the title "Premier" to avoid confusion with the Prime Minister of Canada. Ontario has grown, from its roots in Upper Canada, into a modern jurisdiction. The old titles of the chief law officers, the Attorney-General and the Solicitor-General, remain in use. They both are responsible to the Legislature. The Attorney-General drafts the laws and is responsible for criminal prosecutions and the administration of justice, while the Solicitor-General is responsible for law enforcement and the police services of the province. The Municipal Act is the main statute governing the creation, administration and government of municipalities in the Canadian province of Ontario, other than the City of Toronto. After being passed in 2001, it came into force on January 1, 2003, replacing the previous Municipal Act. Effective January 1, 2007, the Municipal Act (the Act) was significantly amended by the Municipal Statute Law Amendment Act, 2006 (Bill 130). Ontario has numerous political parties which run for election. The four main parties are the centre-right Progressive Conservative Party of Ontario, the social democratic Ontario New Democratic Party (NDP), the centre to centre-left Ontario Liberal Party, and Green Party of Ontario. The Progressive Conservatives, Liberals and New Democrats have each governed the province, while the Greens elected their first member to the Legislative Assembly in 2018. The 2018 provincial election resulted in a Progressive Conservative majority government under party leader Doug Ford, who was sworn in as Premier on June 29. Ontario NDP leader Andrea Horwath was sworn in as the leader of her Majesty's Loyal Opposition. Ontario has three types of first-level administrative divisions. They include single-tier municipalities, upper-tier municipalities (which may be in the form of either regional municipalities or counties), and districts. Upper-tier municipalities and districts are made up of smaller municipalities and other types of administrative divisions. Administrative divisions differ primarily in the services that they provide to their residents, with the differing structures of these administrative regions resulting in disparities among Ontario's different regions. The administrative regions of Ontario are roughly coterminous with the census divisions used by Statistics Canada, although some exceptions do exist. Statistics Canada's measure of a "metro area", the Census Metropolitan Area (CMA), roughly bundles together population figures from the core municipality with those from "commuter" municipalities. *Parts of Quebec (including Gatineau) are included in the Ottawa CMA. The population of the Ottawa CMA, in both provinces, is shown. The Ontario portion of the CMA is about 75% of the total population of the CMA. In Canada, education falls under provincial jurisdiction. Publicly funded elementary and secondary schools are administered by the Ontario Ministry of Education, while colleges and universities are administered by the Ontario Ministry of Training, Colleges and Universities. The Minister of Education is Stephen Lecce, the Minister of Colleges and Universities is Ross Romano, and the Minister of Labour, Training and Skills Development Monte McNaughton. Higher education in Ontario includes post-secondary education and skills training regulated by the Ministry of Training, Colleges, and Universities and provided by universities, colleges of applied arts and technology, and private career colleges. The minister is Merrilee Fullerton. The ministry administers laws covering 22 public universities, 24 public colleges (21 Colleges of Applied Arts and Technology (CAATs) and three Institutes of Technology and Advanced Learning (ITALs)), 17 privately funded religious universities, and over 500 private career colleges. The Canadian constitution provides each province with the responsibility for higher education and there is no corresponding national federal ministry of higher education. Within Canadian federalism the division of responsibilities and taxing powers between the Ontario and Canadian governments creates the need for co-operation to fund and deliver higher education to students. Each higher education system aims to improve participation, access, and mobility for students. There are two central organizations that assist with the process of applying to Ontario universities and colleges: the Ontario Universities' Application Centre and Ontario College Application Service. While application services are centralized, admission and selection processes vary and are the purview of each institution. Admission to many Ontario postsecondary institutions can be highly competitive. Upon admission, students may get involved with regional student representation with the Canadian Federation of Students, the Canadian Alliance of Student Associations, the Ontario Undergraduate Student Alliance, or through the College Student Alliance in Ontario. Outdoor recreation is popular in Ontario and the region is home to numerous cultural events and festivals. There is no single regional dish in Ontario. Local fish and wild game, such as walleye and moose, are sometimes consumed. Poutine, a dish that originated in Quebec, is also popular in Ontario. In 2019, the government of Ontario passed legislation that established the Poet Laureate of Ontario. The largest museum in both Ontario and Canada is the Royal Ontario Museum, located in Toronto and founded in 1912. Receiving over one million visitors each year, it is also Canada's most popular museum. It features 40 exhibits containing "art, culture and nature from around the world." Iconic objects include: the world's largest faceted cerussite gem, Light of the Desert; four large totem poles, Nisga'a and Haida; and a Neo-Babylonian wall relief, Striding Lion. Ontario is also home to a number of national museums, due to the location of Ottawa. These include, among others, the Canadian War Museum, dedicated to Canada's military history, the Canadian Museum of Nature, dedicated to natural history and the Canada Science and Technology Museum, dedicated to the history of science and technology in Canada. There are also numerous other smaller, regional museums located in Ontario. Ontario has a particularly prominent role in Canadian music. The provincial capital city of Toronto, Canada's largest municipality, is home to much of the English Canadian music industry and many individual musicians, and the most popular destination for musicians from other parts of Canada, besides French-Canadian musicians, looking to advance their careers. In classical music, the Toronto Symphony Orchestra, and the National Arts Centre Orchestra are renowned internationally. Many smaller Ontario cities have orchestras of their own as well. The Canadian Opera Company, also based in Toronto, is the country's largest and most influential producer of opera productions. Other institutions in the province include the Royal Conservatory of Music, MuchMusic, National Ballet of Canada and concert venues such as Roy Thomson Hall, Massey Hall, the National Arts Centre, and the Four Seasons Centre. As of 2022, Ontario has 357 newspapers, 32 of which are daily, the highest in any province. Ontario is home to the largest newspaper in Canada, the Toronto Star, and Canada's newspaper of record, The Globe and Mail. Both are daily broadsheets based in Toronto. There are also numerous weekly newspapers for individual communities, though print publications for these papers have been on a downwards trend due to local news being shared on sites like Facebook. In 1973, the first slogan to appear on licence plates in Ontario was "Keep It Beautiful". This was replaced by "Yours to Discover" in 1982, which was originally used as a tourism slogan beginning in 1980. Plates with the French equivalent, Tant à découvrir, were made available to the public beginning in May 2008. (From 1988 to 1990, "Ontario Incredible" gave "Yours to Discover" a brief respite.) In 2020, as part of a license plate redesign, the slogan was changed to "A Place to Grow," inspired by the song A Place to Stand, a Place to Grow. This decision was reversed in the same year, due to visibility concerns. The slogan on licence plates remains "Yours to Discover". A Place to Stand, a Place to Grow is a song commissioned by the government of Ontario for its pavilion in Expo 67, and an unofficial anthem of the province. As a part of the Canada 150 celebrations in 2017, the provincial government released an updated rendition. In 2007, the provincial tourism agency commissioned a new song, "There's No Place Like This" is featured in television advertising, performed by Ontario artists including Molly Johnson, Brian Byrne, Keshia Chanté, as well as Tomi Swick and Arkells. The province has professional sports teams in baseball, basketball, Canadian football, ice hockey, lacrosse, rugby league, rugby union and soccer. Transportation in Ontario is under the purview of the Ministry of Transportation of Ontario and Transport Canada. Infrastructure and laws relating to road transport is the responsibility of the Ministry of Transportation, while infrastructure and laws relating to air, rail and marine transport is the responsibility of Transport Canada. As of October 2023, there are two Transport Canada designated international airports in Ontario They are Toronto Pearson International Airport, the busiest airport in Canada, handling almost 35 million passengers in 2022 and Ottawa Macdonald–Cartier International Airport, Ontario's second largest airport, handling over 2.5 million passengers in 2022. In addition to airports in Ottawa, and Toronto, the province also operates 11 other airports of entry. A number of Ontario cities also have regional airports, many of which have scheduled commuter flights from Jazz Aviation or smaller airlines and charter companies – flights from mid-size cities such as Sudbury and Sault Ste. Marie, to larger airports in Toronto and Ottawa. Bearskin Airlines also runs flights along the northerly east–west route, connecting North Bay, Sudbury, Sault Ste. Marie, and Thunder Bay directly. Remote and isolated towns and settlements in the northern areas of the province rely partly or entirely on air service for travel, goods, and even ambulance services (MEDIVAC), since much of the far northern area of the province cannot be reached by road (or by year-round road) or rail. Via Rail operates the inter-regional passenger train service on the Quebec City–Windsor Corridor, along with The Canadian, a transcontinental rail service from Southern Ontario to Vancouver, and the Sudbury–White River train. Additionally, Amtrak rail connects Ontario with key New York cities including Buffalo, Albany, and New York City. Ontario Northland provides rail service to destinations as far north as Moosonee near James Bay, connecting them with the south. Regional commuter rail is limited to the provincially owned GO Transit, and serves a train-bus network spanning the Golden Horseshoe region, with Union Station in Toronto serving as the transport hub. Freight rail is dominated by the founding cross-country Canadian National Railway and CP Rail companies. As of 2021, there are 19,979 km of railways in operation. There are several city rail-transit systems in the Province. The Toronto Transit Commission operates subways, as well as streetcars (being one of the busiest streetcar systems in North America). OC Transpo operates a light rail metro system in Ottawa. In addition, Waterloo region operates a surface light rail system. Construction on light rail lines are also underway in the Regional Municipality of Peel, and are expected to be completed by late 2024. 400-series highways make up the primary vehicular network in the south of province, and they connect at a number of points to border crossings to the United States, and Quebec, the busiest being the Detroit–Windsor Tunnel and Ambassador Bridge and the Blue Water Bridge (via Highway 402). Some of the primary highways along the southern route are Highway 401, Highway 417, and Highway 400, Highway 401 being the busiest highway in North America. Other provincial highways and regional roads inter-connect the remainder of the province, and the Trans-Canada Highway connects the province to the rest of the country. The Saint Lawrence Seaway, which extends across most of the southern portion of the province and connects to the Atlantic Ocean, is the primary water transportation route for cargo, particularly iron ore and grain. In the past, the Great Lakes and St. Lawrence River were also a major passenger transportation route, but over the past half century passenger travel has been reduced to ferry services and sightseeing cruises. Ontario's three largest ports are the Port of Hamilton, Port of Thunder Bay and the Port of Nanticoke. Ontario's only saltwater port is located in the town of Moosonee on James Bay.
{{Infobox province or territory of Canada | name = Manitoba | other_name = {{native name|cr|Manitou-wapow}}<br>{{native name|oj|Manidoobaa}} | settlement_type = [[Provinces and territories of Canada|Province]] | image_flag = Flag of Manitoba.svg | flag_alt = A red flag with a large Union Jack in the upper left corner and a shield, consisting of St. George's Cross over a left-facing bison standing on a rock, on the right side | image_shield = Coat of arms of Manitoba.svg | shield_alt = A central shield showing a bison standing on a rock, under a St George's Cross. On top of the shield sits a helmet decorated with a red and white billowing veil. On top of the helmet sits a beaver with a crown on its back, holding a [[prairie crocus]]. To the right of the shield is a rearing white unicorn wearing a collar of white and green maple leaves, from which hangs a green cart-wheel pendant. To the left of the shield is a rearing white horse wearing a collar of Indian beadwork, from which hangs a green cycle of life medallion. The animals and shield stand on a mound, with a wheat field beneath the unicorn, prairie crocuses beneath the shield, and spruces beneath the horse. Beneath the mound are white and blue waves, under which is an orange scroll bearing the words "GLORIOSUS ET LIBER" | motto = {{lang-la|Gloriosus et Liber}}<br />{{small|''"Glorious and free"''}}<br />{{small|{{native phrase|fr|"Glorieux et libre"}}}} | image_map = Manitoba in Canada 2.svg | map_alt = Map showing Manitoba's location in the centre of Southern Canada | Label_map = yes | coordinates = {{Coord|55|00|00|N|97|00|00|W|type:adm1st_scale:30000000_region:CA-MB_source:http://www4.rncan.gc.ca/search-place-names/unique/GBGYI|display=inline,title}} | capital = [[Winnipeg]] | largest_city = Winnipeg | largest_metro = [[Winnipeg Capital Region]] | official_lang = [[English language|English]]<ref>{{cite web | title = The legal context of Canada's official languages | publisher = University of Ottawa | url = https://slmc.uottawa.ca/?q=english_french_legal | accessdate = 7 March 2019 | archive-url = https://web.archive.org/web/20171010214715/https://slmc.uottawa.ca/?q=english_french_legal | archive-date = 10 October 2017 | url-status = dead }}</ref> | Viceroy = [[Janice Filmon]] | ViceroyType = Lieutenant Governor | Premier = [[Heather Stefanson]] | PremierParty = [[Progressive Conservative Party of Manitoba|Progressive Conservative]] | Legislature = Legislative Assembly of Manitoba | area_rank = 8th | area_total_km2 = 649950 | area_land_km2 = 548360 | area_water_km2 = 101593 | PercentWater = 15.6 | population_demonym = Manitoban | population_rank = 5th | population_total = 1278365 <!-- 2016 StatCan federal census population only per [[WP:CANPOP]]. Do not update until 2021 census population released. Use "Population_est" below for latest StatCan quarterly estimate. --> | population_ref =<ref name="census2016">{{cite web | url=http://www12.statcan.gc.ca/census-recensement/2016/dp-pd/hlt-fst/pd-pl/Table.cfm?Lang=Eng&T=101&S=50&O=A | title=Population and dwelling counts, for Canada, provinces and territories, 2016 and 2011 censuses | work=[[Statistics Canada]] | date=2 February 2017 | accessdate=30 April 2017 | url-status=live | archiveurl=https://web.archive.org/web/20170504080727/http://www12.statcan.gc.ca/census-recensement/2016/dp-pd/hlt-fst/pd-pl/Table.cfm?Lang=Eng&T=101&S=50&O=A | archivedate=4 May 2017 | df=dmy-all }}</ref> | population_est = 1379584<!-- Latest StatCan quarterly estimate only. --> | pop_est_as_of = 2020 Q4 | pop_est_ref =<ref name=StatCan2016Q1Est>{{cite web | url=http://www5.statcan.gc.ca/cansim/a26?lang=eng&retrLang=eng&id=0510005&&pattern=&stByVal=1&p1=1&p2=31&tabMode=dataTable&csid= | title=Estimates of population, Canada, provinces and territories | publisher=[[Statistics Canada]] | date=28 September 2016 | accessdate=29 September 2018 | url-status=live | archiveurl=https://web.archive.org/web/20171223040248/http://www5.statcan.gc.ca/cansim/a26?lang=eng&retrLang=eng&id=0510005&&pattern=&stByVal=1&p1=1&p2=31&tabMode=dataTable&csid= | archivedate=23 December 2017 | df=dmy-all }}</ref> | DensityRank = 8th | Density_km2 = 2.3<ref name=census2011>{{cite web |url=http://www12.statcan.ca/census-recensement/2011/dp-pd/hlt-fst/pd-pl/Table-Tableau.cfm?LANG=Eng&T=101&S=50&O=A |title=Population and dwelling counts, for Canada, provinces and territories, 2011 and 2006 censuses |publisher=Statistics Canada |date=8 February 2012 |accessdate=8 February 2012 |archivedate=7 January 2019 |archiveurl=https://web.archive.org/web/20190107044952/https://www12.statcan.gc.ca/census-recensement/2011/dp-pd/hlt-fst/pd-pl/Table-Tableau.cfm?LANG=Eng&T=101&S=50&O=A%20 }}</ref> | GDP_year = 2015 | GDP_total = C$65.862&nbsp;billion<ref name=GDP2011>{{cite web | url=http://www.statcan.gc.ca/tables-tableaux/sum-som/l01/cst01/econ15-eng.htm | title=Gross domestic product, expenditure-based, by province and territory (2015) | publisher=Statistics Canada | date=9 November 2016 | accessdate=26 January 2017 | url-status=live| archiveurl=https://web.archive.org/web/20120919211233/http://www.statcan.gc.ca/tables-tableaux/sum-som/l01/cst01/econ15-eng.htm | archivedate=19 September 2012 | df=dmy-all }}</ref> | GDP_rank = 6th | GDP_per_capita = C$50,820 | GDP_per_capita_rank = 9th | AdmittanceOrder = 5th | AdmittanceDate = 15 July 1870 | HouseSeats = 14 | SenateSeats = 6 | timezone1 = [[Central Time Zone|Central]] | utc_offset1 = -06:00 | timezone1_DST = [[Central Time Zone|Central DST]] | utc_offset1_DST = -05:00 | PostalAbbreviation = | PostalCodePrefix = }} '''Manitoba''' is a province of [[Canada]] that is roughly in the country's centre. Manitoba is Canada's sixth-largest province, with an area of {{convert|647797|km2}}. It has the fifth-largest number of people, at 1,379,584 in 2020, and its people are called "Manitobans". ==History== People have been living in Manitoba for thousands of years. Both the [[Hudson's Bay Company]] from [[England]] and many people from [[France]] moved to Manitoba during the 17th, 18th, and 19th centuries. Manitobabecame part of the [[Northwest Territories]] in 1869. The [[Red River Rebellion]], which was started by [[Louis Riel]], began there. Manitoba became part of Canada on 12 May 1870. The province then included only the southern part, which is near the [[United States]], until parts of the Northwest Territories were later added. ==Government== The capital of Manitoba is [[Winnipeg, Manitoba|Winnipeg]]. Other large cities in Manitoba are Steinbach and Brandon. The people of Manitoba elect a legislature. The leader of the government, who is called the [[premier]], is the leader of the largest party in the legislature. There is also a [[Lieutenant Governor|lieutenant governor]], who represents the [[Monarchy of the United Kingdom|British monarch]]. The main [[Political party|political parties]] in Manitoba are the [[New Democratic Party]], the [[Progressive Conservative Party of Manitoba|Progressive Conservative Party]], and the [[Liberal Party of Canada|Liberal Party]]. ==References== {{reflist}} == Other websites == {{sisterlinks|Manitoba}} {{wikivoyage|Manitoba}} * {{dmoz|Regional/North_America/Canada/Manitoba}} {{Provinces and territories of Canada}} {{canada-stub}} [[Category:Manitoba| ]] [[Category:1870 establishments in North America]] [[Category:1870s establishments in Canada]]
Manitoba (/ˌmænɪˈtoʊbə/ MAN-ih-TOH-bə) is a province of Canada at the longitudinal centre of the country. It is Canada's fifth-most populous province, with a population of 1,342,153 as of 2021. Manitoba has a widely varied landscape, from arctic tundra and the Hudson Bay coastline in the north to dense boreal forest, large freshwater lakes, and prairie grassland in the central and southern regions. Indigenous peoples have inhabited what is now Manitoba for thousands of years. In the early 17th century, English and French fur traders began arriving in the area and establishing settlements. The Kingdom of England secured control of the region in 1673 and created a territory named Rupert's Land, which was placed under the administration of the Hudson's Bay Company. Rupert's Land, which included all of present-day Manitoba, grew and evolved from 1673 until 1869 with significant settlements of Indigenous and Métis people in the Red River Colony. Negotiations for the creation of the province of Manitoba commenced in 1869, but deep disagreements over the right to self-determination led to an armed conflict, known as the Red River Rebellion, between the federal government and the people (particularly Métis) of the Red River Colony. The resolution of the conflict and further negotiations led to Manitoba becoming the fifth province to join Canadian Confederation, when the Parliament of Canada passed the Manitoba Act on July 15, 1870. Manitoba's capital and largest city is Winnipeg, the sixth most populous municipality in Canada. Winnipeg is the seat of government, home to the Legislative Assembly of Manitoba and the Provincial Court. Four of the province's five universities, all four of its professional sports teams, and most of its cultural activities (including Festival du Voyageur and Folklorama) are located in Winnipeg. The city has an international airport as well as train and bus stations; a Canadian Forces base, CFB Winnipeg, operates from the airport and is the regional headquarters of the North American Aerospace Defense Command. The name Manitoba possibly derives from either Cree manitou-wapow or Ojibwe manidoobaa, both meaning 'straits of Manitou, the Great Spirit'. Alternatively, it may be from the Assiniboine minnetoba, meaning 'Lake of the Prairie' (the lake was known to French explorers as Lac des Prairies). The name was chosen by Thomas Spence for the new republic he proposed for the area south of the lake. Métis leader Louis Riel preferred the name over the proposed alternative of "Assiniboia". It was accepted in Ottawa under the Manitoba Act, 1870. Modern-day Manitoba was inhabited by the First Nations people shortly after the last ice age glaciers retreated in the southwest about 10,000 years ago; the first exposed land was the Turtle Mountain area. The Ojibwe, Cree, Dene, Sioux, Mandan, and Assiniboine peoples founded settlements, and other tribes entered the area to trade. In Northern Manitoba, quartz was mined to make arrowheads. The first farming in Manitoba was along the Red River, where corn and other seed crops were planted before contact with Europeans. In 1611, Henry Hudson was one of the first Europeans to sail into what is now known as Hudson Bay, where he was abandoned by his crew. Thomas Button travelled this area in 1612 in an unsuccessful attempt to find and rescue Hudson. When the British ship Nonsuch sailed into Hudson Bay in 1668–1669, she became the first trading vessel to reach the area; that voyage led to the formation of the Hudson's Bay Company, to which the British government gave absolute control of the entire Hudson Bay watershed. This watershed was named Rupert's Land, after Prince Rupert, who helped to subsidize the Hudson's Bay Company. York Factory was founded in 1684 after the original fort of the Hudson's Bay Company, Fort Nelson (built in 1682), was destroyed by rival French traders. Pierre Gaultier de Varennes, sieur de La Vérendrye, visited the Red River Valley in the 1730s to help open the area for French exploration and trade. As French explorers entered the area, a Montreal-based company, the North West Company, began trading with the local Indigenous people. Both the North West Company and the Hudson's Bay Company built fur-trading forts; the two companies competed in southern Manitoba, occasionally resulting in violence, until they merged in 1821 (the Hudson's Bay Company Archives in Winnipeg preserve the history of this era). Great Britain secured the territory in 1763 after their victory over France in the North American theatre of the Seven Years' War, better known as the French and Indian War in North America; lasting from 1754 to 1763. The founding of the first agricultural community and settlements in 1812 by Lord Selkirk, north of the area which is now downtown Winnipeg, led to conflict between British colonists and the Métis. Twenty colonists, including the governor, and one Métis were killed in the Battle of Seven Oaks in 1816. Rupert's Land was ceded to Canada by the Hudson's Bay Company in 1869 and incorporated into the Northwest Territories; a lack of attention to Métis concerns caused Métis leader Louis Riel to establish a local provisional government which formed into the Convention of Forty and the subsequent elected Legislative Assembly of Assiniboia on 9 March 1870. This assembly subsequently sent three delegates to Ottawa to negotiate with the Canadian government. This resulted in the Manitoba Act and that province's entry into the Canadian Confederation. Prime Minister Sir John A. Macdonald introduced the Manitoba Act in the House of Commons of Canada, the bill was given Royal Assent and Manitoba was brought into Canada as a province in 1870. Louis Riel was pursued by British army officer Garnet Wolseley because of the rebellion, and Riel fled into exile. The Canadian government blocked the Métis' attempts to obtain land promised to them as part of Manitoba's entry into confederation. Facing racism from the new flood of white settlers from Ontario, large numbers of Métis moved to what would become Saskatchewan and Alberta. Numbered Treaties were signed in the late 19th century with the chiefs of First Nations that lived in the area. They made specific promises of land for every family. As a result, a reserve system was established under the jurisdiction of the federal government. The prescribed amount of land promised to the native peoples was not always given; this led Indigenous groups to assert rights to the land through land claims, many of which are still ongoing. The original province of Manitoba was a square one-eighteenth of its current size, and was known colloquially as the "postage stamp province". Its borders were expanded in 1881, taking land from the Northwest Territories and the District of Keewatin, but Ontario claimed a large portion of the land; the disputed portion was awarded to Ontario in 1889. Manitoba grew to its current size in 1912, absorbing land from the Northwest Territories to reach 60°N, uniform with the northern reach of its western neighbours Saskatchewan, Alberta and British Columbia. The Manitoba Schools Question showed the deep divergence of cultural values in the territory. The Catholic Franco-Manitobans had been guaranteed a state-supported separate school system in the original constitution of Manitoba, but a grassroots political movement among English Protestants from 1888 to 1890 demanded the end of French schools. In 1890, the Manitoba legislature passed a law removing funding for French Catholic schools. The French Catholic minority asked the federal government for support; however, the Orange Order and other anti-Catholic forces mobilized nationwide to oppose them. The federal Conservatives proposed remedial legislation to override Manitoba, but they were blocked by the Liberals, led by Wilfrid Laurier. Once elected Prime Minister in 1896, Laurier implemented a compromise stating Catholics in Manitoba could have their own religious instruction for 30 minutes at the end of the day if there were enough students to warrant it, implemented on a school-by-school basis. By 1911, Winnipeg was the third largest city in Canada, and remained so until overtaken by Vancouver in the 1920s. A boomtown, it grew quickly around the start of the 20th century, with outside investors and immigrants contributing to its success. The drop in growth in the second half of the decade was a result of the opening of the Panama Canal in 1914, which reduced reliance on transcontinental railways for trade, as well as a decrease in immigration due to the outbreak of the First World War. Over 18,000 Manitoba residents enlisted in the first year of the war; by the end of the war, 14 Manitobans had received the Victoria Cross. During the First World War, Nellie McClung started the campaign for women's votes. On January 28, 1916, the vote for women was legalized. Manitoba was the first province to allow women to vote in provincial elections. This was two years before Canada as a country granted women the right to vote. After the First World War ended, severe discontent among farmers (over wheat prices) and union members (over wage rates) resulted in an upsurge of radicalism, coupled with a polarization over the rise of Bolshevism in Russia. The most dramatic result was the Winnipeg general strike of 1919. It began on 15 May and collapsed on 25 June 1919; as the workers gradually returned to their jobs, the Central Strike Committee decided to end the movement. Government efforts to violently crush the strike, including a Royal North-West Mounted Police charge into a crowd of protesters that resulted in multiple casualties and one death, had led to the arrest of the movement's leaders. In the aftermath, eight leaders went on trial, and most were convicted on charges of seditious conspiracy, illegal combinations, and seditious libel; four were deported under the Canadian Immigration Act. The Great Depression (1929–c. 1939) hit especially hard in Western Canada, including Manitoba. The collapse of the world market combined with a steep drop in agricultural production due to drought led to economic diversification, moving away from a reliance on wheat production. The Manitoba Co-operative Commonwealth Federation, forerunner to the New Democratic Party of Manitoba (NDP), was founded in 1932. Canada entered the Second World War in 1939. Winnipeg was one of the major commands for the British Commonwealth Air Training Plan to train fighter pilots, and there were air training schools throughout Manitoba. Several Manitoba-based regiments were deployed overseas, including Princess Patricia's Canadian Light Infantry. In an effort to raise money for the war effort, the Victory Loan campaign organized "If Day" in 1942. The event featured a simulated Nazi invasion and occupation of Manitoba, and eventually raised over C$65 million. Winnipeg was inundated during the 1950 Red River Flood and had to be partially evacuated. In that year, the Red River reached its highest level since 1861 and flooded most of the Red River Valley. The damage caused by the flood led then-Premier Duff Roblin to advocate for the construction of the Red River Floodway; it was completed in 1968 after six years of excavation. Permanent dikes were erected in eight towns south of Winnipeg, and clay dikes and diversion dams were built in the Winnipeg area. In 1997, the "Flood of the Century" caused over C$400 million in damages in Manitoba, but the floodway prevented Winnipeg from flooding. In 1990, Prime Minister Brian Mulroney attempted to pass the Meech Lake Accord, a series of constitutional amendments to persuade Quebec to endorse the Canada Act 1982. Unanimous support in the legislature was needed to bypass public consultation. Cree politician Elijah Harper opposed because he did not believe First Nations had been adequately involved in the Accord's process, and thus the Accord failed. Glen Murray, elected in Winnipeg in 1998, became the first openly gay mayor of a large North American city. The province was impacted by major flooding in 2009 and 2011. In 2004, Manitoba became the first province in Canada to ban indoor smoking in public places. In 2013, Manitoba was the second province to introduce accessibility legislation, protecting the rights of persons with disabilities. Manitoba is bordered by the provinces of Ontario to the east and Saskatchewan to the west, the territory of Nunavut to the north, and the US states of North Dakota and Minnesota to the south. Manitoba is at the centre of the Hudson Bay drainage basin, with a high volume of the water draining into Lake Winnipeg and then north down the Nelson River into Hudson Bay. This basin's rivers reach far west to the mountains, far south into the United States, and east into Ontario. Major watercourses include the Red, Assiniboine, Nelson, Winnipeg, Hayes, Whiteshell and Churchill rivers. Most of Manitoba's inhabited south has developed in the prehistoric bed of Glacial Lake Agassiz. This region, particularly the Red River Valley, is flat and fertile; receding glaciers left hilly and rocky areas throughout the province. The province has a saltwater coastline bordering Hudson Bay and more than 110,000 lakes, covering approximately 15.6 percent or 101,593 square kilometres (39,225 sq mi) of its surface area. Manitoba's major lakes are Lake Manitoba, Lake Winnipegosis, and Lake Winnipeg, the tenth-largest freshwater lake in the world. A total of 29,000 square kilometres (11,000 sq mi) of traditional First Nations lands and boreal forest on Lake Winnipeg's east side were officially designated as a UNESCO World Heritage Site known as Pimachiowin Aki in 2018. Baldy Mountain is the province's highest point at 832 metres (2,730 ft) above sea level, and the Hudson Bay coast is the lowest at sea level. Riding Mountain, the Pembina Hills, Sandilands Provincial Forest, and the Canadian Shield are also upland regions. Much of the province's sparsely inhabited north and east lie on the irregular granite Canadian Shield, including Whiteshell, Atikaki, and Nopiming Provincial Parks. Extensive agriculture is found only in the province's southern areas, although there is grain farming in the Carrot Valley Region (near The Pas). Around 11 percent of Canada's farmland is in Manitoba. Manitoba has an extreme continental climate. Temperatures and precipitation generally decrease from south to north and increase from east to west. Manitoba is far from the moderating influences of mountain ranges or large bodies of water. Because of the generally flat landscape, it is exposed to cold Arctic high-pressure air masses from the northwest during January and February. In the summer, air masses sometimes come out of the Southern United States, as warm humid air is drawn northward from the Gulf of Mexico. Temperatures exceed 30 °C (86 °F) numerous times each summer, and the combination of heat and humidity can bring the humidex value to the mid-40s. Carman, Manitoba, recorded the second-highest humidex ever in Canada in 2007, with 53.0. According to Environment Canada, Manitoba ranked first for clearest skies year round and ranked second for clearest skies in the summer and for the sunniest province in the winter and spring. Southern Manitoba (including the city of Winnipeg), falls into the humid continental climate zone (Köppen Dfb). This area is cold and windy in the winter and often has blizzards because of the open landscape. Summers are warm with a moderate length. This region is the most humid area in the prairie provinces, with moderate precipitation. Southwestern Manitoba, though under the same climate classification as the rest of Southern Manitoba, is closer to the semi-arid interior of Palliser's Triangle. The area is drier and more prone to droughts than other parts of southern Manitoba. This area is cold and windy in the winter and has frequent blizzards due to the openness of the Canadian Prairie landscape. Summers are generally warm to hot, with low to moderate humidity. Southern parts of the province, just north of Tornado Alley, experience tornadoes, with 16 confirmed touchdowns in 2016. In 2007, on 22 and 23 June, numerous tornadoes touched down, the largest an F5 tornado that devastated parts of Elie (the strongest recorded tornado in Canada). The province's northern sections (including the city of Thompson) fall in the subarctic climate zone (Köppen climate classification Dfc). This region features long and extremely cold winters and brief, warm summers with little precipitation. Overnight temperatures as low as −40 °C (−40 °F) occur on several days each winter. Manitoba natural communities may be grouped within five ecozones: boreal plains, prairie, taiga shield, boreal shield and Hudson plains. Three of these—taiga shield, boreal shield and Hudson plain—contain part of the Boreal forest of Canada which covers the province's eastern, southeastern, and northern reaches. Forests make up about 263,000 square kilometres (102,000 sq mi), or 48 percent, of the province's land area. The forests consist of pines (Jack Pine, Red Pine, Eastern White Pine), spruces (White Spruce, Black Spruce), Balsam Fir, Tamarack (larch), poplars (Trembling Aspen, Balsam Poplar), birches (White Birch, Swamp Birch) and small pockets of Eastern White Cedar. Two sections of the province are not dominated by forest. The province's northeast corner bordering Hudson Bay is above the treeline and considered tundra. The tallgrass prairie once dominated the south-central and southeastern regions, including the Red River Valley. Mixed grass prairie is found in the southwestern region. Agriculture has replaced much of the natural vegetation but prairie can still be found in parks and protected areas; some are notable for the presence of the endangered western prairie fringed orchid. Manitoba is especially noted for its northern polar bear population; Churchill is commonly referred to as the "Polar Bear Capital". In the waters off the northern coast of the province are numerous marine species, including the beluga whale. Other populations of animals, including moose, white-tailed deer, mule deer, black and brown bears, coyote, cougar, red fox, Canada lynx, and gray wolf, are distributed throughout the province, especially in the provincial and national parks. There is a large population of red-sided garter snakes near Narcisse; the overwintering dens there are seasonally home to the world's largest concentration of snakes. Manitoba's bird diversity is enhanced by its position on two major migration routes, with 392 confirmed identified species; 287 of these nesting within the province. These include the great grey owl, the province's official bird, and the endangered peregrine falcon. Manitoba's lakes host 18 species of game fish, particularly species of trout, pike, and goldeye, as well as many smaller fish. At the 2021 census, Manitoba had a population of 1,342,153, more than half of which is in Winnipeg. Although initial colonization of the province revolved mostly around homesteading, the last century has seen a shift towards urbanization; Manitoba is the only Canadian province with over fifty-five percent of its population in a single city. The largest ethnic group in Manitoba is English (16.1%), followed by Scottish (14.5%), German (13.6%), Ukrainian (12.6%), Irish (11.0%), French (9.3%), Filipino (7.0%), Métis (6.8%), Polish (6.0%), First Nations (4.5%), Mennonite (3.9%), Russian (3.7%), Dutch (3.3%), Indian (3.0%), and Icelandic (2.4%). Indigenous peoples (including Métis) are Manitoba's fastest-growing ethnic group, representing 13.6 percent of Manitoba's population as of 2001 (some reserves refused to allow census-takers to enumerate their populations or were otherwise incompletely counted). Gimli, Manitoba is home to the largest Icelandic community outside of Iceland. As of the 2021 Canadian Census, the ten most spoken languages in the province included English (1,288,950 or 98.6%), French (111,790 or 8.55%), Tagalog (73,440 or 5.62%), Punjabi (42,820 or 3.28%), German (41,980 or 3.21%), Hindi (26,980 or 2.06%), Spanish (23,435 or 1.79%), Mandarin (16,765 or 1.28%), Cree (16,115 or 1.23%), and Plautdietsch (15,055 or 1.15%). The question on knowledge of languages allows for multiple responses. Most Manitobans belong to a Christian denomination: on the 2021 census, 54.2% reported being Christian, followed by 2.7% Sikh, 2.0% Muslim, 1.4% Hindu, 0.9% Jewish, and 0.8% Indigenous spirituality. 36.7% reported no religious affiliation. The largest Christian denominations by number of adherents were the Roman Catholic Church with 21.2%; the United Church of Canada with 5.8%; and the Anglican Church of Canada with 3.3%. Manitoba has a moderately strong economy based largely on natural resources. Its Gross Domestic Product was C$50.834 billion in 2008. The province's economy grew 2.4 percent in 2008, the third consecutive year of growth. The average individual income in Manitoba in 2006 was C$25,100 (compared to a national average of C$26,500), ranking fifth-highest among the provinces. As of October 2009, Manitoba's unemployment rate was 5.8 percent. Manitoba's economy relies heavily on agriculture, tourism, electricity, oil, mining, and forestry. Agriculture is vital and is found mostly in the southern half of the province, although grain farming occurs as far north as The Pas. The most common agricultural activity is cattle husbandry, followed by assorted grains and oilseed. Manitoba is the nation's largest producer of sunflower seed and dry beans, and one of the leading sources of potatoes. Portage la Prairie is a major potato processing centre. Richardson International, one of the largest oat mills in the world, also has a plant in the municipality. Manitoba's largest employers are government and government-funded institutions, including crown corporations and services like hospitals and universities. Major private-sector employers are The Great-West Life Assurance Company, Cargill Ltd., and Richardson International. Manitoba also has large manufacturing and tourism sectors. Churchill's Arctic wildlife is a major tourist attraction; the town is a world capital for polar bear and beluga whale watchers. Manitoba is the only province with an Arctic deep-water seaport, at Churchill. In January 2018, the Canadian Federation of Independent Business claimed Manitoba was the most improved province for tackling red tape. Manitoba's early economy depended on mobility and living off the land. Indigenous Nations (Cree, Ojibwa, Dene, Sioux and Assiniboine) followed herds of bison and congregated to trade among themselves at key meeting places throughout the province. After the arrival of the first European traders in the 17th century, the economy centred on the trade of beaver pelts and other furs. Diversification of the economy came when Lord Selkirk brought the first agricultural settlers in 1811, though the triumph of the Hudson's Bay Company (HBC) over its competitors ensured the primacy of the fur trade over widespread agricultural colonization. HBC control of Rupert's Land ended in 1868; when Manitoba became a province in 1870, all land became the property of the federal government, with homesteads granted to settlers for farming. Transcontinental railways were constructed to simplify trade. Manitoba's economy depended mainly on farming, which persisted until drought and the Great Depression led to further diversification. CFB Winnipeg is a Canadian Forces Base at the Winnipeg International Airport. The base is home to flight operations support divisions and several training schools, as well as the 1 Canadian Air Division and Canadian NORAD Region Headquarters. 17 Wing of the Canadian Forces is based at CFB Winnipeg; the Wing has three squadrons and six schools. It supports 113 units from Thunder Bay to the Saskatchewan/Alberta border, and from the 49th parallel north to the high Arctic. 17 Wing acts as a deployed operating base for CF-18 Hornet fighter–bombers assigned to the Canadian NORAD Region. The two 17 Wing squadrons based in the city are: the 402 ("City of Winnipeg" Squadron), which flies the Canadian designed and produced de Havilland Canada CT-142 Dash 8 navigation trainer in support of the 1 Canadian Forces Flight Training School's Air Combat Systems Officer and Airborne Electronic Sensor Operator training programs (which trains all Canadian Air Combat Systems Officer); and the 435 ("Chinthe" Transport and Rescue Squadron), which flies the Lockheed C-130 Hercules tanker/transport in airlift search and rescue roles, and is the only Air Force squadron equipped and trained to conduct air-to-air refuelling of fighter aircraft. Canadian Forces Base Shilo (CFB Shilo) is an Operations and Training base of the Canadian Forces 35 kilometres (22 mi) east of Brandon. During the 1990s, Canadian Forces Base Shilo was designated as an Area Support Unit, acting as a local base of operations for Southwest Manitoba in times of military and civil emergency. CFB Shilo is the home of the 1st Regiment, Royal Canadian Horse Artillery, both battalions of the 1 Canadian Mechanized Brigade Group, and the Royal Canadian Artillery. The Second Battalion of Princess Patricia's Canadian Light Infantry (2 PPCLI), which was originally stationed in Winnipeg (first at Fort Osborne, then in Kapyong Barracks), has operated out of CFB Shilo since 2004. CFB Shilo hosts a training unit, 3rd Canadian Division Training Centre. It serves as a base for support units of 3rd Canadian Division, also including 3 CDSG Signals Squadron, Shared Services Unit (West), 11 CF Health Services Centre, 1 Dental Unit, 1 Military Police Regiment, and an Integrated Personnel Support Centre. The base houses 1,700 soldiers. After the control of Rupert's Land was passed from Great Britain to the Government of Canada in 1869, Manitoba attained full-fledged rights and responsibilities of self-government as the first Canadian province carved out of Rupert's Land. The Legislative Assembly of Manitoba was established on 14 July 1870. Political parties first emerged between 1878 and 1883, with a two-party system (Liberals and Conservatives). The United Farmers of Manitoba appeared in 1922, and later merged with the Liberals in 1932. Other parties, including the Co-operative Commonwealth Federation (CCF), appeared during the Great Depression; in the 1950s, Manitoban politics became a three-party system, and the Liberals gradually declined in power. The CCF became the New Democratic Party of Manitoba (NDP), which came to power in 1969. Since then, the Progressive Conservatives and the NDP have been the dominant parties. Like all Canadian provinces, Manitoba is governed by a unicameral legislative assembly. The executive branch is formed by the governing party; the party leader is the premier of Manitoba, the head of the executive branch. The head of state, King Charles III, is represented by the lieutenant governor of Manitoba, who is appointed by the governor general of Canada on advice of the prime minister. The head of state is primarily a ceremonial role, although the lieutenant governor has the official responsibility of ensuring Manitoba has a duly constituted government. The Legislative Assembly consists of the 57 Members elected to represent the people of Manitoba. The premier of Manitoba is Wab Kinew, who was elected in the 2023 provincial election. The province is represented in federal politics by 14 Members of Parliament and six Senators. Manitoba's judiciary consists of the Court of Appeal, the Court of King's Bench, and the Provincial Court. The Provincial Court is primarily for criminal law; 95 per cent of criminal cases in Manitoba are heard here. The Court of King's Bench is the highest trial court in the province. It has four jurisdictions: family law (child and family services cases), civil law, criminal law (for indictable offences), and appeals. The Court of Appeal hears appeals from both benches; its decisions can only be appealed to the Supreme Court of Canada. Both English and French are official languages of the legislature and courts of Manitoba, according to section 23 of the Manitoba Act, 1870 (part of the Constitution of Canada). In April 1890, the Manitoba legislature attempted to abolish the official status of French and ceased to publish bilingual legislation. However, in 1985, the Supreme Court of Canada ruled in the Reference re Manitoba Language Rights that section 23 still applied, and that legislation published only in English was invalid (unilingual legislation was declared valid for a temporary period to allow time for translation). Although French is an official language for the purposes of the legislature, legislation, and the courts, the Manitoba Act does not require it to be an official language for the purpose of the executive branch (except when performing legislative or judicial functions). Hence, Manitoba's government is not completely bilingual. The Manitoba French Language Services Policy of 1999 is intended to provide a comparable level of provincial government services in both official languages. According to the 2006 Census, 82.8 percent of Manitoba's population spoke only English, 3.2 percent spoke only French, 15.1 percent spoke both, and 0.9 percent spoke neither. In 2010, the provincial government of Manitoba passed the Aboriginal Languages Recognition Act, which gives official recognition to seven indigenous languages: Cree, Dakota, Dene, Inuktitut, Michif, Ojibway and Oji-Cree. Manitoba has two Class I railways: Canadian National Railway (CN) and Canadian Pacific Railway (CPR). Winnipeg is centrally located on the main lines of both carriers, and both maintain large inter-modal terminals in the city. Via Rail offers transcontinental and Northern Manitoba passenger service from Winnipeg's Union Station. Numerous small regional and short-line railways also run trains within Manitoba: the Hudson Bay Railway, the Southern Manitoba Railway, Burlington Northern Santa Fe Manitoba, Greater Winnipeg Water District Railway, and Central Manitoba Railway. Winnipeg James Armstrong Richardson International Airport, Manitoba's largest airport, is one of only a few 24-hour unrestricted airports in Canada and is part of the National Airports System. A new, larger terminal opened in October 2011. It is the seventh busiest airport in Canada by passenger traffic, serving 4,484,343 passengers in 2018, and the 11th busiest airport by aircraft movements. The airport handles approximately 195,000 tonnes (430,000,000 lb) of cargo annually, making it the third largest cargo airport in the country. Winnipeg is a major sorting facility for both FedEx and Purolator, and receives daily trans-border service from UPS. The Port of Churchill is the only Arctic deep-water port in Canada. It is nautically closer to ports in Northern Europe and Russia than any other port in Canada. It has four deep-sea berths for the loading and unloading of grain, general cargo and tanker vessels. The port is served by the Hudson Bay Railway. The port and railway came under complete community and Indigenous ownership in 2021, after AGT Food and Ingredients and Fairfax Financial transferred their shares in Arctic Gateway to OneNorth – a consortium of community and Indigenous partners which owned the other fifty percent of Arctic Gateway's shares. Public schools follow a provincially mandated curriculum in either French or English. There are sixty-five funded independent schools in Manitoba, including three boarding schools. These schools must follow the Manitoban curriculum and meet other provincial requirements. There are forty-four non-funded independent schools, which are not required to meet those standards. Public schools in Manitoba fall under the regulation of one of thirty-seven school divisions within the provincial education system (except for the Manitoba Band Operated Schools, which are administered by the federal government). In 2021, the provincial government announced a plan to merge all English-language school divisions into 15 regional catchment areas, overseen by a provincial education authority. There are five universities in Manitoba, regulated by the Ministry of Advanced Education and Literacy. Four of these universities are in Winnipeg: the University of Manitoba, the largest and most comprehensive; the University of Winnipeg, a liberal arts school primarily focused on undergrad studies downtown; Université de Saint-Boniface, the province's only French-language university; and the Canadian Mennonite University, a religious-based institution. The Université de Saint-Boniface, established in 1818 and now affiliated with the University of Manitoba, is the oldest university in Western Canada. Brandon University, formed in 1899 and in Brandon, is the province's only university not in Winnipeg. Manitoba has fifty-four public library systems. Of these, Winnipeg Public Library has the largest collections, at 1.1 million items as of 2020. The Minister of Culture, Heritage, Tourism and Sport is responsible for promoting and, to some extent, financing Manitoban culture. Manitoba is the birthplace of the Red River Jig, a combination of Indigenous pow-wows and European reels popular among early settlers. Manitoba's traditional music has strong roots in Métis and First Nations culture, in particular the old-time fiddling of the Métis. Manitoba's cultural scene also incorporates classical European traditions. The Winnipeg-based Royal Winnipeg Ballet (RWB), is Canada's oldest ballet and North America's longest continuously operating ballet company; it was granted its royal title in 1953 under Queen Elizabeth II. The Winnipeg Symphony Orchestra (WSO) performs classical music and new compositions at the Centennial Concert Hall. Manitoba Opera, founded in 1969, also performs out of the Centennial Concert Hall. Le Cercle Molière (founded 1925) is the oldest French-language theatre in Canada, and Royal Manitoba Theatre Centre (founded 1958) is Canada's oldest English-language regional theatre. Manitoba Theatre for Young People was the first English-language theatre to win the Canadian Institute of the Arts for Young Audiences Award, and offers plays for children and teenagers as well as a theatre school. The Winnipeg Art Gallery (WAG), Manitoba's largest art gallery and the sixth largest in the country, hosts an art school for children; the WAG's permanent collection comprises over twenty thousand works, with a particular emphasis on Manitoban and Canadian art. The 1960s pop group the Guess Who was formed in Manitoba, and later became the first Canadian band to have a No. 1 hit in the United States; Guess Who guitarist Randy Bachman later created Bachman–Turner Overdrive (BTO) with fellow Winnipeg-based musician Fred Turner. Fellow rocker Neil Young, grew up in Manitoba, and later played in Buffalo Springfield, and Crosby, Stills, Nash & Young. Folk rock band Crash Test Dummies formed in the late 1980s in Winnipeg and were the 1992 Juno Awards Group of the Year. Several prominent Canadian films were produced in Manitoba, such as The Stone Angel, based on the Margaret Laurence book of the same title, The Saddest Music in the World, Foodland, For Angela, and My Winnipeg. Major films shot in Manitoba include The Assassination of Jesse James by the Coward Robert Ford and Capote, both of which received Academy Award nominations. Falcon Beach, an internationally broadcast television drama, was filmed at Winnipeg Beach, Manitoba. Manitoba has a strong literary tradition. Bertram Brooker won the first-ever Governor General's Award for Fiction in 1936. Cartoonist Lynn Johnston, author of the comic strip For Better or For Worse, was a finalist for a 1994 Pulitzer Prize and inducted into the Canadian Cartoonist Hall of Fame. Margaret Laurence's The Stone Angel and A Jest of God were set in Manawaka, a fictional town representing Neepawa; the latter title won the Governor General's Award in 1966. Carol Shields won both the Governor General's Award and the Pulitzer Prize for The Stone Diaries. Gabrielle Roy, a Franco-Manitoban writer, won the Governor General's Award three times. A quote from her writings is featured on the Canadian $20 bill. Joan Thomas was nominated for the Governor General's Award twice and won in 2019 for Five Wives. The province has also been home to many of the key figures in Mennonite literature, including Governor General Award-winning Miriam Toews, Giller winner David Bergen, Armin Wiebe and many others. Sandra Birdsell, whose fiction focusses on her Métis and Mennonite heritage, was thrice nominated for the Governor General's Literary Award for English Language Fiction, and also for the Scotiabank Giller Prize in 2001. Festivals take place throughout the province, with the largest centred in Winnipeg. The Winnipeg Folk Festival has an annual attendance of over 70,000. The Festival du Voyageur is an annual ten-day event held in Winnipeg's French Quarter, and is Western Canada's largest winter festival. It celebrates Canada's fur-trading past and French-Canadian heritage and culture. Folklorama, a multicultural festival run by the Folk Arts Council, receives around 400,000 pavilion visits each year, of which about thirty percent are from non-Winnipeg residents. The Winnipeg Fringe Theatre Festival is an annual alternative theatre festival, the second-largest festival of its kind in North America (after the Edmonton International Fringe Festival). Manitoban museums document different aspects of the province's heritage. The Manitoba Museum is the largest museum in Manitoba and focuses on Manitoban history from prehistory to the 1920s. The full-size replica of the Nonsuch is the museum's showcase piece. The Manitoba Children's Museum at The Forks presents exhibits for children. There are two museums dedicated to the native flora and fauna of Manitoba: the Living Prairie Museum, a tall grass prairie preserve featuring 160 species of grasses and wildflowers, and FortWhyte Alive, a park encompassing prairie, lake, forest and wetland habitats, home to a large herd of bison. The Canadian Fossil Discovery Centre houses the largest collection of marine reptile fossils in Canada. Other museums feature the history of aviation, marine transport, and railways in the area. The Canadian Museum for Human Rights is the first Canadian national museum outside of the National Capital Region. Winnipeg has two daily newspapers: the Winnipeg Free Press, a broadsheet with the highest circulation numbers in Manitoba, as well as the Winnipeg Sun, a smaller tabloid-style paper. There are several ethnic weekly newspapers, including the weekly French-language La Liberté, and regional and national magazines based in the city. Brandon has two newspapers: the daily Brandon Sun and the weekly Wheat City Journal. Many small towns have local newspapers. There are five English-language television stations and one French-language station based in Winnipeg. The Global Television Network (owned by Canwest) is headquartered in the city. Winnipeg is home to twenty-one AM and FM radio stations, two of which are French-language stations. Brandon's five local radio stations are provided by Astral Media and Westman Communications Group. In addition to the Brandon and Winnipeg stations, radio service is provided in rural areas and smaller towns by Golden West Broadcasting, Corus Entertainment, and local broadcasters. CBC Radio broadcasts local and national programming throughout the province. Native Communications is devoted to indigenous programming and broadcasts to many of the isolated native communities as well as to larger cities. Manitoba has five professional sports teams: the Winnipeg Blue Bombers (Canadian Football League), the Winnipeg Jets (National Hockey League) and Manitoba Moose (American Hockey League), the Winnipeg Goldeyes (American Association), and Valour FC (Canadian Premier League). The province was previously home to another team called the Winnipeg Jets, which played in the World Hockey Association and National Hockey League from 1972 until 1996, when financial troubles prompted a sale and move of the team to Arizona, where they became the Phoenix Coyotes. A second incarnation of the Winnipeg Jets returned after True North Sports & Entertainment bought the Atlanta Thrashers and moved the team to Winnipeg in time for the 2011 hockey season. Manitoba has one major junior-level ice hockey team, the Western Hockey League's Brandon Wheat Kings, and one junior football team, the Winnipeg Rifles of the Canadian Junior Football League. It is also home to two teams in the Western Women's Canadian Football League: the Manitoba Fearless and Winnipeg Wolfpack. The Manitoba Herd, meanwhile, compete in the National Ringette League. The province is represented in university athletics by the university of Manitoba Bisons, the university of Winnipeg Wesmen, and the Brandon University Bobcats. All three teams compete in the Canada West Universities Athletic Association, a regional division of U Sports. Curling is an important winter sport in the province with Manitoba producing more men's national champions than any other province, while additionally in the top 3 women's national champions, as well as multiple world champions in the sport. The province also hosts the world's largest curling tournament in the MCA Bonspiel. Although not as prominent as ice hockey and curling, long track speed skating also features as a notable and top winter sport in Manitoba. The province has produced some of the world's best female speed skaters including Susan Auch and the country's top Olympic medal earners Cindy Klassen and Clara Hughes.